Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
What can I do if my reconsideration request is rejected?
-
Last week I received an unnatural link warning from Google. Sad times.
I followed the guidelines and reviewed all my inbound links for the last 3 months. All 5000 of them! Along with several genuine ones from trusted sites like BBC, Guardian and Telegraph there was a load of spam. About 2800 of them were junk. As we don't employ any SEO agency and don't buy links (we don't even buy adwords!) I know that all of this spam is generated by spam bots and site scrapers copying our content.
As the bad links have not been created by us and there are 2800 of them I cannot hope to get them removed. There are no 'contact us' pages on these Russian spam directories and Indian scraper sites. And as for the 'adult book marking website' who have linked to us over 1000 times, well I couldn't even contact that site in company time if I wanted to! As a result i did my manual review all day, made a list of 2800 bad links and disavowed them.
I followed this up with a reconsideration request to tell Google what I'd done but a week later this has been rejected "We've reviewed your site and we still see links to your site that violate our quality guidelines." As these links are beyond my control and I've tried to disavow them is there anything more to be done?
Cheers
Steve
-
Tom has given you good advice. I'll put in my 2 cents' worth as well.
There are 3 main reasons for a site to fail at reconsideration:
1. Not enough links were assessed by the site owner to be unnatural.
2. Not enough effort was put into removing links and documenting that to Google.
3. Improper use of the disavow tool.
In most cases #1 is the main cause. Almost every time I do a reconsideration request my client is surprised at what kind of links are considered unnatural. From what I have seen, Google is usually pretty good at figuring out whether you have been manually trying to manipulate the SERPS or whether links are just spam bot type of links.
Here are a few things to consider:
Are you being COMPLETELY honest with yourself about the spammy links you are seeing? How did Russian and porn sites end up linking to you? Most sites don't just get those by accident. Sometimes this can happen when sites use linkbuilding companies that use automated methods to build links. Even still, do all you can to address those links, and then for the ones that you can't get removed, document your efforts, show Google and then disavow them.
Even if these are foreign language sites, many of them will have whois emails that you can contact.
Are you ABSOLUTELY sure that your good links are truly natural? Just because they are from news sources is not a good enough reason. Have you read all the interflora stuff recently? They had a pile of links from advertorials (amongst other things) that now need to be cleaned up.
-
Hi Steve
If Google is saying there are still a few more links, then it might be an idea to manually review a few others that you haven't disavowed. I find the LinkDetox tool very useful for this. It's free with a tweet and will tell you if a link from a site is toxic (the site is deindexed) or if it's suspicious (and why it's suspicious). You still need to use your own judgement on these, but it might help you to find the extra links you're talking about.
However, there is a chance you have gone and disavowed every bad link, but still got the rejection. In this case, I'd keep trying but make your reconsideration request more detailed. Create an excel sheet and list the bad URLs and/or domains and give a reason explaining why you think they're bad links. Then provide information on how you found their contact details. If there are no contact us pages, check the whois registrar's email. After that, say when you contacted them (give a sample of your letter to them too), and if they replied, along with a follow up date if you got silence. If there are no details in the whois, explicitly mention that there are no contact details and so you have proceeded straight to disavowing.
Then list the URLs you've disavowed (upload the .txt file with your reconsideration email). You've now told Google that you've found bad links, why you think their bad (also include how you discovered them), that you've contacted the webmaster on numerous occasions and, if no removal was made, you've disavowed as a last resort. This is a very thorough process and uses the disavow tool in the way that Google wants us to - as a last resort to an unresponsive or anonymous webmaster.
Please forgive me if you've already done all this and it seems like repetition. I only mention it because I've found it's best to be as thorough as possible with Google in these situations. Remember, a reconsideration request is manual and if they see that you've gone through all this effort to be reinstated, you've got a better chance of being approved.
Keep trying, mate. It can be disheartening, but if you think it's worth the time and effort, then keep going for it. I would bear in mind the alternatives, however, such as starting fresh on a new domain. If you find yourself going round the bend with endless reconsiderations, sometimes your time, effort and expertise can be better put elsewhere.
All the best!
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
I am trying to generate GEO meta tag for my website where on one page there are multiple locations My question is, Can I add GEO tagging for every address?
Am I restricted to 1 geo tag per page or can i add multiple geo tags ?
Technical SEO | | lina_digital0 -
How can I get a photo album indexed by Google?
We have a lot of photos on our website. Unfortunately most of them don't seem to be indexed by Google. We run a party website. One of the things we do, is take pictures at events and put them on the site. An event page with a photo album, can have anywhere between 100 and 750 photo's. For each foto's there is a thumbnail on the page. The thumbnails are lazy loaded by showing a placeholder and loading the picture right before it comes onscreen. There is no pagination of infinite scrolling. Thumbnails don't have an alt text. Each thumbnail links to a picture page. This page only shows the base HTML structure (menu, etc), the image and a close button. The image has a src attribute with full size image, a srcset with several sizes for responsive design and an alt text. There is no real textual content on an image page. (Note that when a user clicks on the thumbnail, the large image is loaded using JavaScript and we mimic the page change. I think it doesn't matter, but am unsure.) I'd like that full size images should be indexed by Google and found with Google image search. Thumbnails should not be indexed (or ignored). Unfortunately most pictures aren't found or their thumbnail is shown. Moz is giving telling me that all the picture pages are duplicate content (19,521 issues), as they are all the same with the exception of the image. The page title isn't the same but similar for all images of an album. Example: On the "A day at the park" event page, we have 136 pictures. A site search on "a day at the park" foto, only reveals two photo's of the albums. 3QolbbI.png QTQVxqY.jpg mwEG90S.jpg
Technical SEO | | jasny0 -
Can I set a canonical tag to an anchor link?
I have a client who is moving to a one page website design. So, content from the inner pages is being condensed in to sections on the 'home' page. There will be a navigation that anchor links to each relevant section. I am wondering if I should leave the old pages and use rel=canonical to point them to their relevant sections on the new 'home' page rather than 301 them. Thoughts?
Technical SEO | | Vizergy0 -
Blocked URL parameters can still be crawled and indexed by google?
Hy guys, I have two questions and one might be a dumb question but there it goes. I just want to be sure that I understand: IF I tell webmaster tools to ignore an URL Parameter, will google still index and rank my url? IS it ok if I don't append in the url structure the brand filter?, will I still rank for that brand? Thanks, PS: ok 3 questions :)...
Technical SEO | | catalinmoraru0 -
Can spiders crawl jQuery Fancy Box scripts
Hi Everyone - I'm not a technical person at all. I have some content that will be hidden until a user clicks "learn more" where upon it will be displayed via jQuery Fancy Box script. The content behind the learn more javascript is important and I need it to be crawled by search engine spiders. Does anyone know if there will be a problem with this script?
Technical SEO | | Santaur0 -
Can I remove 301 redirects after some time?
Hello, We have an very large number of 301 redirects on our site and would like to find a way to remove some of them. Is there a time frame after which Google does not need a 301 any more? For example if A is 301 redirected to B, does Google know after a while not to serve A any more, and replaces any requests for A with B? How about any links that go to A? Or: Is the only option to have all links that pointed to A point to B and then the 301 can be removed after some time? Thank you for you you help!
Technical SEO | | Veva0 -
What can I do about missing Meta Description for category pagest etc.?
On all my campaigns I'm returning high levels of 'Missing Meta Description Tags'. The problem with fixing this is they're all for category, tag and author pages. Is there a way to add a meta description to these pages (there are hundreds) or will it not really have any ranking effect?
Technical SEO | | SiliconBeachTraining0 -
On a dedicated server with multiple IP addresses, how can one address group be slow/time out and all other IP addresses OK?
We utilize a dedicated server to host roughly 60 sites on. The server is with a company that utilizes a lady who drives race cars.... About 4 months ago we realized we had a group of sites down thanks to monitoring alerts and checked it out. All were on the same IP address and the sites on the other IP address were still up and functioning well. When we contacted the support at first we were stonewalled, but eventually they said there was a problem and it was resolved within about 2 hours. Up until recently we had no problems. As a part of our ongoing SEO we check page load speed for our clients. A few days ago a client who has their site hosted by the same company was running very slow (about 8 seconds to load without cache). We ran every check we could and could not find a reason on our end. The client called the host and were told they needed to be on some other type of server (with the host) at a fee increase of roughly $10 per month. Yesterday, we noticed one group of sites on our server was down and, again, it was one IP address with about 8 sites on it. On chat with support, they kept saying it was our ISP. (We speed tested on multiple computers and were 22MB down and 9MB up +/-2MB). We ran a trace on the IP address and it went through without a problem on three occassions over about ten minutes. After about 30 minutes the sites were back up. Here's the twist: we had a couple of people in the building who were on other ISP's try and the sites came up and loaded on their machines. Does anyone have any idea as to what the issue is?
Technical SEO | | RobertFisher0