Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Google Search console says 'sitemap is blocked by robots?
- 
					
					
					
					
 Google Search console is telling me "Sitemap contains URLs which are blocked by robots.txt." I don't understand why my sitemap is being blocked? My robots.txt look like this: User-Agent: * 
 Disallow:It's a WordPress site, with Yoast SEO installed. Is anyone else having this issue with Google Search console? Does anyone know how I can fix this issue? 
- 
					
					
					
					
 Nice happy to hear that do you work with Greg Reindel? He is a good friend I looked at your IP that is why I ask? Tom 
- 
					
					
					
					
 I agree with David Hey is your dev Greg Reindel? If so you can call me for help PM me here for my info. Thomas Zickell 
- 
					
					
					
					
 Hey guys, I ended up disabling the sitemap option from YoastSEO, then installed the 'Google (XML) sitemap' plug-in. I re-submitted the sitemap to Google last night, and it came back with no issues. I'm glad to finally have this sorted out. Thanks for all the help! 
- 
					
					
					
					
 Hi Christian, The current robots.txt shouldn't be blocking those URLs. Did you or someone else recently change the robots.txt file? If so, give Google a few days to re-crawl your site. Also, can you check what happens when you do a fetch and render on one of the blocked posts in Search Console? Do you have issues there? Cheers, David 
- 
					
					
					
					
 I think you need to make an https robots.txt file if you are running https if running https https://a-moz.groupbuyseo.org/blog/xml-sitemaps `User-agent: * Disallow: /wp-admin/ Allow: /wp-admin/admin-ajax.php` Sitemap: https://domain.com/index-sitemap.xml(that is a https site map)can you send the sitemap URL or run it though deepcrawl Hope this helps? Did you make a new robots.txt file? 
- 
					
					
					
					
 Thanks for the response. Do you think this is a robots.txt issue? Or could this be caused by the YoastSEO plugin? Do you know if this plug-in works with YoastSEO together? Or will it cause issues? 
- 
					
					
					
					
 Thank you for the response. I just scanned the site using 'Screaming frog'. Under Internal>Directives there were zero 'no index' links. I also check for '404 errors', server 505 errors, or anything 'blocked by robots.txt'. Google search console is still showing me that there are URL's being blocked by my sitemap. (I added a screenshot of this). When I click through, it tells me that the 'post sitemap' has over +300 warnings. I have just deleted the YoastSEO plugin, and I am now re-installing it. hopefully, this fixes the issue. 
- 
					
					
					
					
 No, you do not need to change or plug-in what is happening is Webmaster tools is telling you that you have no index or no follow were robots xTag somewhere on your URLs inside your sitemap. Run your site through Moz, screaming frog Seo spider or deepcrawl and look for no indexed URLs. webmaster tools/search console is telling you that you have no index URLs inside of your XML sitemap not that you robots.txt is blocking it. This would be set in the Yoast plugin. one way to correct it is to look for noindex URLs & filter them inside Yoast so they are not being presented to the crawlers. If you would like you can turn off the sitemap on Yoast and turn it back on if that does not work I recommend completely removing the plug-in and reinstalling it - https://kb.yoast.com/kb/how-can-i-uninstall-my-plugin/
- https://kinsta.com/blog/uninstall-wordpress-plugin/
 Can you send a screenshot of what you're seeing? When you see it in Google Webmaster tools are you talking about the XML sitemap itself mean no indexed because all XML sitemaps are no indexed. Please add this to your robots.txt `User-agent:* Disallow:/wp-admin/ Allow:/wp-admin/admin-ajax.php` Sitemap: http://www.website.com/sitemap_index.xmlI hope this is of help, Tom 
- 
					
					
					
					
 Hi, Use this plugin https://wordpress.org/plugins/wp-robots-txt/ it will remove previous robots.txt and set simple wordpress robots.txt and wait for a day problem can be solved. Also watch this video on the same @ https://www.youtube.com/watch?v=DZiyN07bbBM Thanks 
Browse Questions
Explore more categories
- 
		
		Moz ToolsChat with the community about the Moz tools. 
- 
		
		SEO TacticsDiscuss the SEO process with fellow marketers 
- 
		
		CommunityDiscuss industry events, jobs, and news! 
- 
		
		Digital MarketingChat about tactics outside of SEO 
- 
		
		Research & TrendsDive into research and trends in the search industry. 
- 
		
		SupportConnect on product support and feature requests. 
Related Questions
- 
		
		
		
		
		
		Role of Robots.txt and Search Console parameters settings
 Hi, wondering if anyone can point me to resources or explain the difference between these two. If a site has url parameters disallowed in Robots.txt is it redundant to edit settings in Search Console parameters to anything other than "Let Googlebot Decide"? Technical SEO | | LivDetrick0
- 
		
		
		
		
		
		Crawl solutions for landing pages that don't contain a robots.txt file?
 My site (www.nomader.com) is currently built on Instapage, which does not offer the ability to add a robots.txt file. I plan to migrate to a Shopify site in the coming months, but for now the Instapage site is my primary website. In the interim, would you suggest that I manually request a Google crawl through the search console tool? If so, how often? Any other suggestions for countering this Meta Noindex issue? Technical SEO | | Nomader1
- 
		
		
		
		
		
		Spam URL'S in search results
 We built a new website for a client. When I do 'site:clientswebsite.com' in Google it shows some of the real, recently submitted pages. But it also shows many pages of spam url results, like this 'clientswebsite.com/gockumamaso/22753.htm' - all of which then go to the sites 404 page. They have page titles and meta descriptions in Chinese or Japanese too. Some of the urls are of real pages, and link to the correct page, despite having the same Chinese page titles and descriptions in the SERPS. When I went to remove all the spammy urls in Search Console (it only allowed me to temporarily hide them), a whole load of new ones popped up in the SERPS after a day or two. The site files itself are all fine, with no errors in the server logs. All the usual stuff...robots.txt, sitemap etc seems ok and the proper pages have all been requested for indexing and are slowly appearing. The spammy ones continue though. What is going on and how can I fix it? Technical SEO | | Digital-Murph0
- 
		
		
		
		
		
		Why is Google Webmaster Tools showing 404 Page Not Found Errors for web pages that don't have anything to do with my site?
 I am currently working on a small site with approx 50 web pages. In the crawl error section in WMT Google has highlighted over 10,000 page not found errors for pages that have nothing to do with my site. Anyone come across this before? Technical SEO | | Pete40
- 
		
		
		
		
		
		Blocked URL parameters can still be crawled and indexed by google?
 Hy guys, I have two questions and one might be a dumb question but there it goes. I just want to be sure that I understand: IF I tell webmaster tools to ignore an URL Parameter, will google still index and rank my url? IS it ok if I don't append in the url structure the brand filter?, will I still rank for that brand? Thanks, PS: ok 3 questions :)... Technical SEO | | catalinmoraru0
- 
		
		
		
		
		
		Blocking URL's with specific parameters from Googlebot
 Hi, I've discovered that Googlebot's are voting on products listed on our website and as a result are creating negative ratings by placing votes from 1 to 5 for every product. The voting function is handled using Javascript, as shown below, and the script prevents multiple votes so most products end up with a vote of 1, which translates to "poor". How do I go about using robots.txt to block a URL with specific parameters only? I'm worried that I might end up blocking the whole product listing, which would result in de-listing from Google and the loss of many highly ranked pages. DON'T want to block: http://www.mysite.com/product.php?productid=1234 WANT to block: http://www.mysite.com/product.php?mode=vote&productid=1234&vote=2 Javacript button code: onclick="javascript: document.voteform.submit();" Thanks in advance for any advice given. Regards, Technical SEO | | aethereal
 Asim0
- 
		
		
		
		
		
		Should we use Google's crawl delay setting?
 We’ve been noticing a huge uptick in Google’s spidering lately, and along with it a notable worsening of render times. Yesterday, for example, Google spidered our site at a rate of 30:1 (google spider vs. organic traffic.) So in other words, for every organic page request, Google hits the site 30 times. Our render times have lengthened to an avg. of 2 seconds (and up to 2.5 seconds). Before this renewed interest Google has taken in us we were seeing closer to one second average render times, and often half of that. A year ago, the ratio of Spider to Organic was between 6:1 and 10:1. Is requesting a crawl-delay from Googlebot a viable option? Our goal would be only to reduce Googlebot traffic, and hopefully improve render times and organic traffic. Thanks, Trisha Technical SEO | | lzhao0
 
			
		 
			
		 
			
		 
			
		 
					
				 
					
				 
					
				 
					
				 
					
				 
					
				 
					
				