Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Oh no googlebot can not access my robots.txt file
- 
					
					
					
					
 I just receive a n error message from google webmaster Wonder it was something to do with Yoast plugin. Could somebody help me with troubleshooting this? Here's original message Over the last 24 hours, Googlebot encountered 189 errors while attempting to access your robots.txt. To ensure that we didn't crawl any pages listed in that file, we postponed our crawl. Your site's overall robots.txt error rate is 100.0%. Recommended action If the site error rate is 100%: - Using a web browser, attempt to access http://www.soobumimphotography.com//robots.txt. If you are able to access it from your browser, then your site may be configured to deny access to googlebot. Check the configuration of your firewall and site to ensure that you are not denying access to googlebot.
- If your robots.txt is a static page, verify that your web service has proper permissions to access the file.
- If your robots.txt is dynamically generated, verify that the scripts that generate the robots.txt are properly configured and have permission to run. Check the logs for your website to see if your scripts are failing, and if so attempt to diagnose the cause of the failure.
 If the site error rate is less than 100%: - Using Webmaster Tools, find a day with a high error rate and examine the logs for your web server for that day. Look for errors accessing robots.txt in the logs for that day and fix the causes of those errors.
- The most likely explanation is that your site is overloaded. Contact your hosting provider and discuss reconfiguring your web server or adding more resources to your website.
 After you think you've fixed the problem, use Fetch as Google to fetch http://www.soobumimphotography.com//robots.txt to verify that Googlebot can properly access your site. 
- 
					
					
					
					
 I can open text file but Godaddy told me robots.txt file is not on my server (root level). Also told me that my site is not crawled because robot.txt file is not there. Basically all of those might have resulted from plug in I was using (term optimizer) Based on what Godaddy told me, my .htaccess file was crashed because of that and had to be recreated. So now .htaceess file is good. Now I have to figure out is why my site is not accessible from Googlebot. Let me know Keith if this is a quick fix or need some time to troubleshoot. You can send me a message to discuss about fees if nessary. Thanks again 
- 
					
					
					
					
 Hi, You have a robots.txt file here: http://www.soobumimphotography.com/robots.txt Can you write this again in English so it makes sense? "I called Godaddy and told me if I used any plug ins etc. Godaddy fixed .htaccss file and my site was up and runningjust fine." Yes google xml sitemaps will add the location of your stitemap to the robots.txt file - but there is nothing wrong with your robots.txt file. 
- 
					
					
					
					
 I just called Godaddy and told me that I don't have robots.txt tile. Can anyone help with this issue? So here's what happen: I purchased Joos de Vailk's Term Optimizer to consolidate tags etc. As soon as I installed & opened it, my site crashed. I called Godaddy and told me if I used any plug ins etc. Godaddy fixed .htaccss file and my site was up and runningjust fine. Isn't plugin like the Google XML Sitemaps automatically generates robots.txt file? 
- 
					
					
					
					
 Yes, my site was down. 
- 
					
					
					
					
 I had a .htaccess issue past 24 hour with plug in and Godaddy had fixed it for me. I think this caused problem. I just fetched again and still getting unreachable page. I wonder if I have bad .htaccess file 
- 
					
					
					
					
 Was your site down during this period? I would recommend setting up pingdom.com (free site monitoring), this will email you if your site goes down - I suspect this is a hosting related issue. FYI, I can access your robots.txt fine from here. 
- 
					
					
					
					
 Hi Bistoss, You should log into Google Webmaster Tools to check the day the problem occurred. It is not uncommon for host to have problems that temporarily cause access problems. In some rare cases Google itself could be having problems. For example, in July we had 1 day with a 11% failure rate, it was the host. Since then no problems. If your problems are persistent, then you may have an issue like this: http://blog.jitbit.com/2012/08/fixing-googlebot-cant-access-your-site.html old Analytic code. Other things to look at is any recent changes, specifically anything that had to do with .htaccess Be sure to use the FETCH AS GOOGLE bot after any changes to verify that Google can now crawl your site. Hope this helps 
- 
					
					
					
					
 I also use Robots Meta Configuration plug in 
Browse Questions
Explore more categories
- 
		
		Moz ToolsChat with the community about the Moz tools. 
- 
		
		SEO TacticsDiscuss the SEO process with fellow marketers 
- 
		
		CommunityDiscuss industry events, jobs, and news! 
- 
		
		Digital MarketingChat about tactics outside of SEO 
- 
		
		Research & TrendsDive into research and trends in the search industry. 
- 
		
		SupportConnect on product support and feature requests. 
Related Questions
- 
		
		
		
		
		
		Unsolved Temporary redirect from 302 to 301 for PNG File?
 #302HTTP #temporaryredirect Technical SEO | | Damian_Ed 0
 Hi everyone, Recently I have faced a crawl issue with my media images on website. For example this page url https://intreface.com/wp-content/uploads/2022/12/Horion-screen-side-2.png has 302 HTTP Status and the recommendation is to change it 301. I have read the article on temporary redirections here:
 https://a-moz.groupbuyseo.org/learn/seo/redirection?_ga=2.45324708.1293586627.1702571936-916254120.1702571936
 but its not written here how to redirect in my HTML 1 image url not the landing page.
 Screenshot 2023-12-15 at 11.02.40.png
 I have messaged to MOZ Support but they recommended to go for the MOZ Community!
 Screenshot 2023-12-15 at 11.06.02.png Could you assist me wit this issue please? I can reach HTTML of the necessary page and change what I need for permanent redirection but firstly I need to understand how to do that correctly.0
- 
		
		
		
		
		
		Robots.txt Tester - syntax not understood
 I've looked in the robots.txt Tester and I can see 3 warnings: There is a 'syntax not understood' warning for each of these. XML Sitemaps: Technical SEO | | JamesHancocks1
 https://www.pkeducation.co.uk/post-sitemap.xml
 https://www.pkeducation.co.uk/sitemap_index.xml How do I fix or reformat these to remove the warnings? Many thanks in advance.
 Jim0
- 
		
		
		
		
		
		Multiple robots.txt files on server
 Hi! I have previously hired a developer to put up my site and noticed afterwards that he did not know much about SEO. This lead me to starting to learn myself and applying some changes step by step. One of the things I am currently doing is inserting sitemap reference in robots.txt file (which was not there before). But just now when I wanted to upload the file via FTP to my server I found multiple ones - in different sizes - and I dont know what to do with them? Can I remove them? I have downloaded and opened them and they seem to be 2 textfiles and 2 dupplicates. Names: robots.txt (original dupplicate) Technical SEO | | mjukhud
 robots.txt-Original (original)
 robots.txt-NEW (other content)
 robots.txt-Working (other content dupplicate) Would really appreciate help and expertise suggestions. Thanks!0
- 
		
		
		
		
		
		Block Domain in robots.txt
 Hi. We had some URLs that were indexed in Google from a www1-subdomain. We have now disabled the URLs (returning a 404 - for other reasons we cannot do a redirect from www1 to www) and blocked via robots.txt. But the amount of indexed pages keeps increasing (for 2 weeks now). Unfortunately, I cannot install Webmaster Tools for this subdomain to tell Google to back off... Any ideas why this could be and whether it's normal? I can send you more domain infos by personal message if you want to have a look at it. Technical SEO | | zeepartner0
- 
		
		
		
		
		
		Google insists robots.txt is blocking... but it isn't.
 I recently launched a new website. During development, I'd enabled the option in WordPress to prevent search engines from indexing the site. When the site went public (over 24 hours ago), I cleared that option. At that point, I added a specific robots.txt file that only disallowed a couple directories of files. You can view the robots.txt at http://photogeardeals.com/robots.txt Google (via Webmaster tools) is insisting that my robots.txt file contains a "Disallow: /" on line 2 and that it's preventing Google from indexing the site and preventing me from submitting a sitemap. These errors are showing both in the sitemap section of Webmaster tools as well as the Blocked URLs section. Bing's webmaster tools are able to read the site and sitemap just fine. Any idea why Google insists I'm disallowing everything even after telling it to re-fetch? Technical SEO | | ahockley0
- 
		
		
		
		
		
		No indexing url including query string with Robots txt
 Dear all, how can I block url/pages with query strings like page.html?dir=asc&order=name with robots txt? Thanks! Technical SEO | | HMK-NL0
- 
		
		
		
		
		
		Can I Disallow Faceted Nav URLs - Robots.txt
 I have been disallowing /*? So I know that works without affecting crawling. I am wondering if I can disallow the faceted nav urls. So disallow: /category.html/? /category2.html/? /category3.html/*? To prevent the price faceted url from being cached: /category.html?price=1%2C1000 Technical SEO | | tylerfraser
 and
 /category.html?price=1%2C1000&product_material=88 Thanks!0
- 
		
		
		
		
		
		Subdomain Removal in Robots.txt with Conditional Logic??
 I would like to see if there is a way to add conditional logic to the robots.txt file so that when we push from DEV to PRODUCTION and the robots.txt file is pushed, we don't have to remember to NOT push the robots.txt file OR edit it when it goes live. My specific situation is this: I have www.website.com, dev.website.com and new.website.com and somehow google has indexed the DEV.website.com and NEW.website.com and I'd like these to be removed from google's index as they are causing duplicate content. Should I: a) add 2 new GWT entries for DEV.website.com and NEW.website.com and VERIFY ownership - if I do this, then when the files are pushed to LIVE won't the files contain the VERIFY META CODE for the DEV version even though it's now LIVE? (hope that makes sense) b) write a robots.txt file that specifies "DISALLOW: DEV.website.com/" is that possible? I have only seen examples of DISALLOW with a "/" in the beginning... Hope this makes sense, can really use the help! I'm on a Windows Server 2008 box running ColdFusion websites. Technical SEO | | ErnieB0
 
			
		 
			
		 
			
		 
					
				 
					
				 
					
				 
					
				 
					
				 
					
				 
					
				