Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
My brand name has 2 words but Google only indexing as 1 word. Is there a fix?
- 
					
					
					
					
 Hi all...I'm at a loss. I've never had this happen. Google only shows pages of my site when I search the brand name as one word. When I Google the site as one word BrandBrand- it only shows my blog page and about us page plus Twitter and Facebook on page 1. The homepage does not show up at all. When I Google the site as two words Brand Brand - My Facebook page is on page 1 but nothing else. The homepage isn't showing up at all. When I search both words on Bing and Yahoo both are indexing it as two words and shows on page 1. Any ideas? 
- 
					
					
					
					
 @TexasBlogger Here are a few potential reasons for the issue: Google's indexing: Sometimes, Google's indexing algorithms may take time to properly crawl and index your website. It's possible that your homepage hasn't been fully indexed yet. Brand name competition: If your brand name consists of common words or phrases, it could face tougher competition in search results. Other websites or brands with similar names might be overshadowing your homepage. Search engine preferences: Different search engines can have variations in their algorithms and indexing processes. It's possible that Bing and Yahoo are interpreting your brand name as two separate words, while Google treats it as a single word. This can affect the way your homepage appears in their search results. Website optimization: Double-check your homepage's optimization to ensure that it contains relevant keywords, meta tags, and descriptive content. Make sure your homepage is properly optimized for search engines to understand its relevance. Backlinks and authority: Building high-quality backlinks from reputable websites can improve your website's authority and visibility in search results. Consider reaching out to relevant websites or industry influencers for potential collaborations or partnerships. To address these issues, you can try the following steps: Submit your website's sitemap to Google Search Console to ensure that it is properly indexed. Continue to optimize your homepage by including relevant keywords, improving meta tags, and enhancing the content. Consider enhancing your website's overall SEO strategy, including link-building efforts, to improve its visibility and authority. If feasible, consider consulting with an SEO specialist who can analyze your website's specific situation and provide tailored recommendations. 
- 
					
					
					
					
 Right but most pages of my site have been indexed. They're not ranking well yet and that's what takes time. The indexing happens quickly, it's the ranking that takes time. I appreciate your responses, though. 
- 
					
					
					
					
 In my experience it can even take 2 to 4 months really. It's why i focus more on domains with a longer age then newly registered ones. I have domains of over 10 year old in relation of registration and those index in just one complete night if i renew the website. Just saying. 
- 
					
					
					
					
 Great. I'll wait it out then. The reason I reached out is I've seen aged sites with this same issue. Let's see what happens in a few weeks. Thanks! 
- 
					
					
					
					
 New sites are damped to prevent quick spammers launching up domains on to go. So it takes a while or some time for google to release that and start listing you actively. 
- 
					
					
					
					
 It's a new site. It's doesn't list at all...I checked through page 14 with the two words. But it does with one. There are no backlinks yet as it's new. But it should still be showing results for the brand name search. 
- 
					
					
					
					
 Site age? Incoming links? Did you submit the website to even Google? Does it list at all? Like page 2 or 3 etc? If thats the case, i'd suggest inspecting your incoming link profile and adjust links accordingly. It might be considered spammy. 
Browse Questions
Explore more categories
- 
		
		Moz ToolsChat with the community about the Moz tools. 
- 
		
		SEO TacticsDiscuss the SEO process with fellow marketers 
- 
		
		CommunityDiscuss industry events, jobs, and news! 
- 
		
		Digital MarketingChat about tactics outside of SEO 
- 
		
		Research & TrendsDive into research and trends in the search industry. 
- 
		
		SupportConnect on product support and feature requests. 
Related Questions
- 
		
		
		
		
		
		Google Not Indexing Pages (Wordpress)
 Hello, recently I started noticing that google is not indexing our new pages or our new blog posts. We are simply getting a "Discovered - Currently Not Indexed" message on all new pages. When I click "Request Indexing" is takes a few days, but eventually it does get indexed and is on Google. This is very strange, as our website has been around since the late 90's and the quality of the new content is neither duplicate nor "low quality". We started noticing this happening around February. We also do not have many pages - maybe 500 maximum? I have looked at all the obvious answers (allowing for indexing, etc.), but just can't seem to pinpoint a reason why. Has anyone had this happen recently? It is getting very annoying having to manually go in and request indexing for every page and makes me think there may be some underlying issues with the website that should be fixed. Technical SEO | | Hasanovic1
- 
		
		
		
		
		
		URLs dropping from index (Crawled, currently not indexed)
 I've noticed that some of our URLs have recently dropped completely out of Google's index. When carrying out a URL inspection in GSC, it comes up with 'Crawled, currently not indexed'. Strangely, I've also noticed that under referring page it says 'None detected', which is definitely not the case. I wonder if it could be something to do with the following? https://www.seroundtable.com/google-ranking-index-drop-30192.html - It seems to be a bug affecting quite a few people. Here are a few examples of the URLs that have gone missing: https://www.ihasco.co.uk/courses/detail/sexual-harassment-awareness-training https://www.ihasco.co.uk/courses/detail/conflict-resolution-training https://www.ihasco.co.uk/courses/detail/prevent-duty-training Any help here would be massively appreciated! Technical SEO | | iHasco0
- 
		
		
		
		
		
		Google is still indexing the old domain a year after 301 redirects are put in place
 Hi there, You might have experienced this before but for me this is the first. A client of mine moved from domain A (www.domainA.com) to domain B (www.domainB.com). 301 redirects are all in place for over a year. But the old domain is still showing in Google when you search for "site:domainA.com" The HTTP Header check shows this result for the URL https://www.domainA.com/company/cookie-policy.aspx HTTP/1.1 301 Moved Permanently => Technical SEO | | iQi
 Cache-Control => private
 Content-Length => 174
 Content-Type => text/html; charset=utf-8
 Location => https://www.domain_B_.com/legal/cookie-policy
 Server => Microsoft-IIS/10.0
 X-AspNetMvc-Version => 5.2
 X-AspNet-Version => 4.0.30319
 X-Powered-By => ASP.NET
 Date => Fri, 15 Mar 2019 12:01:33 GMT
 Connection => close Does the redirect look wrong? The change of address request was made on Google Console when the website was moved over a year ago. Edit: Checked the domainA.com on bing and it seems that its not indexed, and replaced with domainB.com, which is the right. Just Google is indexing the old domain! Please let me know your thoughts on why this is happening. Best,0
- 
		
		
		
		
		
		Google tries to index non existing language URLs. Why?
 Hi, I am working for a SAAS client. He uses two different language versions by using two different subdomains. Technical SEO | | TheHecksler
 de.domain.com/company for german and en.domain.com for english. Many thousands URLs has been indexed correctly. But Google Search Console tries to index URLs which were never existing before and are still not existing. de.domain.com**/en/company
 en.domain.com/de/**company ... and an thousand more using the /en/ or /de/ in between. We never use this variant and calling these URLs will throw up a 404 Page correctly (but with wrong respond code - we`re fixing that 😉 ). But Google tries to index these kind of URLs again and again. And, I couldnt find any source of these URLs. No Website is using this as an out going link, etc.
 We do see in our logfiles, that a Screaming Frog Installation and a-moz.groupbuyseo.org w opensiteexplorer were trying to access this earlier. My Question: How does Google comes up with that? From where did they get these URLs, that (to our knowledge) never existed? Any ideas? Thanks 🙂0
- 
		
		
		
		
		
		How can I get a photo album indexed by Google?
 We have a lot of photos on our website. Unfortunately most of them don't seem to be indexed by Google. We run a party website. One of the things we do, is take pictures at events and put them on the site. An event page with a photo album, can have anywhere between 100 and 750 photo's. For each foto's there is a thumbnail on the page. The thumbnails are lazy loaded by showing a placeholder and loading the picture right before it comes onscreen. There is no pagination of infinite scrolling. Thumbnails don't have an alt text. Each thumbnail links to a picture page. This page only shows the base HTML structure (menu, etc), the image and a close button. The image has a src attribute with full size image, a srcset with several sizes for responsive design and an alt text. There is no real textual content on an image page. (Note that when a user clicks on the thumbnail, the large image is loaded using JavaScript and we mimic the page change. I think it doesn't matter, but am unsure.) I'd like that full size images should be indexed by Google and found with Google image search. Thumbnails should not be indexed (or ignored). Unfortunately most pictures aren't found or their thumbnail is shown. Moz is giving telling me that all the picture pages are duplicate content (19,521 issues), as they are all the same with the exception of the image. The page title isn't the same but similar for all images of an album. Example: On the "A day at the park" event page, we have 136 pictures. A site search on "a day at the park" foto, only reveals two photo's of the albums. 3QolbbI.png QTQVxqY.jpg mwEG90S.jpg Technical SEO | | jasny0
- 
		
		
		
		
		
		Fake Links indexing in google
 Hello everyone, I have an interesting situation occurring here, and hoping maybe someone here has seen something of this nature or be able to offer some sort of advice. So, we recently installed a wordpress to a subdomain for our business and have been blogging through it. We added the google webmaster tools meta tag and I've noticed an increase in 404 links. I brought this up to or server admin, and he verified that there were a lot of ip's pinging our server looking for these links that don't exist. We've combed through our server files and nothing seems to be compromised. Today, we noticed that when you do site:ourdomain.com into google the subdomain with wordpress shows hundreds of these fake links, that when you visit them, return a 404 page. Just curious if anyone has seen anything like this, what it may be, how we can stop it, could it negatively impact us in anyway? Should we even worry about it? Here's the link to the google results. https://www.google.com/search?q=site%3Amshowells.com&oq=site%3A&aqs=chrome.0.69i59j69i57j69i58.1905j0j1&sourceid=chrome&es_sm=91&ie=UTF-8 (odd links show up on pages 2-3+) Technical SEO | | mshowells0
- 
		
		
		
		
		
		How to Remove /feed URLs from Google's Index
 Hey everyone, I have an issue with RSS /feed URLs being indexed by Google for some of our Wordpress sites. Have a look at this Google query, and click to show omitted search results. You'll see we have 500+ /feed URLs indexed by Google, for our many category pages/etc. Here is one of the example URLs: http://www.howdesign.com/design-creativity/fonts-typography/letterforms/attachment/gilhelveticatrade/feed/. Based on this content/code of the XML page, it looks like Wordpress is generating these: <generator>http://wordpress.org/?v=3.5.2</generator> Any idea how to get them out of Google's index without 301 redirecting them? We need the Wordpress-generated RSS feeds to work for various uses. My first two thoughts are trying to work with our Development team to see if we can get a "noindex" meta robots tag on the pages, by they are dynamically-generated pages...so I'm not sure if that will be possible. Or, perhaps we can add a "feed" paramater to GWT "URL Parameters" section...but I don't want to limit Google from crawling these again...I figure I need Google to crawl them and see some code that says to get the pages out of their index...and THEN not crawl the pages anymore. I don't think the "Remove URL" feature in GWT will work, since that tool only removes URLs from the search results, not the actual Google index. FWIW, this site is using the Yoast plugin. We set every page type to "noindex" except for the homepage, Posts, Pages and Categories. We have other sites on Yoast that do not have any /feed URLs indexed by Google at all. Side note, the /robots.txt file was previously blocking crawling of the /feed URLs on this site, which is why you'll see that note in the Google SERPs when you click on the query link given in the first paragraph. Technical SEO | | M_D_Golden_Peak0
- 
		
		
		
		
		
		CDN Being Crawled and Indexed by Google
 I'm doing a SEO site audit, and I've discovered that the site uses a Content Delivery Network (CDN) that's being crawled and indexed by Google. There are two sub-domains from the CDN that are being crawled and indexed. A small number of organic search visitors have come through these two sub domains. So the CDN based content is out-ranking the root domain, in a small number of cases. It's a huge duplicate content issue (tens of thousands of URLs being crawled) - what's the best way to prevent the crawling and indexing of a CDN like this? Exclude via robots.txt? Additionally, the use of relative canonical tags (instead of absolute) appear to be contributing to this problem as well. As I understand it, these canonical tags are telling the SEs that each sub domain is the "home" of the content/URL. Thanks! Scott Technical SEO | | Scott-Thomas0
 
			
		 
				
		 
			
		 
					
				 
					
				 
					
				 
					
				 
					
				 
					
				 
					
				