Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Is it better to use XXX.com or XXX.com/index.html as canonical page
-
Is it better to use 301 redirects or canonical page? I suspect canonical is easier. The question is, which is the best canonical page, YYY.com or YYY.com/indexhtml? I assume YYY.com, since there will be many other pages such as YYY.com/info.html, YYY.com/services.html, etc.
-
Glad you got it sorted out. If you're 301-redirecting a lot of domains, I'd suggest doing it gradually or maybe holding off on the lowest-quality domains. Google can see a massive set of redirects as a bit of a red flag (too many people have bought up cheap domains and 301-redirected to consolidate the link equity). If the domains are really all closely related or if you're only talking about a handful (<5) then it's probably not a big issue.
-
I think things may be sorted out, but I am not sure. I actually put in 301-redirects from a bunch of domains that I own to this new domain, the content of which will eventually replace my main domain. But, I need to get the domain properly set up and optimized before I move it to my primary domain to replace the ancient web site. At that time, I will also redirect this site to the new, old site.
I used to have Google ad-words tied to some of the domains that I 301-redirected to the new web site that I am building. Those were just a waste of money, however, so I put them on hold. I also had a lot of problems with semel and buttons for web bouncing off those pages that I re-directed. I put in .htaccess commands to stop those spam sites and that seems to work.
-
Google seems to be indexing 30-ish pages, but when I look at the cached home-page, I'm actually seeing the home-page of http://rfprototype.com/. Did you recently change domains or 301-redirect the old site? The cache data is around Christmas (after the original question was posted), so I think we're missing part of the puzzle here.
-
So, I think I may have had things wrong. For one thing, it seems like moz and Google are only indexing 2 pages, while the site index shows something like 80 pages. (I suspect an image is a page, and there are a lot of images. But, there are about 10 or 12 distinct pages at the moment. Also, Google and moz do not seem to show the correct key words in any sense like they should, leading me to think that they were just spidering 2 pages. I don't know why. I added the following to my index.html header:
and
I assume I put them in the correct place. I also believe I don't need canonical pages anywhere else.
Should these changes to my index.html make the proper changes?
-
Yeah, I'd have to concur - all the evidence and case studies I've seen suggest that rel=canonical almost always passes authority (link equity). There are exceptions, but honestly, there are exceptions with 301s, too.
I think the biggest difference, practically, is the impact on human visitors. 301-redirects take people to a new page, whereas canonical tags don't.
-
In terms of rel=canonical that will pass value the same as a 301 redirect - for evidence have a look here:
http://a-moz.groupbuyseo.org/learn/seo/canonicalization
"Another option for dealing with duplicate content is to utilize the rel=canonical tag. The rel=canonical tag passes the same amount of link juice (ranking power) as a 301 redirect, and often takes much less development time to implement."
See DR Pete's response in this Moz Q&A:
http://googlewebmastercentral.blogspot.co.uk/2009/02/specify-your-canonical.html
https://support.google.com/webmasters/answer/139066?rd=1
http://searchenginewatch.com/sew/how-to/2288690/how-and-when-to-use-301-redirects-vs-canonical
Matts Cutts stated there is not a whole lot of difference between the 301 and the canonical - they will both lose "just a tiny little amount bit, not very much at all" of credit from the referring page.
-
Ok, this is how I look at the situation.
So you have two URLs and the question is either to redirect301 or use canonical? In my opinion 301 is a better solution and this is because it will not only redirect people to the preferred version but the link value as well.
Whereas, with canonicals only search engines will know what is the preferred page but it will not transfer the link value which can help you with organic rankings.
Hope this helps!
-
You would put the canonical link in the index file and I would point that at the xxx.com version rather than the xxx.com/index.html version as people visiting your sites homepage are going to enter the domain and not the specific page so xxx.com rather than xxx.com/index.html.
There are some great articles on Moz explaining all this which I would suggest that you read -
http://a-moz.groupbuyseo.org/learn/seo/canonicalization
Dr Pete also did this post answering common questions on rel=canonical.
http://a-moz.groupbuyseo.org/blog/rel-confused-answers-to-your-rel-canonical-questions
In terms of 301 redirects and canonicalization both pass the same amount of authority gained by different pages. If you are trying to keep it as clean as possible you need to be careful you don't create an issue redirecting your index file to your domain - here is an old post explaining how moz solved this 301 redirect on an Apache server
http://a-moz.groupbuyseo.org/blog/apache-redirect-an-index-file-to-your-domain-without-looping
I personally find that if all your links on your site reference your preferred(canonical) URL for the homepage so in this case xxx.com and you redirect the www version to this or vice versa depending on your preference then you add a canonical in the index.html file pointing at xxx.com in this case unless you prefer to do it the other way round with www.xxx.com for both you will be fine.
Hope this helps
-
I forgot. Of course, there is no xxx.com page, per se. It is actually xxx.com/index.html so if you needed to put the canonical reference on xxx.com, how would you do it?
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
In writing the url, it is better to use the language used by the people of my country or English?
We speak Persian and all people search in Persian on Google. But I read in some sources that the url should be in English. Please tell me which language to use for url writing?
Technical SEO | | ghesta
For example, I brought down two models: 1fb0e134-10dc-4737-904f-bfdf07143a98-image.png https://ghesta.ir/blog/how-to-become-rich/
2)https://ghesta.ir/blog/چگونه-پولدار-شویم/0 -
How to check if an individual page is indexed by Google?
So my understanding is that you can use site: [page url without http] to check if a page is indexed by Google, is this 100% reliable though? Just recently Ive worked on a few pages that have not shown up when Ive checked them using site: but they do show up when using info: and also show their cached versions, also the rest of the site and pages above it (the url I was checking was quite deep) are indexed just fine. What does this mean? thank you p.s I do not have WMT or GA access for these sites
Technical SEO | | linklander0 -
How to Stop Google from Indexing Old Pages
We moved from a .php site to a java site on April 10th. It's almost 2 months later and Google continues to crawl old pages that no longer exist (225,430 Not Found Errors to be exact). These pages no longer exist on the site and there are no internal or external links pointing to these pages. Google has crawled the site since the go live, but continues to try and crawl these pages. What are my next steps?
Technical SEO | | rhoadesjohn0 -
How Does Google's "index" find the location of pages in the "page directory" to return?
This is my understanding of how Google's search works, and I am unsure about one thing in specific: Google continuously crawls websites and stores each page it finds (let's call it "page directory") Google's "page directory" is a cache so it isn't the "live" version of the page Google has separate storage called "the index" which contains all the keywords searched. These keywords in "the index" point to the pages in the "page directory" that contain the same keywords. When someone searches a keyword, that keyword is accessed in the "index" and returns all relevant pages in the "page directory" These returned pages are given ranks based on the algorithm The one part I'm unsure of is how Google's "index" knows the location of relevant pages in the "page directory". The keyword entries in the "index" point to the "page directory" somehow. I'm thinking each page has a url in the "page directory", and the entries in the "index" contain these urls. Since Google's "page directory" is a cache, would the urls be the same as the live website (and would the keywords in the "index" point to these urls)? For example if webpage is found at wwww.website.com/page1, would the "page directory" store this page under that url in Google's cache? The reason I want to discuss this is to know the effects of changing a pages url by understanding how the search process works better.
Technical SEO | | reidsteven750 -
How to determine which pages are not indexed
Is there a way to determine which pages of a website are not being indexed by the search engines? I know Google Webmasters has a sitemap area where it tells you how many urls have been submitted and how many are indexed out of those submitted. However, it doesn't necessarily show which urls aren't being indexed.
Technical SEO | | priceseo1 -
Can you have a /sitemap.xml and /sitemap.html on the same site?
Thanks in advance for any responses; we really appreciate the expertise of the SEOmoz community! My question: Since the file extensions are different, can a site have both a /sitemap.xml and /sitemap.html both siting at the root domain? For example, we've already put the html sitemap in place here: https://www.pioneermilitaryloans.com/sitemap Now, we're considering adding an XML sitemap. I know standard practice is to load it at the root (www.example.com/sitemap.xml), but am wondering if this will cause conflicts. I've been unable to find this topic addressed anywhere, or any real-life examples of sites currently doing this. What do you think?
Technical SEO | | PioneerServices0 -
OK to block /js/ folder using robots.txt?
I know Matt Cutts suggestions we allow bots to crawl css and javascript folders (http://www.youtube.com/watch?v=PNEipHjsEPU) But what if you have lots and lots of JS and you dont want to waste precious crawl resources? Also, as we update and improve the javascript on our site, we iterate the version number ?v=1.1... 1.2... 1.3... etc. And the legacy versions show up in Google Webmaster Tools as 404s. For example: http://www.discoverafrica.com/js/global_functions.js?v=1.1
Technical SEO | | AndreVanKets
http://www.discoverafrica.com/js/jquery.cookie.js?v=1.1
http://www.discoverafrica.com/js/global.js?v=1.2
http://www.discoverafrica.com/js/jquery.validate.min.js?v=1.1
http://www.discoverafrica.com/js/json2.js?v=1.1 Wouldn't it just be easier to prevent Googlebot from crawling the js folder altogether? Isn't that what robots.txt was made for? Just to be clear - we are NOT doing any sneaky redirects or other dodgy javascript hacks. We're just trying to power our content and UX elegantly with javascript. What do you guys say: Obey Matt? Or run the javascript gauntlet?0 -
Sitmap Page - HTML and XML
Hi there I have a domain which has a sitemap in html for regular users and a sitemap in xml for the spiders. I have a warning via seomoz saying that i have too many links on the html version. What do i do here? regards Stef
Technical SEO | | stefanok0