Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Solved Should I consolidate my "www" and "non-www" pages?
-
My page rank for www and non-www is the same. In one keyword instance, my www version performs SO much better.
Wanting to consolidate to one or the other. My question is as to whether all these issues would ultimately resolve to my chosen consolidated domain (i.e. www or non-www) regardless of which one I choose. OR, would it be smart to choose the one where I am already ranking high for this significant keyword phrase?
Thank you in advance for your help.
-
It may be that one version (www or non-www) has more historical links. You say your PageRank for both is the same, but how are you checking that? Google's public PageRank has not been updated in a decade or so.
Either way, I'd generally say that if you pick one version and stick to it (redirect the other, e.g. so every non-www. URL points to its www. equivalent), you should maintain all rankings. There is a theoretical advantage to picking the version with more links, but in my experience in practice this type of migration tends to be smooth.
-
Require the www Options +FollowSymLinks RewriteEngine On RewriteBase / RewriteCond %{HTTP_HOST} !^www\.askapache\.com$ [NC] RewriteRule ^(.*)$ https://www.askapache.com/$1 [R=301,L]
-
Yes. I would recommend picking the version (either www or non-www) that has the historical data showing it performs better than the other version. Check the list of indexed pages for each of the versions to compare. Ideally both the www and non-www version of the website will be indexed in Google so it will help you to decide which version makes the most sense to consolidate to.
Once you identify the preferred version, set 301 redirects from the non-preferred URLs to the preferred version of each URL (the one that has more traffic, links, authority, etc.) of the site. This should be done site-wide so that all URLs are either www or non-www, it shouldn’t be a mix of both. In my experience, I’ve found that between 90-99% of the Site’s SEO Authority is preserved when setting a permanent 301 redirect.
-
@meditationbunny Sorry for the slow reply - but yes, I'd expect Page Authority to increase slightly, if the "other" version had any value to it.
For Page Optimization, yes. For example, for my own site I see:
http://tcapper.co.uk redirects to https://www.tcapper.co.uk/. This on-page analysis is for https://www.tcapper.co.uk/.
-
It may be that one version (www or non-www) has more historical links. You say your PageRank for both is the same, but how are you checking that? Google's public PageRank has not been updated in a decade or so.
Either way, I'd generally say that if you pick one version and stick to it (redirect the other, e.g. so every non-www. URL points to its www. equivalent), you should maintain all rankings. There is a theoretical advantage to picking the version with more links, but in my experience in practice this type of migration tends to be smooth.
-
@tom-capper
Thank you. Yes, I should be more clear. I am calling it page rank, when I am actually referring to Moz's domain authority and Moz's keyword ranking. Still, I believe you answered my question. Under page optimization, I can see what appear to be duplicate listings of my pages along with different SERP ranking. It was confusing until I realized that one was the www and the other was non-www. I have since added code to my .htaccess file that will send everything to www. Can I expect the page optimization section to now only show www versions of the pages? Also, can I expect page authority to increase because it is no longer a mish-mash and is all headed to the same domain and same pages (i.e. www version)? -
It may be that one version ("www" or "non-www") has more historical links. You say your PageRank for both is the same, but how are you checking that? Google's public PageRank has not been updated in a decade or so.
Either way, I'd generally say that if you pick one version and stick to it (redirect the other, e.g. so every non-www. URL points to its www. equivalent), you should maintain all rankings. There is a theoretical advantage to picking the version with more links, but in my experience, in practice, this type of migration tends to be smooth.
-
This post is deleted! -
This post is deleted! -
This post is deleted!
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Page with "random" content
Hi, I'm creating a page of 300+ in the near future, on which the content basicly will be unique as it can be. However, upon every refresh, also coming from a search engine refferer, i want the actual content such as listing 12 business to be displayed random upon every hit. So basicly we got 300+ nearby pages with unique content, and the overview of those "listings" as i might say, are being displayed randomly. Ive build an extensive script and i disabled any caching for PHP files in specific these pages, it works. But what about google? The content of the pages will still be as it is, it is more of the listings that are shuffled randomly to give every business listing a fair shot at a click and so on. Anyone experience with this? Ive tried a few things in the past, like a "Last update PHP Month" in the title which sometimes is'nt picked up very well.
Technical SEO | | Vanderlindemedia0 -
Non Published Wordpress Pages
Hi, Is there any negative SEO consequences from having too many pages private or not published. Can it like slow the site down or does it not matter? Someone in my dept. has so many pages started/not complete and besides being messy, I wonder if it has any negative impact on the site. Thanks
Technical SEO | | aua1 -
301 Redirect non existant pages
Hi I have 100's of URL's appearing in Search Console for example: ?p=1_1 These go to on to 5_200 etc.. I have tried to do htaccess and the mod rewrite is on as I can redirect directories to the root i.e RewriteRule ^web_example(.*)$ /$1 [R=301,N,L] However I have tried all kinds of variations to redirect ?p= and either it doesn't work at all or it crashes the website. Can anyone point me in the right direction to fix this.
Technical SEO | | Cocoonfxmedia0 -
New "Static" Site with 302s
Hey all, Came across a bit of an interesting challenge recently, one that I was hoping some of you might have had experience with! We're currently in the process of a website rebuild, for which I'm really excited. The new site is using Markdown to create an entirely static site. Load-times are fantastic, and the code is clean. Life is good, apart from the 302s. One of the weird quirks I've realized is that with oldschool, non-server-generated page content is that every page of the site is an Index.html file in a directory. The resulting in a www.website.com/page-title will 302 to www.website.com/page-title/. My solution off the bat has been to just be super diligent and try to stay on top of the link profile and send lots of helpful emails to the staff reminding them about how to build links, but I know that even the best laid plans often fail. Has anyone had a similar challenge with a static site and found a way to overcome it?
Technical SEO | | danny.wood1 -
Using the word "FREE" in domain name
Hi, This may seem like a simple question but a new client of mine wishes to use a domain name with the word "free" in it. The website will offer free activity vouchers. I couldn't see this being a problem as there a lot of websites that do this although he was told it may present a problem with the search engines thinking the site was spammy. It won't be and will be offering information and vouchers on local sporting activities. I was wondering if anybody could clarify this please so I can give him a more definitive answer to his question. Thanks for your help.
Technical SEO | | malinkymedia0 -
Splitting Page Authority with two URLs for the same page.
Hello guys, My website is currently holding two different URLs for the same page and I am under the impression such set up is dividing my Page Authority and Link Juice. We currently have the following page with both URLs below: www.wbresearch.com/soldiertechnologyusa/home.aspx
Technical SEO | | JoaoPdaCosta-WBR
www.wbresearch.com/soldiertechnologyusa/ Analysing the page authority and backlinks I identified that we are splitting the amount of backlinks (links from sites, social media and therefore authority). "/home.aspx"
PA: 67
Linking Root Domains: 52
Total Links: 272 "/"
PA: 64
Linking Root Domains: 29
Total Links: 128 I am under the impression that if the URLs were the same we would maximise our backlinks and therefore page authority. My Question: How can I fix this? Should I have a 301 redirect from the page "/" to the "/home.aspx" therefore passing the authority and link juice of “/” directly to “/homes.aspx”? Trying to gather thoughts and ideas on this, suggestions are much appreciated? Thanks!0 -
Determining When to Break a Page Into Multiple Pages?
Suppose you have a page on your site that is a couple thousand words long. How would you determine when to split the page into two and are there any SEO advantages to doing this like being more focused on a specific topic. I noticed the Beginner's Guide to SEO is split into several pages, although it would concentrate the link juice if it was all on one page. Suppose you have a lot of comments. Is it better to move comments to a second page at a certain point? Sometimes the comments are not super focused on the topic of the page compared to the main text.
Technical SEO | | ProjectLabs1 -
Is "last modified" time in XML Sitemaps important?
My Tech lead is concerned that his use of a script to generate XML sitemaps for some client sites may be causing negative issues for those sites. His concern centers around the fact that the script generates a sitemap which indicates that every URL page in the site was last modified at the exact same date and time. I have never heard anything to indicate that this might be a problem, but I do know that the sitemaps I generate for other client sites can choose server response or not. What is the best way to generate the sitemap? Last mod from actual time modified, or all set at one date and time?
Technical SEO | | ShaMenz0