Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Moving from a .com to .co.uk
-
I need to migrate a wordpress site from domainname.com to domainname.co.uk.
If I just put a 301 on every page on the .com will that cover it?
Would it make sense to go and change all the backlinks/profile links to the new .co.uk site or doesn't it matter if you have a 301 redirect on it?
Thanks
-
Thank you, very much appreciate you responses.
-
Hi Joanne,
We did the same thing recently for a client who wanted to move from a .com to a .co.uk. She had a well-established website. Our methodology:
- Create exactly the same URL structure and architecture.
- Copy the exact same content over to each URL
- Notify all websites linking to any pages to update (We actually did this part well in advance).
- Notified Google via its webmaster tools that we are migrating to a new website. You can use their new site section to notify them.
- 301 redirected every page of the old site to its new equivalent page.
The sit back and wait, we found the changeover within the SERPS fairly fast – there were a few hiccups while waiting for Google to fully update, however it worked and all PageRank now on site and older positions are almost all regained.
Hope that helps.
-
Hi Joanne,
When taking such a decision I always read through again and again the article from Google: https://support.google.com/webmasters/bin/answer.py?hl=en&answer=83105
Ideally it would be good to contact each webmaster and ask them to change the url from the .com to the .co.uk, but realistic that is almost impossible.
What I would do is download through OpenSiteExplorer all the incoming links, check in general which are the most important for me and at least contact those webmasters and tell them about the change of address.
I hope it helps.
Istvan
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
GoogleBot still crawling HTTP/1.1 years after website moved to HTTP/2
Whole website moved to https://www. HTTP/2 version 3 years ago. When we review log files, it is clear that - for the home page - GoogleBot continues to only access via HTTP/1.1 protocol Robots file is correct (simply allowing all and referring to https://www. sitemap Sitemap is referencing https://www. pages including homepage Hosting provider has confirmed server is correctly configured to support HTTP/2 and provided evidence of accessing via HTTP/2 working 301 redirects set up for non-secure and non-www versions of website all to https://www. version Not using a CDN or proxy GSC reports home page as correctly indexed (with https://www. version canonicalised) but does still have the non-secure version of website as the referring page in the Discovery section. GSC also reports homepage as being crawled every day or so. Totally understand it can take time to update index, but we are at a complete loss to understand why GoogleBot continues to only go through HTTP/1.1 version not 2 Possibly related issue - and of course what is causing concern - is that new pages of site seem to index and perform well in SERP ... except home page. This never makes it to page 1 (other than for brand name) despite rating multiples higher in terms of content, speed etc than other pages which still get indexed in preference to home page. Any thoughts, further tests, ideas, direction or anything will be much appreciated!
Technical SEO | | AKCAC1 -
I want to move some pages of my website to a folder and nav menu in those pages should only show inner page links, will it hurt SEO?
Hi, My website has a few SaaS products, to make my website simple i want to move my website some pages to its specific folder structure , so eg website.com/product1/features
Technical SEO | | webbeemoz
website.com/product1/pricing
website.com/product1/information and same for product2 and so on, the website.com/product1/.. menu will only show the links of product1 and only one link to homepage (possibly in footer). Please share your opinion will it be a good idea, from UI perspective it will be simple , but i am not sure about SEO perspective, please help thanks1 -
Moving E-Commerce Store to Subdomain?
Hi all, We have a customer who currently uses Square for their in-store point-of-sale system as well as for their e-commerce website. From my understanding, a Square site is a watered-down version of Weebly, and is proving to be highly restrictive from an SEO and content structuring standpoint. It's been an uphill battle to try and get traction for their site in SERPs. Would it be a bad idea to move the entire Square online store to a subdomain, and install WordPress on the root domain? This way their online store would remain as-is, but the primary pages on the site would be on WordPress which would give us a lot more control over the content. I just want to make sure this doesn't negatively impact their SEO. Thanks!
Technical SEO | | suarezventures0 -
Move a Wordpress Site to HTTPS with Bluehost
HI Guys, do you think that the following guide is enoght to move a bluehost wordpress site to https in a seo best practive way? https://www.shoutmeloud.com/free-ssl-certificate-bluehost-hosting.html Basically their steps are: Install SSL on Bluehost panel Install Really Simple SSL Wp Plugin Edit Your .htacess File & Add The Code For HTTP To HTTPS Redirection Update All HTTP URLs In Database To HTTPS Using Search and Replace Plugin Use Broken Link Checker plugin & use its redirection module to find links to 3rd party sites with HTTP that should now be HTTPS. Last thing to do Submit your new HTTPS site to Google Search Console & submit your sitemap. Update your profile link on Google Analytics. Update your website links on social media profiles & anywhere else they exist. This step you can do in pieces in the coming days. Read this guide to learn more about HTTP to HTTPS migration & fixing mixed content. If you disabled Who.Is guard for your domain name, you can enable it now. Do you know a better practical guide for wordrpess? in term of usefull plugins to handle the migration? Tx to everyone!
Technical SEO | | Dreamrealemedia0 -
How well do .ltd.uk domain names rank?
Hi all, do .ltd.uk domain names rank well? How well do they stack up against a .com or .co.uk? I'd really appreciate any feedback from people with any experience with these domains. Thanks John
Technical SEO | | john251810100 -
.ca and. com domains
Hello, currently the main site im working on is a .com, but have the .ca version purchased from register.com. should i have this setup to redirect to the .com site. will google see these as dup content. We have the .ca for our canadian customers but both sites are identical. Thank you
Technical SEO | | TP_Marketing0 -
.biz and .xxx and the latest and greatest.co
We are designing a .BIZ site where we will push a high percentage of the traffic ourselves. In other words, we aren't relying heavily on SEO for being found. I have a two part question: 1. Is there really any reason for a young start up .biz company to buy a $100 / year .XXX extension? 2. We don't have the .com (and can't afford it at the moment) but we do have the .co Does the .co have any upside to it that the .biz does not in SER?And do we get penalized for using the .biz? In other words, do SEs prefer the .com and nothing else will do quite as well? Thanks!
Technical SEO | | webindustry0 -
What is the best method to block a sub-domain, e.g. staging.domain.com/ from getting indexed?
Now that Google considers subdomains as part of the TLD I'm a little leery of testing robots.txt with something like: staging.domain.com
Technical SEO | | fthead9
User-agent: *
Disallow: / in fear it might get the www.domain.com blocked as well. Has anyone had any success using robots.txt to block sub-domains? I know I could add a meta robots tag to the staging.domain.com pages but that would require a lot more work.0