Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Mass URL changes and redirecting those old URLS to the new. What is SEO Risk and best practices?
-
Hello good people of the MOZ community,
I am looking to do a mass edit of URLS on content pages within our sites. The way these were initially setup was to be unique by having the date in the URL which was a few years ago and can make evergreen content now seem dated. The new URLS would follow a better folder path style naming convention and would be way better URLS overall.
Some examples of the **old **URLS would be
https://www.inlineskates.com/Buying-Guide-for-Inline-Skates/buying-guide-9-17-2012,default,pg.html
https://www.inlineskates.com/Buying-Guide-for-Kids-Inline-Skates/buying-guide-11-13-2012,default,pg.html
https://www.inlineskates.com/Buying-Guide-for-Inline-Hockey-Skates/buying-guide-9-3-2012,default,pg.html
https://www.inlineskates.com/Buying-Guide-for-Aggressive-Skates/buying-guide-7-19-2012,default,pg.htmlThe new URLS would look like this which would be a great improvement
https://www.inlineskates.com/Learn/Buying-Guide-for-Inline-Skates,default,pg.html
https://www.inlineskates.com/Learn/Buying-Guide-for-Kids-Inline-Skates,default,pg.html
https://www.inlineskates.com/Learn/Buying-Guide-for-Inline-Hockey-Skates,default,pg.html
https://www.inlineskates.com/Learn/Buying-Guide-for-Aggressive-Skates,default,pg.htmlMy worry is that we do rank fairly well organically for some of the content and don't want to anger the google machine.
The way I would be doing the process would be to edit the URLS to the new layout, then do the redirect for them and push live.
Is there a great SEO risk to doing this?
Is there a way to do a mass "Fetch as googlebot" to reindex these if I do say 50 a day? I only see the ability to do 1 URL at a time in the webmaster backend.
Is there anything else I am missing?I believe this change would overall be good in the long run but do not want to take a huge hit initially by doing something incorrectly. This would be done on 5- to a couple hundred links across various sites I manage.
Thanks in advance,
Chris Gorski -
Hey K.
Happy you found value in everyone's response here.
For your URL, I see your structure resemble a blog structure with your subfolder being "learn" instead of "blog". So if it is, your final URL that you described is fine (learn/buying-guide-for-inline-skates).
As a counterpoint, if you had this content as a hub page (some form of content pillar/topic cluster) for example, it would've been a good possibility to just change the URL structure since you have many Buying Guides. Different types of content, different ways to put it on your site.
Like so: /buying-guide/inline-skatesAt the end of the day, the structure needs to be logical and reflective of where your content is. I think you got it right anyway.
For the execution part;
- I would not recommend using the "crawl as Googlebot" function in search console. It would be way too time-consuming for you, and it is not really designed for that kind of work.
- Instead, update your sitemap with the final URLs and send it again via SearchConsole / Bing Webmastertools.
- Also, don't forget to go ahead and change the internal links pointing to the old URL to point directly to the new ones or else you'll just have a bunch of 301's crawled by Google. Make it seamless.
- Monitor. Monitor. Monitor.
Hope that helped!
-
Thanks all the for responses and I am taking to heart all your suggestions. I will make all the URLs lowercase as this is something I didn't take into account.
Due to site hosting platform limitation for the URL structure I need to have a value in the bold areas below:
https://www.inlineskates.com/learn/buying-guide-for-inline-skates,default,pg.html
The first value "learn" doesn't have to be unique but the 2nd "learn/buying-guide-for-inline-skates" is the driver for the URL and must be unique, so the short URL like https://www.inlineskates.com/buying-guide/Inline-Skates,default,pg.html wouldn't work since I would only be able to use it 1x for inline skates and have several content pieces that I would need to add that to, size guides, charts, etc.
My main concern is about the process of doing the redirects, say I do 50 in one day, what is my next step? Is there a way to run the fetch as googlebot to a handful of pages as I only see the ability to add 1 URL at a time in the webmaster backend. If I go ahead and do this I just want to do it in the smartest way possible.
Thanks, Chris
-
And to add to that, Moz has a great resource on how to write them (emphasis mine);
-
Keeping URLs as simple, relevant, compelling, and accurate as possible is key to getting both your users and search engines to understand them.
-
URLs should be definitive but concise.
-
When necessary for readability, use hyphens to separate words. URLs should not use underscores, spaces, or any other characters to separate words.
-
Use lowercase letters. In some cases, uppercase letters can cause issues with duplicate pages.
-
Avoid the use of URL parameters, if possible, as they can create issues with tracking and duplicate content. If parameters need to be used (UTM codes, e.g.), use them sparingly.
-
-
The new URLs seem pretty verbose.
If this was my site, I would consider...
https://www.inlineskates.com/Buying-Guide/Inline-Skates/
https://www.inlineskates.com/Buying-Guide/Kids-Inline-Skates/
https://www.inlineskates.com/Buying-Guide/Inline-Hockey-Skates/
https://www.inlineskates.com/Buying-Guide/Aggressive-Skates/My breadcrumbs would look similar....
Home >> Buying Guide >> Inline Skates
-
I've heard recent claims that you no longer lose link juice when 301ing, but I've also had personal experience with taking a hit to organic traffic when doing this type of major URL restructuring. You're right, in the long run, it does pay off, but you're also right that there may be a ding in the short term.
While you're in there, I would highly, highly recommend taking the additional step of removing ",default,pg" from your URLs. There are two problems with that string - one is it contains commas, which can be problematic. Anything other than plain text and hyphens is generally discouraged in URLs. The other problem is that "default pg" part is watering down your URL and does not contribute to making the links short and easy to type and understand for humans. Since you are undertaking such a massive restructuring, now is the time to make these additional tweaks - as it's less painful to do All The Things at once and take one hit, than it would be to take a hit now and another hit later when you tackle that separately. I personally also always use only lowercase characters in my URLs as it's easier to type and depending on the server OS improper capitalization may make it so they don't reach the right page, but that is more of a UX preference than an SEO enhancement.
Another couple of tips: I find it extremely helpful to map everything out on a spreadsheet in this type of migration. Helps me make sure I have the old and the new mapped out, and also helps to have a checklist to go through systematically. I also tend to group related pages in batches, as presumably related pages will be linked to each other, and once Google comes crawling back over one page it will see the new URLs and crawl those too, faster than it would crawl unrelated pages. Finally, use a HTTP header check tool (google that and you'll find several) after you add the redirects just to make 100% certain you've set everything up correctly. And where possible, mass-redirect in a single .htaccess rule rather than one by one.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
SEO Best Practices regarding Robots.txt disallow
I cannot find hard and fast direction about the following issue: It looks like the Robots.txt file on my server has been set up to disallow "account" and "search" pages within my site, so I am receiving warnings from the Google Search console that URLs are being blocked by Robots.txt. (Disallow: /Account/ and Disallow: /?search=). Do you recommend unblocking these URLs? I'm getting a warning that over 18,000 Urls are blocked by robots.txt. ("Sitemap contains urls which are blocked by robots.txt"). Seems that I wouldn't want that many urls blocked. ? Thank you!!
Intermediate & Advanced SEO | | jamiegriz0 -
Best practice for deindexing large quantities of pages
We are trying to deindex a large quantity of pages on our site and want to know what the best practice for doing that is. For reference, the reason we are looking for methods that could help us speed it up is we have about 500,000 URLs that we want deindexed because of mis-formatted HTML code and google indexed them much faster than it is taking to unindex them unfortunately. We don't want to risk clogging up our limited crawl log/budget by submitting a sitemap of URLs that have "noindex" on them as a hack for deindexing. Although theoretically that should work, we are looking for white hat methods that are faster than "being patient and waiting it out", since that would likely take months if not years with Google's current crawl rate of our site.
Intermediate & Advanced SEO | | teddef0 -
New Site (redesign) Launched Without 301 Redirects to New Pages - Too Late to Add Redirects?
We recently launched a redesign/redevelopment of a site but failed to put 301 redirects in place for the old URL's. It's been about 2 months. Is it too late to even bother worrying about it at this point? The site has seen a notable decrease in site traffic/visits, perhaps due to this issue. I assume that once the search engines get an error on a URL, it will remove it from displaying in search results after a period of time. I'm just not sure if they will try to re-crawl those old URLs at some point and if so, it may be worth it to have those 301 redirects in place. Thank you.
Intermediate & Advanced SEO | | BrandBuilder0 -
Images with a token in the url, in Drupal. How does it affect to SEO?
Hi everyone! I am checking now a website that works with Drupal, and I found that images have urls like this... http://www.brandname.com/sites/default/files/styles/directory_xyz/public/name-of-the-picture.png?itok=T89RpzrK I was wondering how an URL like that with the token at the and, can affect to SEO. I cound't find anything. Anyone knows? Thank you!
Intermediate & Advanced SEO | | teconsite0 -
Redirecting to a new domain... a second time
Hi all, I help run a website for a history-themed podcast and we just moved it to its second domain in 7 years. We've had very good SEO up until last week, and I'm wondering if I screwed up the way I redirected the domains. It's like this: Originally the site was hosted at "first.com", and it acquired inbound links. However, we then started to host the site on blogger, so we... Redirected the site to "second.blogspot.com". (Thus, 1 --> 2) It stayed here for about 7 years and got lots of traffic. Two weeks ago we moved it off of blogger and into Wordpress, so we 301 redirected everything to... third.com. (Thus, 1 --> 2 --> 3) The redirects worked, and when we Google individual posts, we are now seeing them in Google's index at the new URL. My question: What about the 1--> 2 redirect? There are still lots of links pointing to "first.com". Last week I went into my GoDaddy settings and changed the first redirect, so that first.com now points to third.com. (Thus 1 --> 3, and 2-->3) I was correct in doing that, right? The drop in Google traffic I've seen this past week makes me think that maybe I screwed something up. Should we have kept 1 --> 2 --> 3? (Again, now we have 1-->3 and 2-->3) Thanks for any insights on this! Tom
Intermediate & Advanced SEO | | TomNYC1 -
Can you redirect specific sub domain URLs?
ello! We host our PDFs, Images, CSS all in a sub domain. For the question, let's call this sub.cyto.com. I've noticed a particular PDF doing really well, infact it has gathered valuable external links from high authoritative sites. To top it off, it gets good visits. I've been going back and forth with our developers to move this PDF to a subfolder structure.
Intermediate & Advanced SEO | | Bio-RadAbs
For example: www.cyto.com/document/xxxx.pdf In my perspective, if I move this and set up a permanent redirect, then all the external links the PDF gathered, link juice and future visits will be attributed to the main website. Since the PDF is existing in the subdomain, I can't even track direct visits nor get the link juice. It appears in top position of Google as well. My developer says it is better to keep images, pdf, css in the subdomain. I see his point and an idea I have is to: convert the pdf to a webpage. Set up a 301 redirect from the existing subdomain to this webpage Upload the pdf with a new name and link to it from the webpage, so users can download if they choose to. This should give me the existing rank juice. However, my question is whether you can set up a 301 redirect for just a single subdomain URL to a folder structure URL? sub.cyto.com/xxx.pdf to www.cyto.com/document/xxxx.pdf?0 -
ECommerce product listed in multiple places, best SEO practice?
We have an eCommerce site we have built for a customer and the products are allowed to appear in more than one product category within the web site. Now I know this is a bad idea from a duplicate content point of view, But we are going to allow the customer to select which out of the multiple categories the product appears in will be the default category. This will mean we will have a way of defining what the default url is for a product. So am I correct in thinking all the other urls where the product appears we should add a rel canonical to these pages pointing to the default url to stop duplicate content? Is this the best way?
Intermediate & Advanced SEO | | spiralsites0 -
Best approach to launch a new site with new urls - same domain
www.sierratradingpost.com We have a high volume e-commerce website with over 15K items, an average of 150K visits per day and 12.6 pages per visit. We are launching a new website this spring which is currently on a beta sub domain and we are looking for the best strategy that preserves our current search rankings while throttling traffic (possibly 25% per week) to measure results. The new site will be soft launched as we plan to slowly migrate traffic to it via a load balancer. This way we can monitor performance of the new site while still having the old site as a backup. Only when we are fully comfortable with the new site will we submit the 301 redirects and migrate everyone over to the new site. We will have a month or so of running both sites. Except for the homepage the URL structure for the new site is different than the old site. What is our best strategy so we don’t lose ranking on the old site and start earning ranking on the new site, while avoiding duplicate content and cloaking issues? Here is what we got back from a Google post which may highlight our concerns better: http://www.google.com/support/forum/p/Webmasters/thread?tid=62d0a16c4702a17d&hl=en&fid=62d0a16c4702a17d00049b67b51500a6 Thank You, sincerely, Stephan Woo Cude SEO Specialist [email protected]
Intermediate & Advanced SEO | | STPseo0