Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Creating 100,000's of pages, good or bad idea
-
Hi Folks,
Over the last 10 months we have focused on quality pages but have been frustrated with competition websites out ranking us because they have bigger sites. Should we focus on the long tail again?
One option for us is to take every town across the UK and create pages using our activities. e.g.
Stirling
Stirling paintball
Stirling Go Karting
Stirling Clay shootingWe are not going to link to these pages directly from our main menus but from the site map.
These pages would then show activities that were in a 50 mile radius of the towns. At the moment we have have focused our efforts on Regions, e.g. Paintball Scotland, Paintball Yorkshire focusing all the internal link juice to these regional pages, but we don't rank high for towns that the activity sites are close to.
With 45,000 towns and 250 activities we could create over a million pages which seems very excessive! Would creating 500,000 of these types of pages damage our site? This is my main worry, or would it make our site rank even higher for the tougher keywords and also get lots of traffic from the long tail like we used to get.
Is there a limit to how big a site should be?
-
Hi Mark!
Thanks for asking this good question. While there is no limit to how big a website can be, I think you can see from the general response here that most members would encourage you to stick to manually developing quality pages rather than automating hundreds of thousands of pages, solely for ranking purposes. I second this advice.
Now, I would like to clarify your business model. Are you a physical, actual business that customers come to, either to buy paintball equipment or to play paintball in a gallery? Or, is your business virtual, with no in person transactions? I'm not quite understanding this from your description.
If the former, I would certainly encourage you to develop a very strong, unique page for each of your physical locations. If you have 10 locations (with unique street addresses and phone numbers), then that would be 10 pages. If you've got 20 locations, that would be 20 pages, etc. But don't approach these with a 'just switch out the city name in the title tags' mindset. Make these pages as exceptional as possible. Tell stories, show off testimonials, share pictures and videos, entertain, educate, inspire. These city landing pages will be intimately linked into your whole Local SEM campaign, provided they each represent a business location with a unique dedicated street address and unique local area code phone number.
But, if you are considering simply building a page for every city in the UK, I just can't see justification for doing so. Ask yourself - what is the value?
There are business models (such as carpet cleaners, chimney sweeps, general contractors, etc.) that go to their clients' locations to serve and for which I would be advising that they create city landing pages for each of their service cities, but this would be extremely regional...not statewide or national or International. A carpet cleaner might serve 15 different towns and cities in his region, and I would encourage him to start gathering project notes and testimonials, videos and photos to begin developing a body of content important enough for him to start creating strong, interesting and unique pages for each of these cities. But I've also had local business owners tell me they want to cover every city in California, for instance, because they think it will help them to do so, and I discourage this.
Even if the business is virtual and doesn't have any in-person transactions with clients or physical locations, I would still discourage this blanketing-the-whole-nation-with-pages approach. A national retailer needs to build up its brand so that it becomes known and visible organically for its products rather than your theoretical approach of targeting every city in the nation. In short order, the mindset behind that approach just doesn't make good horse sense.
And, as others have stated, adding thousands of thin, potentially duplicate pages to any site could definitely have a very negative effect on rankings.
My advice is to make the time to start developing a content strategy for cities in which you have a legitimate presence. If budget means you can't hire a copywriter to help you with this and to speed up the work, accept that this project deserves all the time you can give it and that a slow development of exceptional pages is better than a fast automation of poor quality pages.
Hope this helps!
-
Hi Mark,
If A,C, and E's page is similar to B,D, and F's page it is still consider dupllicate content. Based on Webmaster's definiton:
"Duplicate content generally refers to substantive blocks of content within or across domains that either completely match other content or are appreciably similar"
Each of your pages should be unique and different from other pages.
I suggest you to continue providing quality content and target the long tail keywords. That alone will help you generate more traffic. Furthermore, out ranking is not a problem. You should focus on getting to the frist page (providing quality content with long tail or regular keywords) and when you are on the first page, try to get searchers to click on your link using Title tag and Meta descriptions.
Out ranking just means they are ranked 4th and you are ranked 5th, 6th but as long as you have a better title tag and meta description. I believe searchers will click on the more attractive results.
-
Cookie cutter pages like these stopped working in Google about ten years ago.
If you toss them up I think that your entire site will tank.
I would go back to focusing on quality pages.
-
If the user experience awesome, and people are staying on your site and looking around, great. If you think the 100,000 pages will make search engines love you, machines can never provide the love users can give you.
-
Can you mix content up from your website e.g. paintball site A, C and E on one page and B,D and F on another if the towns are close together? What I'm not sure about is how different in % terms the content actually has to be.
If we have less written content then the amounts of words we have to actually change would be much less.
The challenge we have is we have build the site this time with filtering in mind, so rather than making customers navigate we allow them to be able to search which is much better in terms of getting the activities they want. The downside is now our site does not show for the long tail as we reduced the pages massively.
-
so we dont have the resources if we did it manually but what would happen is the content would be different on each page as we would only show activity sites within a 50 miles radius. And we would make certain text, h1 etc different and relate to the town.
Below are some examples of sites I see doing well ie number 1 using this method
Our content would be much better than say http://www.justpaintballDOTcoDOTuk/site_guide/Aberfeldy.htm or http://www.goballisticDOTcoDOTuk/paintball_in_/ABERFELDY.asp
But as you say getting this wrong is my worry.
-
Hi Mark,
Creating 100,000 pages is definitely good for Search Engine because you have a lot more contents for them to crawl and have more chances your pages might show up on related keywords. However, the problem is do you have enough unique contents you can post on all those 100,000 pages. If you use similar content, I am afraid it will be duplicate contents. You may think changing up the town names will be enough but it might be risky.
If you can create 100,000 unique contents, Sure go ahead. If not, don't take the risk of duplicate contents.
-
Do you have the resources to create unique content for all those pages? Because adding 500,000 pages of duplicate content will absolutely damage your site.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Should you 'noindex' Checkout Pages?
Today I was reviewing my Moz analytics and suddenly noticed 1,000 issues with pages without a meta description. I reviewed the list and learned it is 1,000 checkout pages. That's because my website has thousands of agency pages from which you can buy a product, and it reflects that difference on each version of the checkout. So, I was thinking about no-indexing (but continuing to 'follow') these checkout pages, but wondering if it has any knock-on effects I may be unaware of? Any assistance is much appreciated. Luke
Intermediate & Advanced SEO | | Luke_Proctor0 -
What's the best way to noindex pages but still keep backlinks equity?
Hello everyone, Maybe it is a stupid question, but I ask to the experts... What's the best way to noindex pages but still keep backlinks equity from those noindexed pages? For example, let's say I have many pages that look similar to a "main" page which I solely want to appear on Google, so I want to noindex all pages with the exception of that "main" page... but, what if I also want to transfer any possible link equity present on the noindexed pages to the main page? The only solution I have thought is to add a canonical tag pointing to the main page on those noindexed pages... but will that work or cause wreak havoc in some way?
Intermediate & Advanced SEO | | fablau3 -
Will disallowing URL's in the robots.txt file stop those URL's being indexed by Google
I found a lot of duplicate title tags showing in Google Webmaster Tools. When I visited the URL's that these duplicates belonged to, I found that they were just images from a gallery that we didn't particularly want Google to index. There is no benefit to the end user in these image pages being indexed in Google. Our developer has told us that these urls are created by a module and are not "real" pages in the CMS. They would like to add the following to our robots.txt file Disallow: /catalog/product/gallery/ QUESTION: If the these pages are already indexed by Google, will this adjustment to the robots.txt file help to remove the pages from the index? We don't want these pages to be found.
Intermediate & Advanced SEO | | andyheath0 -
How do I get rel='canonical' to eliminate the trailing slash on my home page??
I have been searching high and low. Please help if you can, and thank you if you spend the time reading this. I think this issue may be affecting most pages. SUMMARY: I want to eliminate the trailing slash that is appended to my website. SPECIFIC ISSUE: I want www.threewaystoharems.com to showing up to users and search engines without the trailing slash but try as I might it shows up like www.threewaystoharems.com/ which is the canonical link. WHY? and I'm concerned my back-links to the link without the trailing slash will not be recognized but most people are going to backlink me without a trailing slash. I don't want to loose linkjuice from the people and the search engines not being in consensus about what my page address is. THINGS I"VE TRIED: (1) I've gone in my wordpress settings under permalinks and tried to specify no trailing slash. I can do this here but not for the home page. (2) I've tried using the SEO by yoast to set the canonical page. This would work if I had a static front page, but my front page is of blog posts and so there is no advanced page settings to set the canonical tag. (3) I'd like to just find the source code of the home page, but because it is CSS, I don't know where to find the reference. I have gone into the css files of my wordpress theme looking in header and index and everywhere else looking for a specification of what the canonical page is. I am not able to find it. I'm thinking it is actually specified in the .htaccess file. (4) Went into cpanel file manager looking for files that contain Canonical. I only found a file called canonical.php . the only thing that seemed like it was worth changing was changing line 139 from $redirect_url = home_url('/'); to $redirect_url = home_url(''); nothing happened. I'm thinking it is actually specified in the .htaccess file. (5) I have gone through the .htaccess file and put thes 4 lines at the top (didn't redirect or create the proper canonical link) and then at the bottom of the file (also didn't redirect or create the proper canonical link) : RewriteEngine on
Intermediate & Advanced SEO | | Dillman
RewriteCond %{HTTP_HOST} ^([a-z.]+)?threewaystoharems.com$ [NC]
RewriteCond %{HTTP_HOST} !^www. [NC]
RewriteRule .? http://www.%1threewaystoharems.com%{REQUEST_URI} [R=301,L] Please help friends.0 -
Remove URLs that 301 Redirect from Google's Index
I'm working with a client who has 301 redirected thousands of URLs from their primary subdomain to a new subdomain (these are unimportant pages with regards to link equity). These URLs are still appearing in Google's results under the primary domain, rather than the new subdomain. This is problematic because it's creating an artificial index bloat issue. These URLs make up over 90% of the URLs indexed. My experience has been that URLs that have been 301 redirected are removed from the index over time and replaced by the new destination URL. But it has been several months, close to a year even, and they're still in the index. Any recommendations on how to speed up the process of removing the 301 redirected URLs from Google's index? Will Google, or any search engine for that matter, process a noindex meta tag if the URL's been redirected?
Intermediate & Advanced SEO | | trung.ngo0 -
How to check a website's architecture?
Hello everyone, I am an SEO analyst - a good one - but I am weak in technical aspects. I do not know any programming and only a little HTML. I know this is a major weakness for an SEO so my first request to you all is to guide me how to learn HTML and some basic PHP programming. Secondly... about the topic of this particular question - I know that a website should have a flat architecture... but I do not know how to find out if a website's architecture is flat or not, good or bad. Please help me out on this... I would be obliged. Eagerly awaiting your responses, BEst Regards, Talha
Intermediate & Advanced SEO | | MTalhaImtiaz0 -
There's a website I'm working with that has a .php extension. All the pages do. What's the best practice to remove the .php extension across all pages?
Client wishes to drop the .php extension on all their pages (they've got around 2k pages). I assured them that wasn't necessary. However, in the event that I do end up doing this what's the best practices way (and easiest way) to do this? This is also a WordPress site. Thanks.
Intermediate & Advanced SEO | | digisavvy0 -
Xml sitemap advice for website with over 100,000 articles
Hi, I have read numerous articles that support submitting multiple XML sitemaps for websites that have thousands of articles... in our case we have over 100,000. So, I was thinking I should submit one sitemap for each news category. My question is how many page levels should each sitemap instruct the spiders to go? Would it not be enough to just submit the top level URL for each category and then let the spiders follow the rest of the links organically? So, if I have 12 categories the total number of URL´s will be 12??? If this is true, how do you suggest handling or home page, where the latest articles are displayed regardless of their category... so I.E. the spiders will find l links to a given article both on the home page and in the category it belongs to. We are using canonical tags. Thanks, Jarrett
Intermediate & Advanced SEO | | jarrett.mackay0