Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Best way to noindex long dynamic urls?
- 
					
					
					
					
 I just got a Mozcrawl back and see lots of errors for overly dynamic urls. The site is a villa rental site that gives users the ability to search by bedroom, amenities, price, etc, so I'm wondering what the best way to keep these types of dynamically generated pages with urls like /property-search-page/?location=any&status=any&type=any&bedrooms=9&bathrooms=any&min-price=any&max-price=any from indexing. Any assistance will be greatly appreciated : ) 
- 
					
					
					
					
 If you have a page that lists all the villas outside the search results, then you don't lose anything by blocking that folder on the robots.txt But still, somebody, the guy that wrote the custom theme knows how to do the changes needed. If you want I can help you with it, for free  Just PM me (I'll need FTP access). Just PM me (I'll need FTP access).
- 
					
					
					
					
 Having some trouble... Because the site is Wordpress, which dynamically generates pages, there is no /property-search-page/ nor is there a property-search-page.php in the editor files, so the only option I have is to put disallow: /property-search-page/ in the robots.txt file, correct? 
- 
					
					
					
					
 You guys rock! I'll try these out tomorrow. Thanks a million. 
- 
					
					
					
					
 If you have a /all-villas/ page then you should go ahead and noindex the search results as Google Guidelines suggests. You can either do it in the /property-search-page/ or using the robots.txt file. In the robots.txt, add: disallow: /property-search-page/ The robots method guarantees that no page inside that folder is indexed or even crawled (including /property-search-page/?whatever). Or on the page /property-search-page/ you can add the meta noindex as such: Then check if that meta tag is shown in all search results (just check a couple of them). Hope that works! 
- 
					
					
					
					
 Yes, it will. Also looks like custom code, it depends on how the header is coded. But it should work. Test it, if you can. This should solve your problems relatively easily. If nothing works, you can always do a robots.txt deny for /property-search-page/?* pages, but that's not a recommended solution. Try the canonical way to see if it works first. 
- 
					
					
					
					
 We already have Yoast installed, but the errors are still showing up in the Moz report. To clarify, let's assume we have another page that lists all the villas (/all-villas/). If I go to the property-search-page php file and canonical=rel it to /all-villas/, will it canonical=rel all /property-search-page/?whatever pages to the /all-villas/ page? 
- 
					
					
					
					
 Well, that will make a little easier from one side and harder from the other. You can try installing SEO by Yoast, that will put all the canonical tags for you, however, I think it won't link the search result pages to the canonical page that lists them all. That might require a little coding. If there's another page, outside /property-search-page/ folder that lists all villas, then you can disallow that folder in the robots.txt file, and that should fix it. If there isn't, well, then you will need to edit the /property-search-page/ page to use a static canonical tag that points to the page that lists all the villas removing any kind of filtering. Hope that helps! 
- 
					
					
					
					
 Thanks for the response. The site is Wordpress - is there an easy way to write some sort of rule that would canonical any of these types of pages to a category page? How would you go about doing that? 
- 
					
					
					
					
 Thanks for the response. The site is Wordpress - is there an easy way to write some sort of rule that would canonical any of these types of pages to a category page? How would you go about doing that? 
- 
					
					
					
					
 I agree with Federico one hundred percent. Figure out what your primary SEO friendly URLs are for these kinds of pages and canonical them back to that page. 
- 
					
					
					
					
 I wouldn't put a noindex meta on them, instead I would consider using a canonical tag pointing to the page that lists all the villas. Anyway, what programming language are you using? 
Browse Questions
Explore more categories
- 
		
		Moz ToolsChat with the community about the Moz tools. 
- 
		
		SEO TacticsDiscuss the SEO process with fellow marketers 
- 
		
		CommunityDiscuss industry events, jobs, and news! 
- 
		
		Digital MarketingChat about tactics outside of SEO 
- 
		
		Research & TrendsDive into research and trends in the search industry. 
- 
		
		SupportConnect on product support and feature requests. 
Related Questions
- 
		
		
		
		
		
		Best way to change URL for already ranking pages
 Hello. I have a lot of pages that I'm optimising. The ones I'm focusing on right now is already ranking, but the URLs could be better (they don't include the keywords right now). However I'm worried that if I change the URLs they will drop in rankings or have to start over. I would of course set up 301 redirect, but is there more I need to do? What is the best way to change URL for already ranking pages? Technical SEO | | GoMentor0
- 
		
		
		
		
		
		Url folder structure
 I work for a travel site and we have pages for properties in destinations and am trying to decide how best to organize the URLs basically we have our main domain, resort pages and we'll also have articles about each resort so the URL structure will actually get longer: Technical SEO | | Vacatia_SEO
 A. domain.com/main-keyword/state/city-region/resort-name
 _ domain.com/family-condo-for-rent/orlando-florida/liki-tiki-village_ _ domain.com/main-keyword-in-state-city/resort-name-feature _
 _ domain.com/family-condo-for-rent/orlando-florida/liki-tiki-village/kid-friend-pool_ B. Another way to structure would be to remove the location and keyword folders and combine. Note that some of the resort names are long and spaces are being replaced dynamically with dashes.
 ex. domain.com/main-keyword-in-state-city/resort-name
 _ domain.com/family-condo-for-rent-in-orlando-florida/liki-tiki-village_ _ domain.com/main-keyword-in-state-city/resort-name-feature_
 _ domain.com/family-condo-for-rent-in-orlando-florida/liki-tiki-village-kid-friend-pool_ Question: is that too many folders or should i combine or break up? What would you do with this? Trying to avoid too many dashes.0
- 
		
		
		
		
		
		How Does Dynamic Content for a Specific URL Impact SEO?
 Example URL: http://www.sja.ca/English/Community-Services/Pages/Therapy Dog Services/default.aspx The above page is generated dynamically depending on what province the visitor visits from. For example, a visitor from BC would see something quite different than a visitor from Nova Scotia; the intent is that the information shown should be relevant to the user of that province. How does this effect SEO? How (or from what location) does Googlebot decide to crawl the page? I have considered a subdirectory for each province, though that comes with its challenges as well. One such challenge is duplicate content when different provinces may have the same information for some pages. Any suggestions for this? Technical SEO | | ey_sja0
- 
		
		
		
		
		
		Redirect URLS with 301 twice
 Hello, I had asked my client to ask her web developer to move to a more simplified URL structure. There was a folder called "home" after the root which served no purpose. I asked for the URLs to be redirected using 301 to the new URLs which did not have this structure. However, the web developer didn't agree and decided to just rename the "home" folder "p". I don't know why he did this. We argued the case and he then created the URL structure we wanted. Initially he had 301 redirected the old URLS (the one with "Home") to his new version (the one with the "p"). When we asked for the more simplified URL after arguing, he just redirected all the "p" URLS to the PAGE NOT FOUND. However, remember, all the original URLs are now being redirected to the PAGE NOT FOUND as a result. The problems I see are these unless he redirects again: The new simplified URLS have to start from scratch to rank 2)We have duplicated content - two URLs with the same content Customers clicking products in the SERPs will currently find that they are being redirect to the 404 page. I understand that redirection has to occur but my questions are these: Is it ok to redirect twice with 301 - so old URL to the "p" version then to final simplified version. Will link juice be lost doing this twice? If he redirects from the original URLS to the final version missing out the "p" version, what should happen to the "p" version - they are currently indexed. Any help would be appreciated. Thanks Technical SEO | | AL123al0
- 
		
		
		
		
		
		Noindex user profile
 I have a social networking site with user- and company profiles. Some profiles have little to no content. One of the users here at moz suggested noindex-ing these profiles. I am still investigating this issue and have some follow up questions: What is the possible gain of no-indexing uninteresting profiles? Especially interested in this since these profiles do bring in long-tail traffic atm. How "irreversable" is introducing a noindex directive? Would everything "return to normal" if I remove te noindex directive? When determining the treshold for having profiles indexed, how should the following items be weighed Sum of number of words on the page (comprised of one or more of the following: full name, city, 0 to N company names, bio, activity) (unique) Profile picture (Nofollowed) Links to user's profiles on social networks or user's own site. Embedded Google Map Thanks! Technical SEO | | thomasvanderkleij0
- 
		
		
		
		
		
		Best Way To Clean Up Unruly SubDomain?
 Hi, I have several subdomains that present no real SEO value, but are being indexed. They don't earn any backlinks either. What's the best way of cleaning them up? I was thinking the following: 1. Verify them all in Webmaster Tools. 2. Remove all URLs from the index via the Removal Tool in WMT 3. Add site-wide no-index, follow directive. Also, to remove the URLs in WMT, you usually have to block the URLs via /robots.txt. If I'd like to keep Google crawling through the subdomains and remove their URLs, is there a way to do so? Technical SEO | | RocketZando0
- 
		
		
		
		
		
		Cyrillic letter in URL - Encoding
 Hi all We are launching our site in Russia. As far as I can see by searching Google all sites have URLs in latin letters. Is there a special reason for this? - It seems that cyrillic letters also work. My technical staff says that it might give some encoding problems. Can anyone give me some insight into this? Thanks in advance.. / Kenneth Technical SEO | | Kennethskonto0
 
			
		 
			
		 
			
		 
					
				 
					
				 
					
				 
					
				 
					
				 
					
				 
					
				