Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
How to change noindex to index?
-
Hey,
I've recently upgraded to a pro SEOmoz account and have realised i have 14574 issues to do with 'blocked by meta-robot' and that 'This page is being kept out of the search engine indexes by the meta tag , which may have a value of "noindex", keeping this page out of the index.'
How can i change this so my pages get indexed?
I read somewhere that i need to change my privacy settings but that thread was 3 years old and now the WP Dashboard has updated..
Please let me know
Many thanks, Jamie
P.s Im using WordPress 3.5
And i have the XML sitemap plugin
And i have no idea where to look for this robots.txt file..
-
Answered below as well, but wanted to drop this here in case anyone else is looking. WP has changed the location of what used to be "Privacy" under settings. The functionality (which blocks search engines from your wordpress installation) is now under Settings->Reading
[Screenshot](Hi Mark Did you find it? I struggled too for a bit, but it's moved to Settings->Reading See this screenshot -Dan)
-Dan
-
Hi Mark
Did you find it? I struggled too for a bit, but it's moved to Settings->Reading
-Dan
-
Just updated it to this: http://gyazo.com/4a8a008055abbd563f96bf29b6b259a6.png?1357651763
And then checked my page sources and they're still 'noindex' - why can't i correct this?!
-
Just installed it and now its added this field into my settings>Reading
http://gyazo.com/0be601793fc1cb866d918ea61e7d8ec1.png?1357649141
What do i need to change to allow it to index all my pages?
(Don't want to type something in that will block all my pages
) -
Just asked the WordPress Forums and one of their reply was to install this plugin: http://wordpress.org/extend/plugins/wp-robots-txt/
Just seems to add the privacy tab again so i can set the settings to: I would like my blog to be visible to everyone, including search engines (like Google, Bing, Technorati) and archivers
Like you first stated

Will install it now and see how it goes

-
It could be that the older wordpress had a setting that this new version has decided to ignore. This is typical of programmers!
The next possibility is to look in the database, but the options part of the database is hard to read.
Another idea is to look in the code of the the theme and hack it, so it is permanently index, follow or just remove that altogether.
Maybe someone else has a better idea?
Alan
-
If i remember correctly my pages were still not being indexed before i installed the all in one SEO pack.
Here is my settings for the SEO pack: http://gyazo.com/6b4dddacb307bdacfdd7741894e0356b.png?1357647136
As you can see they are as you explained.
Any other ideas?
-
Yes, I would have them indexed in that case too.
I think it is the categories that are noindex.
I think this is an 'All in one SEO pack' adjustable feature.
In the setup for that, look for a checkbox:
"use noindex for categories"
uncheck that if it is checked.
If that isn't it, I don't know the answer
-
Thanks again for your reply Alan.
Currently the site is still in its final stages of development and once my automated system is finally built and implemented then I won't need to be changing any of the index pages except posting a few blogs once in a week or so.
So i think it would benefit me more to have each of my index pages getting indexed but then again I'm not sure on how to go about allowing them to be indexed due to WordPress' update.
My plugins are all highly downloaded and i use the 'All in one SEO pack' - if that may be the problem? I've gone through all the settings and the noindex buttons are all anticked.
Perhaps it could be the initial theme i used?
-
Thank you Mark
Nice looking site!
Your front page is index, follow.
Index pages are noindex, follow
Final pages are index, follow
I do something very close to this on my site.
Often, index pages are useless to searchers, because the index page changes so quickly that by the time the info gets into a search result, the information is no longer on that page, and the searcher will either just click away, cursing you and your site, or they will go looking through a few index pages and then curse you when they can't find what they wanted.
So I agree with the way you're doing that - if it is the case that the content changes quickly. If the index pages are just collectors of groups of items, then index, follow would be better, provided that you have enough text on the page to make it worthwhile.
As to how to make that happen, it isn't obvious.
I need to upgrade some of my sites to 3.5.
It could be that you have a plugin or a "custom field" that sets the index follow.
I suggest you edit a post and a page and scroll down to see if you have a field that handles it, such as "robotsmeta" that is set to "noindex,follow" for those pages
-
Hi Alan, thanks for your quick response.
My website is: www.FifaCoinStore.com
Here is a printscreen of my settings: http://gyazo.com/0cd3d21c5ec1797873a5c7cacc85a588.png?1357600674
I believe since the WordPress 3.5 update they have removed this privacy option which is why i can't seem to find it. I read this page from WordPress on it: http://codex.wordpress.org/Settings_Reading_Screen
Or am i just looking in the wrong place?
Thanks again
-
Hello Mark.
Please send me a bitly shortened link to your website so I can see what you are seeing
It probably isn't your robots file.
First try this.
In the Admin section, you should see "Settings" on the left navigation
Click that and you should see "Privacy"
Click that and you should see two radio buttons
<label for="blog-public">I would like my blog to be visible to everyone, including search engines (like Google, Bing, Technorati) and archivers</label>
<label for="blog-norobots">I would like to block search engines, but allow normal visitors</label>
Obviously, choose the top one and save it.
Then, refresh your front page or inner pages and look in the code to see if it still says noindex
If you have a cache, you will need to flush it.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Sudden Indexation of "Index of /wp-content/uploads/"
Hi all, I have suddenly noticed a massive jump in indexed pages. After performing a "site:" search, it was revealed that the sudden jump was due to the indexation of many pages beginning with the serp title "Index of /wp-content/uploads/" for many uploaded pieces of content & plugins. This has appeared approximately one month after switching to https. I have also noticed a decline in Bing rankings. Does anyone know what is causing/how to fix this? To be clear, these pages are **not **normal /wp-content/uploads/ but rather "index of" pages, being included in Google. Thank you.
Technical SEO | | Tom3_150 -
Resubmit sitemaps on every change?
Hello Mozers, Our sitemaps were submitted to Google and Bing, and are successfully indexed. Every time pages are added to our store (ecommerce), we re-generate the xml sitemap. My question is: should we be resubmitting the sitemaps every time their content change, or since they were submitted once can we assume that the crawlers will re-download the sitemaps by themselves (I don't like to assume). What are best practices here? Thanks!
Technical SEO | | yacpro131 -
Z-indexed content
I have some content on a page that I am not using any type of css hiding techniques, but I am using an image with a higher z-index in order to prevent the text from being seen until a user clicks a link to have the content scroll down. Are there any negative repercussions for doing this in regards to SEO?
Technical SEO | | cokergroup0 -
Investigating a huge spike in indexed pages
I've noticed an enormous spike in pages indexed through WMT in the last week. Now I know WMT can be a bit (OK, a lot) off base in its reporting but this was pretty hard to explain. See, we're in the middle of a huge campaign against dupe content and we've put a number of measures in place to fight it. For example: Implemented a strong canonicalization effort NOINDEX'd content we know to be duplicate programatically Are currently fixing true duplicate content issues through rewriting titles, desc etc. So I was pretty surprised to see the blow-up. Any ideas as to what else might cause such a counter intuitive trend? Has anyone else see Google do something that suddenly gloms onto a bunch of phantom pages?
Technical SEO | | farbeseo0 -
Noindex user profile
I have a social networking site with user- and company profiles. Some profiles have little to no content. One of the users here at moz suggested noindex-ing these profiles. I am still investigating this issue and have some follow up questions: What is the possible gain of no-indexing uninteresting profiles? Especially interested in this since these profiles do bring in long-tail traffic atm. How "irreversable" is introducing a noindex directive? Would everything "return to normal" if I remove te noindex directive? When determining the treshold for having profiles indexed, how should the following items be weighed Sum of number of words on the page (comprised of one or more of the following: full name, city, 0 to N company names, bio, activity) (unique) Profile picture (Nofollowed) Links to user's profiles on social networks or user's own site. Embedded Google Map Thanks!
Technical SEO | | thomasvanderkleij0 -
De-indexing millions of pages - would this work?
Hi all, We run an e-commerce site with a catalogue of around 5 million products. Unfortunately, we have let Googlebot crawl and index tens of millions of search URLs, the majority of which are very thin of content or duplicates of other URLs. In short: we are in deep. Our bloated Google-index is hampering our real content to rank; Googlebot does not bother crawling our real content (product pages specifically) and hammers the life out of our servers. Since having Googlebot crawl and de-index tens of millions of old URLs would probably take years (?), my plan is this: 301 redirect all old SERP URLs to a new SERP URL. If new URL should not be indexed, add meta robots noindex tag on new URL. When it is evident that Google has indexed most "high quality" new URLs, robots.txt disallow crawling of old SERP URLs. Then directory style remove all old SERP URLs in GWT URL Removal Tool This would be an example of an old URL:
Technical SEO | | TalkInThePark
www.site.com/cgi-bin/weirdapplicationname.cgi?word=bmw&what=1.2&how=2 This would be an example of a new URL:
www.site.com/search?q=bmw&category=cars&color=blue I have to specific questions: Would Google both de-index the old URL and not index the new URL after 301 redirecting the old URL to the new URL (which is noindexed) as described in point 2 above? What risks are associated with removing tens of millions of URLs directory style in GWT URL Removal Tool? I have done this before but then I removed "only" some useless 50 000 "add to cart"-URLs.Google says themselves that you should not remove duplicate/thin content this way and that using this tool tools this way "may cause problems for your site". And yes, these tens of millions of SERP URLs is a result of a faceted navigation/search function let loose all to long.
And no, we cannot wait for Googlebot to crawl all these millions of URLs in order to discover the 301. By then we would be out of business. Best regards,
TalkInThePark0 -
How to remove a sub domain from Google Index!
Hello, I have a website having many subdomains having same copy of content i think its harming my SEO for that site since abc and xyz sub domains do have same contents. Thus i require to know i have already deleted required subdomain DNS RECORDS now how to have those pages removed from Google index as well ? The DNS Records no more exists for those subdomains already.
Technical SEO | | anand20100 -
Index forum sites
Hi Moz Team, somehow the last question i raised a few days ago not only wasnt answered up until now, it was also completely deleted and the credit was not "refunded" - obviously there was some data loss involved with your restructuring. Can you check whether you still find the last question and answer it quickly? I need the answer 🙂 Here is one more question: I bought a website that has a huge forum, loads of pages with user generated content. Overall around 500.000 Threads with 9 Million comments. The complete forum is noindex/nofollow when i bought the site, now i am thinking about what is the best way to unleash the potential. The current system is vBulletin 3.6.10. a) Shall i first do an update of vbulletin to version 4 and use the vSEO tool to make the URLs clean, more user and search engine friendly before i switch to index/follow? b) would you recommend to have the forum in the folder structure or on a subdomain? As far as i know subdomain does take lesser strenght from the TLD, however, it is safer because the subdomain is seen as a separate entity from the regular TLD. Having it in he folder makes it easiert to pass strenght from the TLD to the forum, however, it puts my TLD at risk c) Would you release all forum sites at once or section by section? I think section by section looks rather unnatural not only to search engines but also to users, however, i am afraid of blasting more than a millionpages into the index at once. d) Would you index the first page of a threat or all pages of a threat? I fear duplicate content as the different pages of the threat contain different body content but the same Title and possibly the same h1. Looking forward to hear from you soon! Best Fabian
Technical SEO | | fabiank0