Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Should I remove the ?replytocom variables in wordpress?
-
I'm using Yoast's wordpress plugin and there is an option to remove the replytocom variables. I'm curious what everyone's thoughts were on that, and if I should do it.
Here's the site if you need to see it.
Thanks!
-
Hey, guys, this is a very old post but because it might be useful to people in the future I thought I would update the URL that Ryan did a. great job posting but no longer goes to the correct place. No one can control what I third-party site changes there URL structure to rite?
you can use this plug-in called
?replytocom= replaced with #replytocom= -
Mine is not indexed in the Google Search, but in Google Webmasters and SEOmoz they showed the error !
Should i remove those links via URL parameters ?
Not indexed, someone here told me some plugins are helping you, i have All-in seo plugin and removed the comluv plugin(for more spam in my blog). other plugins are like share social network That's it.
Any help for me and my blog will be much appreciated !
-
Thanks Ryan! Have a great 4th of July!

-
What do people not using this plug in do? I'm assuming not many people do this, right?
I presume they accept the WP default options. Our practice and understanding of SEO is what allows us to analyze and make decisions regarding tidbits such as the one you mentioned.
You think there is any benefit to doing it, or just one of those "hey why not" sort of things?
I do think there is a benefit. You are impacting a LOT of links. Every comment on your site. It may be a tiny 1% benefit type of thing, but the change applies site wide and will presumably be in place for years.
-
I''ll try to find the link where he talks about using pages instead of posts and share it. Curious to hear your thoughts on it.
I'll go ahead and select that option, thanks for your help. (On a side note, what do people not using this plug in do? I'm assuming not many people do this, right?)
You think there is any benefit to doing it, or just one of those "hey why not" sort of things?
-
Regarding the new pages instead of posts idea, do you have a link to share?
Regarding the comment url, the page with the comment should be fully indexed either way. By changing the link, you are helping search engines better understand your site. The comment links do not represent a new page or new information.
Google clearly understands WP sites exceptionally well. I am confident you can choose various options and they will still understand those links represent comments. With that said, I would still go with Yoast on this one.
Actually, SEOmoz does it too. Take a look at their blog comments.
-
Thanks for taking the time to check into it. One I'm concerned with is how this will effect long tail seo / indexing of the comments. How will this effect my organic traffic? (will it hurt it?)
I don't see these sorts of pages coming up in google now, so I'm not sure what selecting that option does (and how it effects the site.)
Yoast does a few things different with his site, and I don't always follow his lead. For example he suggets making new pages instead of new posts for your blog posts. He's the only one I've ever heard say this, or do this.

-
I just took a look at Yoast's site and I now better understand the option to remove the variables. I recommend selecting that option. From the Yoast site:
method remove_reply_to_com [line 939]
string remove_reply_to_com( string $link)
Removes the ?replytocom variable from the link, replacing it with a #comment- <number>anchor.</number> Tags: access: public Parameters: string $link The comment link as a string.
Example: http://yoast.com/user-contact-fields-wordpress/#comment-110294
-
hmm..thanks for the feedback. So do you suggest not blocking those? (and I'll message yoast also and see what his thoughts are.)
Thanks.
-
I understand the logic behind blocking removing the variables. They are a lot of extra links on the page which some webmasters might prefer to manage.
What I would prefer is to reform the link so it was something like: http://noahsdad.com/treadmill-training-progress#replytocom=22729
I am guessing the "respond" portion of the URL acts as if someone pressed the reply button which seems unnecessary. If someone clicks the link whether in search results or otherwise and is taken directly to the comment, they should be quite happy. If they wish to reply they can hit the reply button.
Google ignores anything after the # character in a URL. Therefore Google would see these as simply a link to the page which should already be indexed.
Perhaps you can ask Yoast about his thoughts.
-
Thanks for the kind words, I agree, he is a cutie.
Will blocking those cause the comments not to be indexed though? -
Yup - removing those will save you the trouble of duplicate content - since Google by default is crawling those as different URLs. By default, if you have comments enabled, there's a link at the bottom of posts with that parameter in the url (the same as the blog post url - see here ---> http://noahsdad.com/treadmill-training-progress/?replytocom=22729#respond ).
Noah is cute!
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Should I use noindex or robots to remove pages from the Google index?
I have a Magento site and just realized we have about 800 review pages indexed. The /review directory is disallowed in robots.txt but the pages are still indexed. From my understanding robots means it will not crawl the pages BUT if the pages are still indexed if they are linked from somewhere else. I can add the noindex tag to the review pages but they wont be crawled. https://www.seroundtable.com/google-do-not-use-noindex-in-robots-txt-20873.html Should I remove the robots.txt and add the noindex? Or just add the noindex to what I already have?
Intermediate & Advanced SEO | | Tylerj0 -
Removing .html from URLs - impact of rankings?
Good evening Mozzers. Couple of questions which I hope you can help with. Here's the first. I am wondering, are we likely to see ranking changes if we remove the .html from the sites URLs. For example website.com/category/sub-category.html Change to: website.com/category/sub-category/ We will of course make sure we 301 redirect to the new, user friendly URLs, but I am wondering if anyone has had previous experience of implementing this change and how it has effected rankings. By having the .html in the URLs, does this stop link juice being flowed back to the root category? Second question: If one page can be loaded with and without a forward slash "/" at the end, is this a duplicate page, or would Google consider this as the same page? Would like to eliminate duplicate content issues if this is the case. For example: website.com/category/ and website.com/category Duplicate content/pages?
Intermediate & Advanced SEO | | Jseddon920 -
Best way to remove full demo (staging server) website from Google index
I've recently taken over an in-house role at a property auction company, they have a main site on the top-level domain (TLD) and 400+ agency sub domains! company.com agency1.company.com agency2.company.com... I recently found that the web development team have a demo domain per site, which is found on a subdomain of the original domain - mirroring the site. The problem is that they have all been found and indexed by Google: demo.company.com demo.agency1.company.com demo.agency2.company.com... Obviously this is a problem as it is duplicate content and so on, so my question is... what is the best way to remove the demo domain / sub domains from Google's index? We are taking action to add a noindex tag into the header (of all pages) on the individual domains but this isn't going to get it removed any time soon! Or is it? I was also going to add a robots.txt file into the root of each domain, just as a precaution! Within this file I had intended to disallow all. The final course of action (which I'm holding off in the hope someone comes up with a better solution) is to add each demo domain / sub domain into Google Webmaster and remove the URLs individually. Or would it be better to go down the canonical route?
Intermediate & Advanced SEO | | iam-sold0 -
Removing poor domain authority backlinks worth it?
Hey Moz, I am working with a client on more advanced SEO tactics. This client has a reputable domain authority of 67 and 50,000+ backlinks. We're wanting to continue SEO efforts and stay on top of any bad backlinks that may arise. Would it be worth asking websites (below 20 domain authority) to remove our links? Then, use the disavow tool if they do not respond. Is this a common SEO practice for continued advanced efforts? Also, what would your domain authority benchmark be? I used 20 just as an example. Thanks so much for your help. Cole
Intermediate & Advanced SEO | | ColeLusby1 -
Removing index.php
I have question for the community and whether or not this is a good or bad idea. I currently have a Joomla site that displays www.domain.com/index.php in all the URLs with the exception of the home page. I have read that it's better to not have index.php showing in the URL at all. Does it really matter if I have index.php in my URL? I've read that it is a bad practice. I am thinking about installing the sh404SEF component on my site and removing the index.php. However, I rank pretty high for the keywords I want in Google, Bing and Yahoo. All of the URLs that show up in the searches have index.php as part of the URL. Has anyone ever used sh404SEF to remove the index.php and how did you overcome not loosing your search engine links? I don't want an existing search showing www.domain.com/index.php/sales and it not linking to the correct page which would now be www.domain.com/sales. I guess I could insert the proper redirects in the htaccess file. But I was hoping to avoid having every page of my site in the htaccess file for redirecting. Any help or advice appreciated.
Intermediate & Advanced SEO | | MedGroupMedia0 -
How long takes to a page show up in Google results after removing noindex from a page?
Hi folks, A client of mine created a new page and used meta robots noindex to not show the page while they are not ready to launch it. The problem is that somehow Google "crawled" the page and now, after removing the meta robots noindex, the page does not show up in the results. We've tried to crawl it using Fetch as Googlebot, and then submit it using the button that appears. We've included the page in sitemap.xml and also used the old Google submit new page URL https://www.google.com/webmasters/tools/submit-url Does anyone know how long will it take for Google to show the page AFTER removing meta robots noindex from the page? Any reliable references of the statement? I did not find any Google video/post about this. I know that in some days it will appear but I'd like to have a good reference for the future. Thanks.
Intermediate & Advanced SEO | | fabioricotta-840380 -
Limit on Google Removal Tool?
I'm dealing with thousands of duplicate URL's caused by the CMS... So I am using some automation to get through them - What is the daily limit? weekly? monthly? Any ideas?? thanks, Ben
Intermediate & Advanced SEO | | bjs20100 -
Removing Content 301 vs 410 question
Hello, I was hoping to get the SEOmoz community’s advice on how to remove content most effectively from a large website. I just read a very thought-provoking thread in which Dr. Pete and Kerry22 answered a question about how to cut content in order to recover from Panda. (http://www.seomoz.org/q/panda-recovery-what-is-the-best-way-to-shrink-your-index-and-make-google-aware). Kerry22 mentioned a process in which 410s would be totally visible to googlebot so that it would easily recognize the removal of content. The conversation implied that it is not just important to remove the content, but also to give google the ability to recrawl that content to indeed confirm the content was removed (as opposed to just recrawling the site and not finding the content anywhere). This really made lots of sense to me and also struck a personal chord… Our website was hit by a later Panda refresh back in March 2012, and ever since then we have been aggressive about cutting content and doing what we can to improve user experience. When we cut pages, though, we used a different approach, doing all of the below steps:
Intermediate & Advanced SEO | | Eric_R
1. We cut the pages
2. We set up permanent 301 redirects for all of them immediately.
3. And at the same time, we would always remove from our site all links pointing to these pages (to make sure users didn’t stumble upon the removed pages. When we cut the content pages, we would either delete them or unpublish them, causing them to 404 or 401, but this is probably a moot point since we gave them 301 redirects every time anyway. We thought we could signal to Google that we removed the content while avoiding generating lots of errors that way… I see that this is basically the exact opposite of Dr. Pete's advice and opposite what Kerry22 used in order to get a recovery, and meanwhile here we are still trying to help our site recover. We've been feeling that our site should no longer be under the shadow of Panda. So here is what I'm wondering, and I'd be very appreciative of advice or answers for the following questions: 1. Is it possible that Google still thinks we have this content on our site, and we continue to suffer from Panda because of this?
Could there be a residual taint caused by the way we removed it, or is it all water under the bridge at this point because Google would have figured out we removed it (albeit not in a preferred way)? 2. If there’s a possibility our former cutting process has caused lasting issues and affected how Google sees us, what can we do now (if anything) to correct the damage we did? Thank you in advance for your help,
Eric1