Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Adding index.php at the end of the url effect it's rankings
-
I have just had my site updated and we have put index.php at the end of all the urls. Not long after the sites rankings dropped. Checking the backlinks, they all go to (example) http://www.website.com and not http://www.website.com/index.php. So could this change have effected rankings even though it redirects to the new url?
-
Lot's of good advice here, so I'll just weigh in with my two cents...
Instead of redirecting all your files to /index.php, why not rewrite those in .htaccess to redirect back to the original (without the /index.php)
This has the dual effect of preserving your link equity to those original urls, and there's a slight correlation between shorter URLs and higher rankings (in part possibly because shorter URLs have a higher click-thru rate)
Regardless, I suspect a perfect storm of factors contributed to your rankings, as you stated yourself:
1. Site was down when on old US host for a minimum of 3 hours one day and perhaps longer
2. Changed from US host to host based in Spain
3. Analytics stopped recording data for 3 days and site was down briefly after the change of host
4. All original URLs now have /index.php at the end
I purposely left out a Google Algorithm change, because of Occam's Razor - the simplest explanation is usually correct, and an algorythm change would be too much coincidence.
As Robert said, make sure you're targeting the right country in Google Webmaster. Other than that, I would try very hard to return all URLs, hosting and settings back to their original state before all these changes.
-
Hi Robert,
Thanks for your input on this.
The webmaster changed the hosting from US to Spain without my knowledge, the first I knew of it was when I saw the rankings drop and called him to see if there were any changes done to the site.
The site also seemed to be down at various times during the process and the analytics stopped recording data for 3 days.
We had excellent rankings in UK & US (both our target markets) but the day we changed host, the rankings all dropped from between 5 to 30 positions and so far are showing no real sign of returing to their original rankings even though we have now changed to a new US host.
I don't know if you have ever had the same experience but I wonder how much of an impact this will have in the long run for the rankings and will they even return without having to significantly promote the site again.
There are a number of factors which took place over this period:
1. Site was down when on old US host for a minimum of 3 hours one day and perhaps longer
2. Changed from US host to host based in Spain
3. Analytics stopped recording data for 3 days and site was down briefly after the change of host
4. New Google algorithm change
5. All original URLs now have /index.php at the end
Can it be a combination of all these factors or is there one main culprit?
I will speak with our webmaster Monday and ensure that he has set the target language to EN-US but we are also targeting the UK market and prior to this we were ranked very well in both countries.
Again, thanks for all your feedback!
-
Authority, You just named the issue. Changed from US to Europe. I am assuming the site is in English, what country are you targeting?
If US, and you changed to Europe, you would have had to go into GWMT and change language settings to EN-US. As a .com is not a ccTLD, and a server residing in Europe will be presumed to be targeting there, if you are not set up with GWMT as EN-US, your rankings will drop for a US search.
So, now, no matter what you are targeting, go into GWMT and go to site config, settings, language and choose the correct language config. Even if you are US and you are hosted in US, I urge you to insure this is done.
LMK
-
Thanks all for you input!
We have done redirected the old urls to the new ones ie from www.site.com/keyword/ - to - www.site.com/keyword/index.php
We changed host on the 5th Feb. and literally the same day all rankings dropped. I know there have been recent Google updates but finding the real cause of this is still difficult. If there were no changes to the site, then I may have leant more towards a google algorithm update but the rankings dropped as soon as the sites hosting was changed from US to Europe. Hosting has since been changed back to US based.... 3 days on though and no significant improvements although some keywords are moving up 5 places or so.
Any more input appreciated
-
Authority
This may be implied or I may be missing something, but as to your links and 301's, if you are saying you did a single 301 of http://www.website.com to http://www.website.com/index.php then your rankings drop is because of that. For each url with links, you need to do the redirect of .com/url-a, url-b, etc. to .com/url-a/index.php, .com/url-b to .com/url-b/index.php, and so forth. This should be done in the .htaccess file. You will not transfer link juice by tranferring domain to domain, etc. You must do it url to url.
Hope it helps.
-
You say you have put index.php at the end of all the URLs? So each page is in it's own directory?
Harald is 100% correct but I am wondering; did you always have each page in it's own directory or was that part of the recent change? If the file names used to be more SEO friendly (i.e. keywords in the file name) and now they are just named index.php then that could have a lot to do with your rankings dropping.
Just wanted to add that, cheers.
-
Hi Authority Sitebuilder, First of all Google doesn't seem to care about these but for users' sake, for the sake of conformity and as a good practice, it is best to do a redirect
from(example)
to
http://www.website.com/index.php
In other words, select one URL and stick with it, redirect all others. Make a 301 redirect of your old URLs to your new URLs. Then it should not affect your ranking unless you will do some other changes on your pages As you said earlier that all the back links go to the http://www.website.com (old url), so it is better to redirects o the new url i.e http://www.website.com/index.php
I hope that your query had been solved.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
How can I make a list of all URLs indexed by Google?
I started working for this eCommerce site 2 months ago, and my SEO site audit revealed a massive spider trap. The site should have been 3500-ish pages, but Google has over 30K pages in its index. I'm trying to find a effective way of making a list of all URLs indexed by Google. Anyone? (I basically want to build a sitemap with all the indexed spider trap URLs, then set up 301 on those, then ping Google with the "defective" sitemap so they can see what the site really looks like and remove those URLs, shrinking the site back to around 3500 pages)
Intermediate & Advanced SEO | | Bryggselv.no0 -
Wrong URLs indexed, Failing To Rank Anywhere
I’m struggling with a client website that's massively failing to rank. It was published in Nov/Dec last year - not optimised or ranking for anything, it's about 20 pages. I came onboard recently, and 5-6 weeks ago we added new content, did the on-page and finally changed from the non-www to the www version in htaccess and WP settings (while setting www as preferred in Search Console). We then did a press release and since then, have acquired about 4 partial match contextual links on good websites (before this, it had virtually none, save for social profiles etc.) I should note that just before we added the (about 50%) new content and optimised, my developer accidentally published the dev site of the old version of the site and it got indexed. He immediately added it correctly to robots.txt, and I assumed it would therefore drop out of the index fairly quickly and we need not be concerned. Now it's about 6 weeks later, and we’re still not ranking anywhere for our chosen keywords. The keywords are around “egg freezing,” so only moderate competition. We’re not even ranking for our brand name, which is 4 words long and pretty unique. We were ranking in the top 30 for this until yesterday, but it was the press release page on the old (non-www) URL! I was convinced we must have a duplicate content issue after realising the dev site was still indexed, so last week, we went into Search Console to remove all of the dev URLs manually from the index. The next day, they were all removed, and we suddenly began ranking (~83) for “freezing your eggs,” one of our keywords! This seemed unlikely to be a coincidence, but once again, the positive sign was dampened by the fact it was non-www page that was ranking, which made me wonder why the non-www pages were still even indexed. When I do site:oursite.com, for example, both non-www and www URLs are still showing up…. Can someone with more experience than me tell me whether I need to give up on this site, or what I could do to find out if I do? I feel like I may be wasting the client’s money here by building links to a site that could be under a very weird penalty 😕
Intermediate & Advanced SEO | | Ullamalm0 -
What's the best possible URL structure for a local search engine?
Hi Mozzers, I'm working at AskMe.com which is a local search engine in India i.e if you're standing somewhere & looking for the pizza joints nearby, we pick your current location and share the list of pizza outlets nearby along with ratings, reviews etc. about these outlets. Right now, our URL structure looks like www.askme.com/delhi/pizza-outlets for the city specific category pages (here, "Delhi" is the city name and "Pizza Outlets" is the category) and www.askme.com/delhi/pizza-outlets/in/saket for a category page in a particular area (here "Saket") in a city. The URL looks a little different if you're searching for something which is not a category (or not mapped to a category, in which case we 301 redirect you to the category page), it looks like www.askme.com/delhi/search/pizza-huts/in/saket if you're searching for pizza huts in Saket, Delhi as "pizza huts" is neither a category nor its mapped to any category. We're also dealing in ads & deals along with our very own e-commerce brand AskMeBazaar.com to make the better user experience and one stop shop for our customers. Now, we're working on URL restructure project and my question to you all SEO rockstars is, what can be the best possible URL structure we can have? Assume, we have kick-ass developers who can manage any given URL structure at backend.
Intermediate & Advanced SEO | | _nitman0 -
Removing index.php
I have question for the community and whether or not this is a good or bad idea. I currently have a Joomla site that displays www.domain.com/index.php in all the URLs with the exception of the home page. I have read that it's better to not have index.php showing in the URL at all. Does it really matter if I have index.php in my URL? I've read that it is a bad practice. I am thinking about installing the sh404SEF component on my site and removing the index.php. However, I rank pretty high for the keywords I want in Google, Bing and Yahoo. All of the URLs that show up in the searches have index.php as part of the URL. Has anyone ever used sh404SEF to remove the index.php and how did you overcome not loosing your search engine links? I don't want an existing search showing www.domain.com/index.php/sales and it not linking to the correct page which would now be www.domain.com/sales. I guess I could insert the proper redirects in the htaccess file. But I was hoping to avoid having every page of my site in the htaccess file for redirecting. Any help or advice appreciated.
Intermediate & Advanced SEO | | MedGroupMedia0 -
Using the same content on different TLD's
HI Everyone, We have clients for whom we are going to work with in different countries but sometimes with the same language. For example we might have a client in a competitive niche working in Germany, Austria and Switzerland (Swiss German) ie we're going to potentially rewrite our website three times in German, We're thinking of using Google's href lang tags and use pretty much the same content - is this a safe option, has anyone actually tries this successfully or otherwise? All answers appreciated. Cheers, Mel.
Intermediate & Advanced SEO | | dancape1 -
Brackets vs Encoded URLs: The "Same" in Google's eyes, or dup content?
Hello, This is the first time I've asked a question here, but I would really appreciate the advice of the community - thank you, thank you! Scenario: Internal linking is pointing to two different versions of a URL, one with brackets [] and the other version with the brackets encoded as %5B%5D Version 1: http://www.site.com/test?hello**[]=all&howdy[]=all&ciao[]=all
Intermediate & Advanced SEO | | mirabile
Version 2: http://www.site.com/test?hello%5B%5D**=all&howdy**%5B%5D**=all&ciao**%5B%5D**=all Question: Will search engines view these as duplicate content? Technically there is a difference in characters, but it's only because one version encodes the brackets, and the other does not (See: http://www.w3schools.com/tags/ref_urlencode.asp) We are asking the developer to encode ALL URLs because this seems cleaner but they are telling us that Google will see zero difference. We aren't sure if this is true, since engines can get so _hung up on even one single difference in character. _ We don't want to unnecessarily fracture the internal link structure of the site, so again - any feedback is welcome, thank you. 🙂0 -
Best way to permanently remove URLs from the Google index?
We have several subdomains we use for testing applications. Even if we block with robots.txt, these subdomains still appear to get indexed (though they show as blocked by robots.txt. I've claimed these subdomains and requested permanent removal, but it appears that after a certain time period (6 months)? Google will re-index (and mark them as blocked by robots.txt). What is the best way to permanently remove these from the index? We can't use login to block because our clients want to be able to view these applications without needing to login. What is the next best solution?
Intermediate & Advanced SEO | | nicole.healthline0 -
How to deal with old, indexed hashbang URLs?
I inherited a site that used to be in Flash and used hashbang URLs (i.e. www.example.com/#!page-name-here). We're now off of Flash and have a "normal" URL structure that looks something like this: www.example.com/page-name-here Here's the problem: Google still has thousands of the old hashbang (#!) URLs in its index. These URLs still work because the web server doesn't actually read anything that comes after the hash. So, when the web server sees this URL www.example.com/#!page-name-here, it basically renders this page www.example.com/# while keeping the full URL structure intact (www.example.com/#!page-name-here). Hopefully, that makes sense. So, in Google you'll see this URL indexed (www.example.com/#!page-name-here), but if you click it you essentially are taken to our homepage content (even though the URL isn't exactly the canonical homepage URL...which s/b www.example.com/). My big fear here is a duplicate content penalty for our homepage. Essentially, I'm afraid that Google is seeing thousands of versions of our homepage. Even though the hashbang URLs are different, the content (ie. title, meta descrip, page content) is exactly the same for all of them. Obviously, this is a typical SEO no-no. And, I've recently seen the homepage drop like a rock for a search of our brand name which has ranked #1 for months. Now, admittedly we've made a bunch of changes during this whole site migration, but this #! URL problem just bothers me. I think it could be a major cause of our homepage tanking for brand queries. So, why not just 301 redirect all of the #! URLs? Well, the server won't accept traditional 301s for the #! URLs because the # seems to screw everything up (server doesn't acknowledge what comes after the #). I "think" our only option here is to try and add some 301 redirects via Javascript. Yeah, I know that spiders have a love/hate (well, mostly hate) relationship w/ Javascript, but I think that's our only resort.....unless, someone here has a better way? If you've dealt with hashbang URLs before, I'd LOVE to hear your advice on how to deal w/ this issue. Best, -G
Intermediate & Advanced SEO | | Celts180