Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Using geolocation for dynamic content - what's the best practice for SEO?
-
Hello
We sell a product globally but I want to use different keywords to describe the product based on location.
For this example let’s say in USA the product is a "bathrobe" and in Canada it’s a "housecoat" (same product, just different name).
What this means… I want to show "bathrobe" content in USA (lots of global searches) and "housecoat" in Canada (less searches).
I know I can show the content using a geolocation plugin (also found a caching plugin which will get around the issue of people seeing cached versions), using JavaScript or html5.
I want a solution which enables someone in Canada searching for "bathrobe" to be able to find our site through Google search though too. I want to rank for "bathrobe" in BOTH USA and Canada.
I have read articles which say Google can read the dynamic content in JavaScript, as well as the geolocation plugin. However the plugins suggest Google crawls the content based on location too. I don’t know about JavaScript.
Another option is having two separate pages (one for “bathrobe” and one for “housecoat”) and using geolocation for the main menu (if they find the other page i.e. bathrobe page through a Canadian search, they will still see it though). This may have an SEO impact splitting the traffic though.
Any suggestions or recommendations on what to do?? What do other websites do? I’m a bit stuck.
Thank you so much!
Laura
Ps. I don’t think we have enough traffic to add subdomains or subdirectories.
-
Hello Benjamin,
This is an interesting problem. I'm going to provide my opinion, but I highly recommend studying up on International SEO, which you can do at the links below:
https://a-moz.groupbuyseo.org/learn/seo/international-seo
http://www.aleydasolis.com/en/I don't know what the plugin does, but if it generates a new URL (e.g. adds ?loc=ca or something like that) for the location change you'll want to use rel="alternate" hreflang="*" tags, which would look something like this:
Google recommends putting one language per page, so that would be a different URL for each version, as highlighted in red here.
**However, it sounds to me like all of this is done client-side using JavaScript, and that the URL doesn't change, only the content.** If this is the case - as long as you are serving the same content to Googlebot crawling in Canada as you are to a visitor in Canada you probably won't have any issues, as described here: https://webmasters.googleblog.com/2008/06/how-google-defines-ip-delivery.html .
For the situation you described, it seems like you could put both keyword variations in the content and that would be good enough. But then you don't want to spell specialising with an S in one line and specializing with a z in the next.
Another thing to look into is whether both versions of the content appear in the code, or just one or the other. You definitely don't want to have multiple versions of the content in the source code. But you also don't want to hide both versions via JavaScript, only to load one or the other client-side. That creates even more problems.
One would think there would be a Vary: Location response header, similar to how responses are provided when content varies by user-agent or cookie: https://www.fastly.com/blog/best-practices-for-using-the-vary-header . Alas, I can't find any use cases of this and it's not a "thing". I'm not sure why this is, but maybe an International SEO Expert like Alayda Solis would know. I'll ping her into the thread if she has time.
-
Thank you. That link was helpful.
-
This is a great opportunity to test some of your ideas. It may be a good idea to create unique landing pages based on the most highly search keyword per region and target them on the corresponding geo-page. Read more about great ways to rank geo-targeted pages--and make them convert: https://a-moz.groupbuyseo.org/blog/scaling-geo-targeted-local-landing-pages-that-really-rank-and-convert-whiteboard-friday
However, it may be agood idea to optimize the homepage, about page, and service description pages for the products rather than the locations.And, since you're a national brand, it may be smart to try so PPC ads to geo-target your advertising and use those keywords accordingly.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
What's your proudest accomplishment in regards to SEO?
After many years in the industry, you come to realize a few things. One of of the biggest pain points for us at web daytona was being able to give clients a quick keyword ranking cost estimation. After multiple trial and error and relying on API data from one of the most reliable SEO softwares in our industry, we were able to develop an SEO tool that allows us to quickly and accurately get the estimated cost for a given keyword (s) using multiple variables. Most agencies can relate to that story. It’s something my colleagues and I at Web Daytona have been through before. Finding the cost and amount of time needed to rank for a keyword is a time consuming process. That’s why it’s a common practice to sell SEO packages of 5-10 keywords for about $1000-2000 / month. The problem is not all keywords are equally valuable, and most clients know this. We constantly get questions from clients asking: “how much to rank for this specific keyword?” It’s difficult to answer that question with a pricing model that treats the cost of ranking every keyword equally. So is the answer to spend a lot more time doing tedious in-depth keyword research? If we did we could give our clients more precise estimates. But being that a decent proposal can take as long as 2-5 hours to make, and agency life isn’t exactly full of free time, that wouldn’t be ideal. That’s when we asked a question. What if we could automate the research needed to find the cost of ranking keywords? We looked around for a tool that did, but we couldn’t find it. Then we decided to make it ourselves. It wasn’t going to be easy. But after running an SEO agency for over a decade, we knew we had the expertise to create a tool that wouldn’t just be fast and reliable, it would also be precise. Fast forward to today and we’re proud to announce that The Keyword Cost Estimator is finally done. Now we’re releasing it to the public so other agencies and businesses can use it too. You can see it for yourself here. Keyword-Rank-Cost-Ectimator-Tool-by-Web-Daytona-Agency.png
Local Website Optimization | | WebDaytona0 -
Local SEO Over Optimization
We are targeting a bunch of services for our local business that works in and around their location. I'm concerned about over optimization and need some guidance on whether these points should be resolved. The company is based in a city and works mostly in the city but also in the surrounding areas. Currently, the site has 6 services pages (accessible via main nav) targeting the same location i.e. “Made Up Service London”, “Imaginary Service London” (with URLs and H1 tags etc. in place containing this location). However this is soon going to become 9 services pages, I am concerned that the repetition of this one location is starting to look spammy, especially as its where the company is based. Initially, I also wanted pages targeting the same services in other nearby areas. For example “Made Up Service Surrey”, “Imaginary Service Essex”. This has not happened as the info available has been too sporadic. I was going to add links to relevant case studies into these pages to beef up the content and add interest. To that end, we came up with case studies, but after a while, I noticed that these are also largely focused on the primary location. So out of 32 case studies, we have 19 focused on the primary location again with URL’s and H1 tags etc containing the location keyword. So in total, we have 25 pages optimized for the location (soon to be 28 and more if further case studies are added). My initial feeling was that the inclusion of pages targeting services in other locations would legitimize what we have done with the main pages. But obviously we have not got these pages in place and I question whether we ever will. What is my best course of action moving forward?
Local Website Optimization | | GrouchyKids1 -
SEO Best Practice for Managing a Businesses NAP with Multiple Addresses
I have a client with multiple business addresses - 3 across 3 states, from an SEO perspective what would be the best approach for displaying a NAP on the website? So far I've read that its best: to get 3 GMB account to point to 3 location pages & use a local phone number as opposed to a 1300 number. Display all 3 locations in the footer, run of site
Local Website Optimization | | jasongmcmahon1 -
Content writing for single entity business (The use of I)
Most of my clients consist of single entity law firms in which my clients repeatedly use the pronoun "I" to describe every service they provide. I have always preferred using the business name The Law Office of..." put lawyer name here". Is it ok to repetitively use the pronoun "I" in the content. To me it feels lack luster and childish not very professional, however I have a hard time convincing the lawyers of this. What are your thoughts? Can good content be written with the repetitive use of "I"? If not is the business name sufficient or maybe another pronoun? I will be showing responses to my clients if that is ok.
Local Website Optimization | | donsilvernail0 -
Local SEO - Adding the location to the URL
Hi there, My client has a product URL: www.company.com/product. They are only serving one state in the US. The existing URL is ranking in a position between 8-15 at the moment for local searches. Would it be interesting to add the location to the URL in order to get a higher position or is it dangerous as we have our rankings at the moment. Is it really giving you an advantage that is worth the risk? Thank you for your opinions!
Local Website Optimization | | WeAreDigital_BE
Sander0 -
Call Tracking, DNI Script & Local SEO
Hi Moz! I've been reading about this a lot more lately - and it doesn't seem like there's exactly a method that Google (or other search engines) would consider to be "best practices". The closest I've come to getting some clarity are these Blumenthals articles - http://blumenthals.com/blog/2013/05/14/a-guide-to-call-tracking-and-local/ & the follow-up piece from CallRail - http://blumenthals.com/blog/2014/11/25/guide-to-using-call-tracking-for-local-search/. Assuming a similar goal of using an existing phone number with a solid foundation in the local search ecosystem, and to create the ability to track how many calls are coming organically (not PPC or other paid platform) to the business directly from the website for an average SMB. For now, let's also assume we're also not interested in screening the calls, or evaluating customer interaction with the staff - I would love to hear from anyone who has implemented the DNI call tracking info for a website. Were there negative effects on Local SEO? Did the value of the information (# of calls/month) outweigh any local search conflicts? If I was deploying this today, it seems like the blueprint for including DNI script, while mitigating risk for losing local search visibility might go something like this: Hire reputable call-tracking service, ensure DNI will match geographic area-code & be "clean" numbers Insert DNI script on key pages on site Maintain original phone number (non-DNI) on footer, within Schema & on Contact page of the site ?? Profit Ok, those last 2 bullet points aren't as important, but I would be curious where other marketers land on this issue, as I think there's not a general consensus at this point. Thanks everyone!
Local Website Optimization | | Etna1 -
SEO: .com vs .org vs .travel Domain
Hi there, I am new to MOZ Q&A and first of all I appreciate all the folks here that share their expertise and make everyone understand 'the WWW' a bit better. My question: I have been developing a 'travel guide' site for a city in the U.S. and now its time to choose the right domain name. I put a strong focus on SEO in terms of coding, site performance as well as content and to round things up I'd like to register the _best _domain name in terms of SEO. Let's suppose the city is Atlanta. I have found the following domain names that are available and I was wondering whether you guys could give me some inside on which domain name would perform best. discoveratlanta.org
Local Website Optimization | | kinimod
atlantaguide.org
atlanta.travel
atlantamag.com Looking at the Google Adwords Keyword tool the term that reaches the highest search queries is obviously "Atlanta" itself. Sites that are already ranking high are atlanta.com and atlanta.gov. So basically I am wondering whether I should aim for a new TLD like atlanta.travel or rather go with a .org domain. I had a look around and it seems that .org domains generally work well for city guides (at least a lot of such sites use .org domains). However, I have also seen a major US city that uses .travel and ranks first. On the other hand in New York, nycgo.com ranks well. Is it safe to assume that from the domain names I mentioned it really doesn't matter which one I use since it wouldn't significantly affect my ranking (good or bad)? Or would you still choose one above the other? What do you generally thing about .travel domain names (especially since they are far more expensive then the rest)? I really appreciate your response to my question! Best,
kinimod0 -
International Site Geolocation Redirection (best way to redirect and allow Google bots to index sites)
I have a client that has an international website. The website currently has IP detection and redirects you to the subdomain for your country. They have currently only launched the Australian website and are not yet open to the rest of the world: https://au.domain.com/ Google is not indexing the Australian website or pages, instead I believe that the bots are being blocked by the IP redirection every time they try to visit one of the Australian pages. Therefore only the US 'coming soon' page is being properly indexed. So, I would like to know the best way to place a geolocation redirection without creating a splash page to select location? User friendliness is most important (so we don't want cookies etc). I have seen this great Whiteboard Friday video on Where to Host and How to Target, which makes sense, but what it doesn't tell me is exactly the best method for redirection except at about 10:20 where it tells me what I'm doing is incorrect. I have also read a number of other posts on IP redirection, but none tell me the best method, and some are a little different examples... I need for US visitors to see the US coming soon page and for Google to index the Australian website. I have seen a lot about JS redirects, IP redirects and .htaccess redirects, but unfortunately my technical knowledge of how these affect Google's bots doesn't really help. Appreciate your answers. Cheers, Lincoln
Local Website Optimization | | LincolnSmith0