Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Low text-HTML ratios
-
Are low text-HTML ratios still a negative SEO ranking factor?
Today I ran SEMRUSH site audit that showed 344 out of 345 pages on our website (www.nyc-officespace-leader.com) show an text-HTML ratio that ranges from 8% to 22%. This is characterized as a warning on SEMRUSH. This error did not exist in April when the last SEMRUSH audit was conducted.
Is it worthwhile to try to externalize code in order to improve this ratio? Or to add text (major project on a site of this size)? These pages generally have 200-400 words of text.
Certain URLs, for example www.nyc-officespace-leader.com/blog/nycofficespaceforlease more text, yet it still shows an text-HTML ratio of only 16%.
We recently upgraded to the WordPress 4.2.1. Could this have bloated the code (CSS etcetera) to the detriment of the text-HTML ratio?
If Google has become accustomed to more complex code, is this a ratio that I can ignore.
Thanks, Alan
-
Hi Cynthia,
A text-to-HTML ratio of between 15 and 70 percent is generally considered ideal.
Try this tool, http://tools.seochat.com/tools/code-to-text-ratio/
It is a pretty good tool, but like Andy stated, the best bet is to focus on improving the content on your site, followed by running these pages through https://validator.w3.org/ and follow whatever recommendations possible to reduce code bloat etc...
Cheers,
Kevin
-
We have the same issues on two websites, and I have had difficulty determining what is actually wrong or causing this? Is there another tool that may give more light to what SEMrush is "finding" as low text-html ratios? I am not sure what to address on some of the pages that SEMrush has flagged that as a warning.
Thanks, Cindy
-
Hi,
First of all, you should be aware that SEMRush changed their algorithm a week or so ago and a client of mine also saw the same issue. They have really set the bar low when it comes to showing this factor now, to the point I don't agree with it.
However, I would use this as a signal just to allow you to see if the content is lacking a little on the page, rather than worrying if it will affect SEO. Of course, if you have a page with no content, this is bad for both usability and SEO, but text / HTML ratio isn't a direct ranking signal. See it as more of an indicator to see if the code can be cleaned and to help increase page load times.
There is never any harm in trying to speed up your pages either. If you think that some of the code is slowing things down, then look at how this can be improved. You can always do more to better the user experience, and with that will come indirect SEO improvements.
Write amazing content for your pages (where possible) and don't worry too much about text / HTML ratio. I know which one Google will be paying more attention to

-Andy
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
PDF ranking higher than HTML pages, solution?
Hello Moz community I know this question has been asked before but it seems there is no real answer other than putting a summary of the PDF on the HTML page. My problem is other websites are using my PDFs, I have some PDFs with very high authority links and I would like to either pass the link juice on to my product/category page or do rel=canonical somehow. I'm using bigcommerce as my platform. My website is cwwltd.com. Any help would be greatly appreciated. Thank you
Intermediate & Advanced SEO | | Neverstop1231 -
How to outrank a directory listing with high DA but low PA?
My site is at 4th place, 3 places above it is a gumtree (similar to yell, yelp) listing. How can you figure out how difficult it would be outrank those pages? I mean obviously the pages would have low PA and they are top based on the high DA of the site. This also seems to go back to keyword research and difficulty, when I'm doing keyword research and I see a wikipedia site in top 5 rank, or a yell.com or perhaps an article in forbes.com outranks your site. Typically the problem seems to be Google giving a lot of credit to these pages rankings based on the high DA rather than PA of the pages. How would you gauge the difficulty of that keyword then if the competition are pages with very high DA which is impossible to compete with but low PA? Thanks
Intermediate & Advanced SEO | | magusara2 -
Should I set up no index no follow on low quality pages?
I know it is a good idea for duplicate pages, blog tags, etc. but I remember somewhere that you can help the overall link juice of a website by adding no index no follow or no index follow low quality content pages of your website. Is it still a good idea to do this or was it never a good idea to begin with? Michael
Intermediate & Advanced SEO | | Michael_Rock0 -
How much does dirty html/css etc impact SEO?
Good Morning! I have been trying to clean up this website and half the time I can't even edit our content without breaking the WYSIWYG Editor. Which leads me to the next question. How much, if at all, is this impacting our SEO. To my knowledge this isn't directly causing any broken pages for the viewer, but still, it certainly concerns me. I found this post on Moz from last year: http://a-moz.groupbuyseo.org/community/q/how-much-impact-does-bad-html-coding-really-have-on-seo We have a slightly different set of code problems but still wanted to revisit this question and see if anything has changed. I also can't imagine that all this broken/extra code is helping our page load properly. Thanks everybody!
Intermediate & Advanced SEO | | HashtagHustler0 -
Is it normal to initially rank low in the SERPs, then over time gain rank?
We just released a very targeted page for a specific item about 18 hours ago. For the main keyword as well as multiple variations, we currently are ranking around # 40 to # 50 depending on what the exact query is. Is it normal to initially rank lower in the SERPs and then as the page ages, gain? Thank you for your insights!
Intermediate & Advanced SEO | | DJ1231 -
Google Analytics: how to filter out pages with low bounce rate?
Hello here, I am trying to find out how I can filter out pages in Google Analytics according to their bounce rate. The way I am doing now is the following: 1. I am working inside the Content > Site Content > Landing Pages report 2. Once there, I click the "advanced" link on the right of the filter field. 3. Once there, I define to "include" "Bounce Rate" "Greater than" "0.50" which should show me which pages have a bounce rate higher of 0.50%.... instead I get the following warning on the graph: "Search constraints on metrics can not be applied to this graph" I am afraid I am using the wrong approach... any ideas are very welcome! Thank you in advance.
Intermediate & Advanced SEO | | fablau0 -
301 redirect from .html to non .html?
Previously our site was using this as our URL structure: www.site.com/page.html. A few months ago we updated our URL structure to this: www.site.com/page & we're not using the .html. I've read over this guide & don't see anywhere that discusses this: http://www.seomoz.org/learn-seo/redirection. I've currently got a programmer looking into, but am always a bit weary with their workarounds, as I'd previously had them cause more problems then fix it. Here is the solution he is looking to do: The way that I am doing the redirect is fine. The problem is of where to put the code. The issue is that the files are .html files that need to be redirected to the same url with out a .html on them. I can see if I can add that to the 404 redirect page if there is one inside of there and see if that does the trick. That way if there is no page that exists without the .html then it will still be a 404 page. However if it is there then it will work as normal. I will see what I can find and get back. Any help would be greatly appreciated. Thanks, BJ
Intermediate & Advanced SEO | | seointern0 -
How to resolve Duplicate Page Content issue for root domain & index.html?
SEOMoz returns a Duplicate Page Content error for a website's index page, with both domain.com and domain.com/index.html isted seperately. We had a rewrite in the htacess file, but for some reason this has not had an impact and we have since removed it. What's the best way (in an HTML website) to ensure all index.html links are automatically redirected to the root domain and these aren't seen as two separate pages?
Intermediate & Advanced SEO | | ContentWriterMicky0