Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Reasons Why Our Website Pages Randomly Loads Without Content
-
I know this is not a marketing question but this community is very dev savvy so I'm hoping someone can help me. At random times we're finding that our website pages load without the main body content. The header, footer and navigation loads just fine. If you refresh, it's fine but that's not a solution.
- Happens on Chrome, IE and Firefox, testing with multiple browser versions
- Happens across various page types - but seems to be only the main content section/container
- Happens while on the company network, as well as externally
- Happens after deleting cookies, temporary internet files and restarting computer
- We are using a CMS that is virtually unheard of - Bridgeline/Iapps
- Codebase is .net
Our IT/Dev group keeps pushing back, blaming it on cookies or Chrome plugins because they apparently are unable to "recreate the problem". This has been going on for months and it's a terrible experience for the user to have. It's also not great when landing PPC visitors on pages that load with no content. If anyone has ideas as to why this may be happening I would really appreciate it.
I'm not sure if links are allowed, by today the issue happened on this page serversdirect.com/dm/geek-biz
Linking to an image example below
-
Thanks Gazzerman - This helps as well. Yes, I agree the site definitely needs SEO attention. It was not until recently that a more experience team was brought on to 'fix the site'.
-
I took one look at the code and was amazed by how much is inline. Most of it should be in js and css files. I noticed the solutions pages many of them don't even have title tags! What's going on there? That needs seo attention and is a big no no for coding. This leads me to believe the code needs very close inspection. It surly cant take that long to strip a whole bunch of that code out of each page or template.
If a page is slow to load or pulls in external js files as well you can have scripts try to execute before the required code has loaded. This is probably what is happening some of the time and would explain the inconsistent nature of it.
Run the site through a bunch of validators. I am sure you will find a bunch of stuff.
-
Oddly enough it only seemed to happen on the /Servers page - for me at least. The first time I loaded the page, I got a broken image and an X-Out/Close box. That may help you hone in on the problem.
-
Travis - Thanks for your response. This gives me a better idea of where to start looking. Honestly this site was very poorly developed in my opinion. It was before my time and much of it outsourced to junior teams overseas, which is evident when looking at the source code. But I am stuck with it for now.
-
Found the site. I'm able to recreate the problem. Check the image attachment.
I ran it through tools.pingdom.com, a chrome extension called cast_sender.js isn't loading. But something makes me think there's an issue with the CSS that only crops up from time to time, like the JS blocks it every once in a while. There's both a Global.css and a Global.js in the head. There are also folder levels called /Script%20Library/ and /Style%20Library/.
There's a ton of JavaScript on the site, which I don't have time to get into. And honestly, I'm more of a WordPress guy. I managed to make the page load fail a few times out of a dozen or so tries. I think something in the JavaScript is causing the issue.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Reason for robots.txt file blocking products on category pages?
Hi I have a website with thosands of products. On the category pages, all the products are linked to with the code “?cgid” in the URL. But “?cgid” is also blocked in the robots.txt file for some reason. So I'm thinking it's stopping all my products getting crawled by Google. Am I right here? Is there any reason why a website would want to limit so many URL's? I'm only here a week and the sites getting great traffic, so don't want to go breaking it!!! Thanks
Web Design | | Frankie-BTDublin0 -
How to prevent development website subdomain from being indexed?
Hello awesome MOZ Community! Our development team uses a sub-domain "dev.example.com" for our SEO clients' websites. This allows changes to be made to the dev site (U/X changes, forms testing, etc.) for client approval and testing. An embarrassing discovery was made. Naturally, when you run a "site:example.com" the "dev.example.com" is being indexed. We don't want our clients websites to get penalized or lose killer SERPs because of duplicate content. The solution that is being implemented is to edit the robots.txt file and block the dev site from being indexed by search engines. My questions is, does anyone in the MOZ Community disagree with this solution? Can you recommend another solution? Would you advise against using the sub-domain "dev." for live and ongoing development websites? Thanks!
Web Design | | SproutDigital0 -
Problems preventing Wordpress attachment pages from being indexed and from being seen as duplicate content.
Hi According to a Moz Crawl, it looks like the Wordpress attachment pages from all image uploads are being indexed and seen as duplicate content..or..is it the Yoast sitemap causing it? I see 2 options in SEO Yoast: Redirect attachment URLs to parent post URL. Media...Meta Robots: noindex, follow I set it to (1) initially which didn't resolve the problem. Then I set it to option (2) so that all images won't be indexed but search engines would still associate those images with their relevant posts and pages. However, I understand what both of these options (1) and (2) mean, but because I chose option 2, will that mean all of the images on the website won't stand a chance of being indexed in search engines and Google Images etc? As far as duplicate content goes, search engines can get confused and there are 2 ways for search engines
Web Design | | SEOguy1
to reach the correct page content destination. But when eg Google makes the wrong choice a portion of traffic drops off (is lost hence errors) which then leaves the searcher frustrated, and this affects the seo and ranking of the site which worsens with time. My goal here is - I would like all of the web images to be indexed by Google, and for all of the image attachment pages to not be indexed at all (Moz shows the image attachment pages as duplicates and the referring site causing this is the sitemap url which Yoast creates) ; that sitemap url has been submitted to the search engines already and I will resubmit once I can resolve the attachment pages issues.. Please can you advise. Thanks.0 -
One Page Guide vs. Multiple Individual Pages
Howdy, Mozzers! I am having a battle with my inner-self regarding how to structure a resources section for our website. We're building out several pieces of content that are meant to be educational for our clients and I'm having trouble deciding how to layout the content structure. We could either layout all eight short sections on a single page, or create individual pages for each section. The goal is obviously to attract new potential clients by targeting these terms that they may be searching for in an information gathering stage. Here's my dilemma...
Web Design | | jpretz
With the single page guide, it would be nice because it will have a lot of content (and of course, keywords) to be picked up by the SERPS but I worry that it is going to be a bit crammed (because of eight sections) for the user. The individual pages would be much better organized and you can target more specific keywords, but I worry that it may get flagged for light content as some pages may have as little as a 150 word description. I have always been mindful of writing copy for searchers over spiders, but now I'm at a more technical crossroads as far as potentially getting dinged for not having robust content on each page. Here's where you come in...
What do you think is the better of the two options? I like the idea of having the multiple pages because of the ability to hone-in on a keyword and the clean, organized feel, but I worry about the lack of content (and possibly losing out on long-tail opportunities). I'd love to hear your thoughts. Please and thank you. Ready annnnnnnnnnnnd GO!0 -
Link colour on page?
I always thought that the link colour has to be different from text colour? I have come across a site http://www.printandpackaging.co.uk/ and it has made me question this belief, they seem to only have bolded the link which would be very nice if this is fine.
Web Design | | BobAnderson0 -
Best layout pages for SEO
Dear all, what would be the ideal layout of a webpage for SEO? How would a homepage and landingspage look like? Thanks in advance! Best regards, Ben
Web Design | | HMK-NL0 -
Indexing Dynamic Pages
Hi, I am having an issues among others, regarding indexing dynamic pages. Our website, www.me-by-melia, was just put live and I am concerned the bottom naviagtion pages (http://www.me-by-melia.com/#store, http://www.me-by-melia.com/#facebook, etc) will not be indexed and create duplicate pages. Also, when you open these pages in a new tab, it takes you to homepage. The website was created in HTML5. Please advise.
Web Design | | Melia0 -
Combining web pages and it's affects on SEO?
We are looking into amending a website we are working on to try and combine 2 or 3 current pages onto one page. This site is similar to an estate agents site and currently has images, map, floor plan sub pages etc. Can anyone tell me, if we were to combine these pages and include the above details on one page, how that would affect the current search engine rankings?
Web Design | | SoundinTheory0