Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Page speed tools
-
Working on reducing page load time, since that is one of the ranking factors that Google uses. I've been using Page Speed FireFox plugin (requires FireBug), which is free. Pretty happy with it but wondering if others have pointers to good tools for this task.
Thanks...
-
If you're using wordpress or a CMS of some type, look for posts and tools related to that particular software. The SmushIt tool is available as a wordpress plugin, and it will compress pictures on upload, and will compress (on demand) pictures in the media library.
I use Irfanview for batch resizing images (in the same directory).
-
Have you considered using a CDN? This will cache your site in multiple locations so end users site experience will be better.
-
YSlow for firebug is very useful, don't forget to run SmushIt (Tools in YSlow) to have optimized images. Spites + GZip + 1 single packed CSS + 1 single packed JS + expires headers are very effective. Use media domain if you have a lot of pictures on the page.
See the real load time in the Net tab of firebug as Gary said.
-
I have gone through this same process recently.
FireBug's NET tab, the page speed plugin and Yahoo YSlow plugins were the way to go.
I had a page load time of about 10 seconds before optimising. I reduced this to about 6 seconds by doing the following:
- Using different domain names (e.g. static.mysite.com, static2.mysite.com, static3.mysite.com) to increase parallel downloads of CSS/images etc.
- Gzip
- Reducing size of all images
I then managed to decrease this further to about 1.5 seconds by changing server hosts.
-
decrease size of your pictures by 30% in most cases you wont even see the difference, but it helps a lot with load time.
Also check if you have any errors: http://validator.w3.org/ and http://jigsaw.w3.org/css-validator/
Good Luck.
-
yslow for firefox
and good old Google Webmaster Tools
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
On-page SEO
This is a question for the organic SEO experts, once you added the main keyword that you want to rank for in the homepage title, meta title plus meta description, perhaps once or twice in the text on the homepage. How often do you then write it in the content marketing, say blog posts, we want to rank higher on Google for "SEO agencies Cardiff" however if you mention this in the blog posts too much say once a week, this could lead to over optimisation issues?
On-Page Optimization | | sarahwalsh1 -
site speed
i use mid-quality pic and... but my site speed is low
On-Page Optimization | | zlbvasgabc
any suggestion?
my site is:
https://bandolini.ir/0 -
Is it better to keep a glossary or terms on one page or break it up into multiple pages?
We have a very large glossary of over 1000 industry terms on our site with links to reference material, embedded video, etc. Is it better for SEO purposes to keep this on one page or should we break it up into multiple pages, a different page for each letter for example? Thanks.
On-Page Optimization | | KenW0 -
What to do with multiple forms and thank you pages
Hi Everyone, I'm wondering what to do with form and thank you pages. I asked a question a long time ago about the contact page as noindex and was told by people that its better to leave it and write content for it and thats what we did. Now I have a client that does self storage and they have 4 locations and each location has a reservation page with a basic form but no content. Each page also redirects to a thank you page with tracking codes. There's a total of 6 "thank you" pages with different codes (this was done by yellow book). 4 "reserve your storage pages", 2 pages to pay for storage with iframes to 3rd party payment portals. I was told to noindex these pages but I'm not sure so I'm asking here. I was also told to nofollow and remove them from sitemaps. Thanks Aron
On-Page Optimization | | aronwp0 -
Noindex child pages (whose content is included on parent pages)?
I'm sorry if there have been questions close to this before... I've using WordPress less like a blogging platform and more like a CMS for years now... For content management purposes we organize a lot of content around Parent/Child page (and custom-post-type) relationships; the Child pages are included as tabbed content on the Parent page. Should I be noindexing these child pages, since their content is already on the site, in full, on their Parent pages (ie. duplicate content)? Or does it not matter, since the crawlers may not go to all of the tabbed content? None of the pages have shown up in Moz's "High Priority Issues" as duplicate content but it still seems like I'm making the Parent pages suffer needlessly... Anything obvious I'm not taking into consideration? By the by, this is my first post here @ Moz, which I'm loving; this site and the forums are such a great resource! Anyways, thanks in advance!
On-Page Optimization | | rsigg0 -
Duplicate Content on Event Pages
My client has a pretty popular service of event listings and, in hope of gathering more events, they opened up the platform to allow users to add events. This works really well for them and they are able to garner a lot more events this way. The major problem I'm finding is that many event coordinators and site owners will take the copy from their website and copy and paste it, duplicating a lot of the content. We have editor picks that contain a lot of unique content but the duplicate content scares me. It hasn't hurt our page ranking (we have a page ranking of 7) but I'm wondering if this is something that we should address. We don't have the manpower to eliminate all the duplication but if we cut down the duplication would we experience a significant advantage over people posting the same event?
On-Page Optimization | | mattdinbrooklyn0 -
Category page canonical tag
I know this question has been asked a few times on here but I'm looking for very specific advice. Currently when you go to a category, say http://www.bronterose.co.uk/range.html, a canonical tag is added to the head of the page. There are plenty of "variant" pages which carry the same tag, for example: /range.html?p=2
On-Page Optimization | | crichardson9
/range.html?p=3
/range.html?dir=asc&order=price
/range.html?dir=asc&limit=all&order=price Is it wise to push the "link juice" for each of these variant pages to the top level page? Or should each variant page have its own unique canonical tag? After reading many blog posts, guides and papers I'm truly confused! Any general guidance or recommendations would be much appreciated. Chris.1 -
Avoiding "Duplicate Page Title" and "Duplicate Page Content" - Best Practices?
We have a website with a searchable database of recipes. You can search the database using an online form with dropdown options for: Course (starter, main, salad, etc)
On-Page Optimization | | smaavie
Cooking Method (fry, bake, boil, steam, etc)
Preparation Time (Under 30 min, 30min to 1 hour, Over 1 hour) Here are some examples of how URLs may look when searching for a recipe: find-a-recipe.php?course=starter
find-a-recipe.php?course=main&preperation-time=30min+to+1+hour
find-a-recipe.php?cooking-method=fry&preperation-time=over+1+hour There is also pagination of search results, so the URL could also have the variable "start", e.g. find-a-recipe.php?course=salad&start=30 There can be any combination of these variables, meaning there are hundreds of possible search results URL variations. This all works well on the site, however it gives multiple "Duplicate Page Title" and "Duplicate Page Content" errors when crawled by SEOmoz. I've seached online and found several possible solutions for this, such as: Setting canonical tag Adding these URL variables to Google Webmasters to tell Google to ignore them Change the Title tag in the head dynamically based on what URL variables are present However I am not sure which of these would be best. As far as I can tell the canonical tag should be used when you have the same page available at two seperate URLs, but this isn't the case here as the search results are always different. Adding these URL variables to Google webmasters won't fix the problem in other search engines, and will presumably continue to get these errors in our SEOmoz crawl reports. Changing the title tag each time can lead to very long title tags, and it doesn't address the problem of duplicate page content. I had hoped there would be a standard solution for problems like this, as I imagine others will have come across this before, but I cannot find the ideal solution. Any help would be much appreciated. Kind Regards5