Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
How to find links to 404 pages?
-
I know that I used to be able to do this, but I can't seem to remember.
One of the sites I am working on has had a lot of pages moving around lately. I am sure some links got lost in the fray that I would like to recover, what is the easiest way to see links going to a domain that are pointing to 404 pages?
-
where is that little button next to my crawl warnings that lets me open urls, or explore links to that url using OSE?
-
Specifically in Open Site Explorer, check out the "Top Pages" tab to see if any of your top linked to pages are returning a 404. This tab is actually the first one I look at when running analysis of a site.
-
Sorry for the delay in my answer.
When you have detected all the 404 of your website, you can use the "Explore URL" search in Siteexplorer. If are still existing backlinks to those pages, Yahoo Siteexplorer will show them.
To be sure I just did a try with an 404 of a new client of mine, and just discovered that one 404 page was linked by a Yale University page... obviously I've just made an 301

-
As familiar as I am with Yahoo SiteExplorer I have never used it to find external links that go to pages that are no longer there. How can I do this with that tool?
-
Hello Spencer,
I recommend two tools
1. Xenu link sleuth (http://home.snafu.de/tilman/xenulink.html#Download)
2. Gsitecrawler ( http://gsitecrawler.com/en/download/)
Both will report all the linked pages throwing a 404 error and other status codes including "forbidden request", "no connection", "no such host" and more.
Hope this helps.
Sameer
-
Did you look into the Google Webmaster Tools already? There you can see them as well - of course not all. But you have to check from time to time - they don't show up all together. If you fix some perpaps some more will come up ...
-
Hi Spencer:
I don't know if this qualifies as the easiest way , but it ranks right up there:
-
You can use Open Site Explorer, but i suggest you to widen the discovery using also Yahoo! SiteExplorer
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
How can I stop a tracking link from being indexed while still passing link equity?
I have a marketing campaign landing page and it uses a tracking URL to track clicks. The tracking links look something like this: http://this-is-the-origin-url.com/clkn/http/destination-url.com/ The problem is that Google is indexing these links as pages in the SERPs. Of course when they get indexed and then clicked, they show a 400 error because the /clkn/ link doesn't represent an actual page with content on it. The tracking link is set up to instantly 301 redirect to http://destination-url.com. Right now my dev team has blocked these links from crawlers by adding Disallow: /clkn/ in the robots.txt file, however, this blocks the flow of link equity to the destination page. How can I stop these links from being indexed without blocking the flow of link equity to the destination URL?
Technical SEO | | UnbounceVan0 -
Blog Page Titles - Page 1, Page 2 etc.
Hi All, I have a couple of crawl errors coming up in MOZ that I am trying to fix. They are duplicate page title issues with my blog area. For example we have a URL of www.ourwebsite.com/blog/page/1 and as we have quite a few blog posts they get put onto another page, example www.ourwebsite.com/blog/page/2 both of these urls have the same heading, title, meta description etc. I was just wondering if this was an actual SEO problem or not and if there is a way to fix it. I am using Wordpress for reference but I can't see anywhere to access the settings of these pages. Thanks
Technical SEO | | O2C0 -
Can you use Screaming Frog to find all instances of relative or absolute linking?
My client wants to pull every instance of an absolute URL on their site so that they can update them for an upcoming migration to HTTPS (the majority of the site uses relative linking). Is there a way to use the extraction tool in Screaming Frog to crawl one page at a time and extract every occurrence of _href="http://" _? I have gone back and forth between using an x-path extractor as well as a regex and have had no luck with either. Ex. X-path: //*[starts-with(@href, “http://”)][1] Ex. Regex: href=\”//
Technical SEO | | Merkle-Impaqt0 -
Why is Google Webmaster Tools showing 404 Page Not Found Errors for web pages that don't have anything to do with my site?
I am currently working on a small site with approx 50 web pages. In the crawl error section in WMT Google has highlighted over 10,000 page not found errors for pages that have nothing to do with my site. Anyone come across this before?
Technical SEO | | Pete40 -
How to inform Google to remove 404 Pages of my website?
Hi, I want to remove more than 6,000 pages of my website because of bad keywords, I am going to drop all these pages and making them ‘404’ I want to know how can I inform google that these pages does not exists so please don’t send me traffic from those bad keywords? Also want to know can I use disavow tool of google website to exclude these 6,000 pages of my own website?
Technical SEO | | renukishor4 -
Is the Authority of Individual Pages Diluted When You Add New Pages?
I was wondering if the authority of individual pages is diluted when you add new pages (in Google's view). Suppose your site had 100 pages and you added 100 new pages (without getting any new links). Would the average authority of the original pages significantly decrease and result in a drop in search traffic to the original pages? Do you worry that adding more pages will hurt pages that were previously published?
Technical SEO | | Charlessipe0 -
Footer Links with same anchor text on all pages
We have different websites targeted at the different services our company provides. (e.g. For our document storage services, we have www.ukdocumentstorage.com. For document management, we have www.document-management-solutions.co.uk). If we take the storage site for example, every single page has a link in the footer to our document management site, with the anchor text 'Cleardata Document Management' SEOMoz is telling me that these are seen as external links (as they are on a different URL's), and I'm just clarifying that would this be a major possible factor in the website not ranking highly? How should I rectify this issue?
Technical SEO | | janc0 -
Thoughts about stub pages - 200 & noindex ok, or 404?
With large database/template driven websites it is often possible to get a lot of pages with no content on them. What are the current thoughts regarding these pages with no content, options; Return a 200 header code with noindex meta tag Return a 404 page & header code Something else? Thanks
Technical SEO | | slingshot0