Link rot

From Wikipedia, the free encyclopedia

Link rot (or linkrot), also known as link death or link breaking, is an informal term for the process by which hyperlinks (either on individual websites or the Internet in general) point to web pages, servers or other resources that have become permanently unavailable. The phrase also describes the effects of failing to update out-of-date web pages that clutter search engine results. A link that does not work any more is called a broken link, dead link, or dangling link.

Causes

A link may become broken for several reasons. The simplest and most common reason is that the web page it links to doesn't exist anymore. The most common result of a dead link is a 404 error, which indicates that the web server responded, but the specific page could not be found.

Some news sites contribute to the problem of link rot by keeping only recent news articles freely accessible, then removing them or moving them to a paid subscription area. This causes a heavy loss of supporting links in sites discussing newsworthy events and using news sites as references.[citation needed]

Another type of dead link occurs when the server that hosts the target page stops working or relocates to a new domain name. In this case the browser may return a DNS error, or it may display a site unrelated to the content sought. The latter can occur when a domain name is allowed to lapse and is subsequently reregistered by another party. Domain names acquired in this manner are attractive to those who wish to take advantage of the stream of unsuspecting surfers that will inflate hit counters and PageRanking.

A link might also be broken because of some form of blocking such as content filters or firewalls. Dead links commonplace on the Internet can also occur on the authoring side, when website content is assembled, copied, or deployed without properly verifying the targets, or simply not kept up to date. Dead links can also occur when a website without Clean URLs is "re-organized".[1]

Prevalence

The 404 "Not Found" response is familiar to even the occasional web user. A number of studies have examined the prevalence of link rot on the web, in academic literature, and in digital libraries. In a 2003 experiment, Fetterly et al. discovered that about one link out of every 200 disappeared each week from the internet. McCown et al. (2005) discovered that half of the URLs cited in D-Lib Magazine articles were no longer accessible 10 years after publication, and other studies have shown link rot in academic literature to be even worse (Spinellis, 2003, Lawrence et al., 2001). Nelson and Allen (2002) examined link rot in digital libraries and found that about 3% of the objects were no longer accessible after one year.

Discovering

Detecting link rot for a given URL is difficult using automated methods. If a URL is accessed and returns an HTTP 200 (OK) response, it may be considered accessible, but the contents of the page may have changed and may no longer be relevant. Some web servers also return a soft 404, a page returned with a 200 (OK) response (instead of a 404 that indicates the URL is no longer accessible). Bar-Yossef et al. (2004) [2] developed a heuristic for automatically discovering soft 404s.

Combating

Due to the unprofessional image that dead links bring to both sites linking and linked to, there are multiple solutions that are available to tackle them: some working to prevent them in the first place, and others trying to resolve them when they have occurred. There are several tools that have been developed to help combat link rot.

Server side

  • Avoiding unmanaged hyperlink collections
  • Avoiding links to pages deep in a website ("deep linking")
  • Using redirection mechanisms (e.g. "301: Moved Permanently") to automatically refer browsers and crawlers to the new location of a URL
  • Content management systems may offer builtin solutions to the management of links, e.g. links are updated when content is changed or moved on the site.
  • WordPress guards against link rot by replacing non-canonical URLs with their canonical versions.[3]
  • IBM's Peridot attempts to automatically fix broken links.
  • Permalinking stops broken links by guaranteeing that the content will not move for the foreseeable future. Another form of permalinking is linking to a permalink that then redirects to the actual content, ensuring that even though the real content may be moved etc., links pointing to the resources stay intact.

User side

  • The Linkgraph widget gets the URL of the correct page based upon the old broken URL by using historical location information.
  • The Google 404 Widget employs Google technology to 'guess' the correct URL, and also provides the user a Google search box to find the correct page.
  • When a user receives a 404 response, the Google Toolbar attempts to assist the user in finding the missing page.[4]
  • Deadurl.com[5] gathers and ranks alternate urls for a broken link using Google Cache, the Internet Archive, and user submissions.[6] Typing deadurl.com/ left of a broken link in the browser's address bar and pressing enter loads a ranked list of alternate urls, or (depending on user preference) immediately forwards to the best one.[7]

Web archiving

To combat link rot, web archivists are actively engaged in collecting the Web or particular portions of the Web and ensuring the collection is preserved in an archive, such as an archive site, for future researchers, historians, and the public. The largest web archiving organization is the Internet Archive[citation needed], whose goal is to maintain an archive of the entire Web, taking periodic snapshots of pages that can then be accessed for free via the Wayback Machine and without registration many years later simply by typing in the URL, or automatically by using browser extensions.[8] National libraries, national archives and various consortia of organizations are also involved in archiving culturally important Web content.

Individuals may also use a number of tools that allow them to archive web resources that may go missing in the future:

  • WebCite, a tool specifically for scholarly authors, journal editors and publishers to permanently archive "on-demand" and retrieve cited Internet references (Eysenbach and Trudel, 2005).
  • Archive-It, a subscription service that allows institutions to build, manage and search their own web archive
  • Some social bookmarking websites such as peeep.us,[9] or evernote.com[10] allow users to make online clones of any web page on the internet, creating a copy at an independent url which remains online even if the original page goes down.
  • Google keeps a text-based cache (temporary copy) of the pages it has crawled, which can be used to read the information of recently removed pages. However, unlike in archiving services, cached pages are not stored permanently.
  • The WayBack Machine, at the Internet Archive,[11] is a free website that archives old web pages. It does not archive websites whose owners have stated they do not want their website archived.
  • archive.is, a personal WayBack Machine, which saves text and screenshot of a webpage and keeps it online even if the original page is rot. In most aspects it is a WebCite analog, but also saves images and can save pages from Web 2.0 sites (like Twitter).

Authors citing URLs

A number of studies have shown how widespread link rot is in academic literature (see below). Authors of scholarly publications have also developed best practices for combating link rot in their work:

  • Avoiding URL citations that point to resources on a researcher's personal home page (McCown et al., 2005)
  • Using Persistent Uniform Resource Locators (PURLs) and digital object identifiers (DOIs) whenever possible
  • Using web archiving services (e.g. WebCite) to permanently archive and retrieve cited Internet references (Eysenbach and Trudel, 2005).

See also

Further reading

Link rot on the Web

In academic literature

In digital libraries

  • Nelson, Michael L.; Allen, B. Danette (2002). "Object Persistence and Availability in Digital Libraries". D-Lib Magazine 8 (1). doi:10.1045/january2002-nelson. 

References

  1. Tim Berners-Lee (1998). ""Cool URIs don't change"". Retrieved 2013-10-07. 
  2. Bar-Yossef, Ziv; Broder, Andrei Z.; Kumar, Ravi; Tomkins, Andrew (2004). "Sic transit gloria telae: towards an understanding of the web's decay". Proceedings of the 13th conference on World Wide Web - WWW '04. p. 328. doi:10.1145/988672.988716. ISBN 158113844X. 
  3. Rønn-Jensen, Jesper (2007-10-05). "Software Eliminates User Errors And Linkrot". Justaddwater.dk. Retrieved 2007-10-05. 
  4. Mueller, John (2007-12-14). "FYI on Google Toolbar's Latest Features". Google Webmaster Central Blog. Retrieved 2008-07-09. 
  5. DeadURL.com. "Make Any Dead URL Work Again". DeadURL.com. Retrieved 2013-10-07. 
  6. "DeadURL.com". Retrieved 2011-03-17. "DeadURL.com gathers as many backup links as possible for each dead url, via Google cache, Archive.org, and user submissions." 
  7. "DeadURL.com". Retrieved 2011-03-17. "Just type deadurl.com/ in front of a link that doesn't work, and hit Enter." 
  8. Lazar Kovacevic. "404-Error ? 1.4". Addons.mozilla.org. Retrieved 2013-10-07. 
  9. "Peeep.us". Peeep.us. 2009-09-30. Retrieved 2013-10-07. 
  10. "Remember everything with Evernote, Skitch and our other great apps". Evernote. Retrieved 2013-10-07. 
  11. "Internet Archive: Digital Library of Free Books, Movies, Music & Wayback Machine". Archive.org. 2001-03-10. Retrieved 2013-10-07. 

External links

This article is issued from Wikipedia. The text is available under the Creative Commons Attribution/Share Alike; additional terms may apply for the media files.