Nutch

From Wikipedia, the free encyclopedia
Apache Nutch

Developer(s) Apache Software Foundation
Stable release 1.5.1 and 2.1 / October 5, 2012 (2012-10-05)
Development status Active
Written in Java
Operating system Cross-platform
Type Search Engine
License Apache License 2.0
Website nutch.apache.org

Nutch is an effort to build an open source web search engine based on Lucene and Java for the search and index component.

Features

Nutch is coded entirely in the Java programming language, but data is written in language-independent formats. It has a highly modular architecture, allowing developers to create plug-ins for media-type parsing, data retrieval, querying and clustering.

The fetcher ("robot" or "web crawler") has been written from scratch specifically for this project.

History

Nutch originated with Doug Cutting, creator of both Lucene and Hadoop, and Mike Cafarella.

In June, 2003, a successful 100-million-page demonstration system was developed. To meet the multimachine processing needs of the crawl and index tasks, the Nutch project has also implemented a MapReduce facility and a distributed file system. The two facilities have been spun out into their own subproject, called Hadoop.

In January, 2005, Nutch joined the Apache Incubator, from which it graduated to become a subproject of Lucene in June of that same year. Since April, 2010, Nutch has been considered an independent, top level project of the Apache Software Foundation.[1]

Advantages

Advantages of Nutch over a simple fetcher include[2]

  • highly scalable and relatively feature rich crawler
  • features like politeness which obeys robots.txt rules
  • robust and scalable - Nutch can run on a cluster of up to 100 machines
  • quality - crawling can be biased to fetch "important" pages first

Scalability

IBM Research studied the performance[3] of Nutch/Lucene as part of its Commercial Scale Out (CSO) project.[4] Their findings were that a scale-out system, such as Nutch/Lucene, could achieve a performance level on a cluster of blades that was not achievable on any scale-up computer such as the Power5.

The ClueWeb09 dataset (used in e.g. TREC) was gathered using Nutch, with an average speed of 755.31 documents per second.[5]

Related projects

  • Hadoop - Java framework that supports distributed applications running on large clusters

Search engines built with Nutch

See also

  • Faceted Search
  • Information Extraction
  • Enterprise Search

References

Bibliography

External links

This article is issued from Wikipedia. The text is available under the Creative Commons Attribution/Share Alike; additional terms may apply for the media files.