| |
1. Heritrix | By: | | License: | GNU Library or Lesser General Public License (LGPL) | URL: | http://crawler.archive.org/ | Description: | Heritrix is the Internet Archive's open-source, extensible, web-scale, archival-quality web crawler project.
Heritrix (sometimes spelled heretrix, or misspelled or missaid as heratrix/heritix/ heretix/heratix) is an archaic word for heiress (woman who inherits). Since our crawler seeks to collect and preserve the digital artifacts of our culture for the benefit of future researchers and generations, this name seemed apt.
|
2. WebSPHINX | By: | | License: | Apache Software License | URL: | http://www-2.cs.cmu.edu/~rcm/websphinx/ | Description: | WebSPHINX ( Website-Specific Processors for HTML INformation eXtraction) is a Java class library and interactive development environment for web crawlers. A web crawler (also called a robot or spider) is a program that browses and processes Web pages automatically.
WebSPHINX consists of two parts: the Crawler Workbench and the WebSPHINX class library.
|
3. JSpider | By: | | License: | GNU Library or Lesser General Public License (LGPL) | URL: | http://j-spider.sourceforge.net/ | Description: | JSpider is:
* A highly configurable and customizable Web Spider engine.
* Developed under the LGPL Open Source license
* In 100% pure Java
|
6. WebLech | By: | | License: | MIT License | URL: | http://weblech.sourceforge.net/ | Description: | WebLech is a fully featured web site download/mirror tool in Java, which supports many features required to download websites and emulate standard web-browser behaviour as much as possible. WebLech is multithreaded and will feature a GUI console.
Similar in some aspects to tools such as wget (in recursive retrieval mode), WebSuck or Teleport Pro, WebLech allows you to "spider" a website and to recursively download all the pages on it. You can then browse the site offline for your convenience, or even "mirror" the website and re-publish it yourself. Note that WebLech is not suited to downloading single URLs -- use wget for this kind of thing.
|
7. Arachnid | By: | | License: | GNU General Public License (GPL) | URL: | http://arachnid.sourceforge.net/ | Description: | Arachnid is a Java-based web spider framework. It includes a simple HTML parser object that parses an input stream containing HTML content. Simple Web spiders can be created by sub-classing Arachnid and adding a few lines of code called after each page of a Web site is parsed. Two example spider applications are included to illustrate how to use the framework.
|
8. JoBo | By: | | License: | GNU Library or Lesser General Public License (LGPL) | URL: | http://www.matuschek.net/software/jobo/index.html | Description: | JoBo is a simple program to download complete websites to your local computer. Internally it is basically a web spider. The main advantage to other download tools is that it can automatically fill out forms (e.g. for automated login) and also use cookies for session handling. Compared to other products the GUI seems to be very simple, but the internal features matters ! Do you know any download tool that allows it to login to a web server and download content if that server uses a web forms for login and cookies for session handling ? It also features very flexible rules to limit downloads by URL, size and/or MIME type.
|
|