Norconex / crawlersLinks
Norconex Crawlers (or spiders) are flexible web and filesystem crawlers for collecting, parsing, and manipulating data from the web or filesystem to various data repositories such as search engines.
☆194Updated 3 weeks ago
Alternatives and similar repositories for crawlers
Users that are interested in crawlers are comparing it to the libraries listed below
Sorting:
- A set of reusable Java components that implement functionality common to any web crawler☆246Updated last month
- A scalable, mature and versatile web crawler based on Apache Storm☆939Updated this week
- Carrot2: Text Clustering Algorithms and Applications