seomoz / g-crawl-py
Gevent Crawling in Python, with Utilities
☆23Updated 10 years ago
Alternatives and similar repositories for g-crawl-py:
Users that are interested in g-crawl-py are comparing it to the libraries listed below
- Scraper built with Scrapy.☆17Updated 8 months ago
- Fast Python Bloom Filter using Mmap☆13Updated 12 years ago
- [UNMAINTAINED] Deploy, run and monitor your Scrapy spiders.☆11Updated 10 years ago
- ☆32Updated last year
- Small set of utilities to simplify writing Scrapy spiders.☆49Updated 9 years ago
- common data interchange format for document processing pipelines that apply natural language processing tools to large streams of text☆35Updated 8 years ago
- ☆33Updated 3 years ago
- Examples of distributed computation using Celery☆33Updated 13 years ago
- Find which links on a web page are pagination links☆29Updated 8 years ago
- Crawlera tools☆26Updated 9 years ago
- Bringing sanity to world of messed-up data☆66Updated 10 years ago
- iCQA - Intelligent Community Question Answering Framework☆31Updated 8 years ago
- collection of modules to build distributed and reliable concurrent systems in Python.☆205Updated 11 years ago
- a SQL-like command line client for elasticsearch☆46Updated 6 years ago
- High performance( 2.5 times to MySQLDb ) Python Mysql Driver, using Python native socket layer. pure C implemented.☆55Updated 4 years ago
- A scrapy pipeline which send items to Elastic Search server☆98Updated 7 years ago
- High Level Kafka Scanner☆19Updated 7 years ago
- This is a Python binding to the tokenizer Ucto. Tokenisation is one of the first step in almost any Natural Language Processing task, yet…☆29Updated 4 months ago
- Python implementation of the Parsley language for extracting structured data from web pages☆92Updated 7 years ago
- Running scrapy spider programmatically.☆47Updated 8 years ago
- Slides to learn a little natural language processing (NLP) with Python. Written in reST with S5/Docutils.☆28Updated 12 years ago
- MongoDB extensions for Scrapy☆44Updated 10 years ago
- Entry for the Third Annual GitHub Data Challenge☆35Updated 10 years ago
- Convert URL's to a normalized unicode format☆67Updated 7 years ago
- A simple and fast search engine☆70Updated 2 years ago
- Software stack used to run Portia spiders in Scrapinghub cloud☆11Updated 5 years ago
- Tornado Web Crawler☆66Updated 12 years ago
- Blog crawler for the blogforever project.☆22Updated 11 years ago
- A Flask extension that adds HTTP based caching to Flask apps☆43Updated 9 years ago
- Modularly extensible semantic metadata validator☆84Updated 9 years ago