Parsely / serpextract
Easy extraction of keywords and engines from search engine results pages (SERPs).
☆90Updated 3 years ago
Alternatives and similar repositories for serpextract:
Users that are interested in serpextract are comparing it to the libraries listed below
- Paginating the web☆37Updated 11 years ago
- A python library detect and extract listing data from HTML page.☆108Updated 8 years ago
- Modularly extensible semantic metadata validator☆84Updated 9 years ago
- A Python library for extracting titles, images, descriptions and canonical urls from HTML.☆150Updated 4 years ago
- Python package to detect and return RSS / Atom feeds for a given website. The tool supports major blogging platform including Wordpress, …☆21Updated 3 years ago
- A python implementation of DEPTA☆83Updated 8 years ago
- Find which links on a web page are pagination links☆29Updated 8 years ago
- Automatic Item List Extraction☆87Updated 8 years ago
- CoCrawler is a versatile web crawler built using modern tools and concurrency.☆190Updated 3 years ago
- python library for extracting html microdata☆166Updated last year
- Automatically extracts and normalizes an online article or blog post publication date☆117Updated last year
- Python implementation of the Parsley language for extracting structured data from web pages☆92Updated 7 years ago
- Exporters is an extensible export pipeline library that supports filter, transform and several sources and destinations☆40Updated 11 months ago
- Word analysis, by domain, on the Common Crawl data set for the purpose of finding industry trends☆56Updated last year
- A scrapy extension to store requests and responses information in storage service☆26Updated 3 years ago
- A Scrapy pipeline to categorize items using MonkeyLearn☆38Updated 8 years ago
- A tiny library for Python text normalisation. Useful for ad-hoc text processing.☆151Updated 3 months ago
- Modern robots.txt Parser for Python☆194Updated last year
- Restrict crawl and scraping scope using matchers.☆25Updated 8 years ago
- A component that tries to avoid downloading duplicate content☆27Updated 6 years ago
- ☆59Updated 3 years ago
- [UNMAINTAINED] Deploy, run and monitor your Scrapy spiders.☆11Updated 10 years ago
- A library to interface with the Linkscape API.☆40Updated 6 years ago
- Scrapy middleware which allows to crawl only new content☆80Updated 2 years ago
- An attempt at creating a silver/gold standard dataset for backtesting yesterday & today's content-extractors☆34Updated 10 years ago
- Crawlera tools☆26Updated 9 years ago
- WebAnnotator is a tool for annotating Web pages. WebAnnotator is implemented as a Firefox extension (https://addons.mozilla.org/en-US/fi…☆48Updated 3 years ago
- Frontera backend to guide a crawl using PageRank, HITS or other ranking algorithms based on the link structure of the web graph, even whe…☆55Updated 11 months ago
- Scrapy middleware for the autologin☆37Updated 6 years ago
- ☆50Updated 3 years ago