roliygu / CNKICrawler
A crawler of CNKI. It collects data for NLP and other ML/DL experiment.
☆33Updated 7 years ago
Related projects ⓘ
Alternatives and complementary repositories for CNKICrawler
- sina weibo capture and sentiment classification☆53Updated 8 years ago
- 微博搜索结果爬取工具☆26Updated 9 years ago
- A distributed Sina Weibo Search spider base on Scrapy and Redis.☆143Updated 11 years ago
- Sample of using proxies to crawl baidu search results.☆117Updated 6 years ago
- Chinese Word Similarity Computation based on HowNet☆27Updated 6 years ago
- a text analyzing (match, rewrite, extract) engine (python edition)☆80Updated 7 years ago
- 中国高校更名记录合并☆13Updated 9 years ago
- Get Data Reused☆21Updated 7 years ago
- A command-line manager programed in python, help with managing your local academic papers.☆89Updated 6 years ago
- SNS用户交互学习行为研究☆45Updated 9 years ago
- 微博主题搜索分析,上海租房☆115Updated 8 years ago
- A Python package for pullword.com☆83Updated 4 years ago
- A Scrapy Project 中文门户网站新闻和评论抓取——重启维护工作☆14Updated last year
- tyccl(同义词词林) is a ruby gem that provides friendly functions to analyse similarity between Chinese Words.☆46Updated 10 years ago
- This open source project is a python wrapper for NLPIR.☆82Updated 9 years ago
- Recurrent Neural Networks(GRU) for character-level language models on Chinese, in Python/Theano☆64Updated 7 years ago
- 《知网》中文词语语义相似度算法☆42Updated 11 years ago
- Pure python NLP toolkit☆55Updated 8 years ago
- Yet another Chinese word segmentation package based on character-based tagging heuristics and CRF algorithm☆243Updated 11 years ago
- ☆46Updated 8 years ago
- Scrapy Spider for 新闻联播☆37Updated 9 years ago
- 对红楼梦的各回目进行分类☆35Updated 7 years ago
- 中国爬盟出品的微博备份神器:用于备份新浪微博指定用户全部微博的备份工具☆192Updated 10 years ago
- this is a high performance cuda porting of cbow model of word2vec☆17Updated 10 years ago
- 微信公众号批量抓取器☆55Updated 8 years ago
- 基于Scrapy框架,用于抓取新浪微博数据,主要包括微博内容,评论以及用户信息☆27Updated 9 years ago
- 中文分词程序,可以在没有中文语料库的情况下通过相关性将一段文本中的中文词汇抽取出来☆52Updated 11 years ago