The-Orizon / nlputils
Utility scripts or libraries for various Natural Language Processing tasks.
☆39Updated 2 years ago
Related projects ⓘ
Alternatives and complementary repositories for nlputils
- an open solution for collecting n-gram Chinese lexicon and n-gram statistics☆74Updated 8 years ago
- A tool for ancient Chinese segmentation.☆53Updated 5 years ago
- 一个轻量且功能全面的中文分词器,帮助学生了解分词器的工作原理。MicroTokenizer: A lightweight Chinese tokenizer designed for educational and research purposes. Provides a…☆147Updated 3 weeks ago
- classic Chinese punctuate experiment with keras using daizhige(殆知阁古代文献藏书) dataset☆33Updated last year
- ☆91Updated this week
- Hanzi Converter for Traditional and Simplified Chinese☆180Updated 4 years ago
- 绝对有趣的中文发音引擎 funny chinese text to speech enginee☆50Updated 11 years ago
- THU Chinese Keyphrase Extraction Toolkit☆124Updated 6 years ago
- Self complemented Word Collocation using MI method which is tested to be effective..基于互信息算法的词语搭配抽取☆29Updated 6 years ago
- 中文分词软件基准测试 | Chinese tokenizer benchmark☆23Updated 6 years ago
- tools for chinese word segmentation and pos tagging written in python☆38Updated 10 years ago
- ☆28Updated 5 months ago
- Corpus creator for Chinese Wikipedia☆42Updated 3 years ago
- Berserker - BERt chineSE woRd toKenizER☆17Updated 5 years ago
- chinese anti semantic word search interface based on dict crawled from online resources, ChineseAntiword,针对中文词语的反义词查询接口☆58Updated 6 years ago
- Chinese Tokenizer module for Python☆16Updated 6 years ago
- worddict crawler and transfer for sougpuinput wordict , 搜狗输入法词库抓取与格式转换☆25Updated 6 years ago
- A Public Corpus for Machine Learning☆44Updated 6 years ago
- Conceptual Keyboard☆28Updated last year
- 中文分词工具评估☆59Updated last year
- This directory contains the training, test, and gold-standard data used in the 2nd International Chinese Word Segmentation Bakeoff. Also …☆66Updated 6 years ago
- 中文生成式预训练模型☆98Updated 4 years ago
- (WIP) My humble contribution to the democratization of the Chinese NLP technology☆46Updated 5 years ago
- Chinese word segmentation algorithm based on entropy(基于熵,无需语料库的中文分词)☆12Updated 6 years ago
- 汉字数据集,包括汉字的相关信息,例如笔画数、部首、拼音、英文释义/同义词等。☆92Updated 4 years ago
- 大规模中文语料☆38Updated 5 years ago
- a shell script to decode sougou cell dict / 一个用来解码搜狗细胞词库的 shell 脚本☆20Updated 11 years ago
- NMT for chinese-english using tensor2tensor☆47Updated 6 years ago
- Conversion of UD_Chinese-GSD to simplified Chinese characters.☆35Updated 6 months ago
- This is a corpus of Chinese abbreviation, including negative full forms.☆189Updated 3 years ago