The-Orizon / nlputilsLinks
Utility scripts or libraries for various Natural Language Processing tasks.
☆39Updated 3 years ago
Alternatives and similar repositories for nlputils
Users that are interested in nlputils are comparing it to the libraries listed below
Sorting:
- ☆39Updated last year
- Berserker - BERt chineSE woRd toKenizER☆16Updated 6 years ago
- an open solution for collecting n-gram Chinese lexicon and n-gram statistics☆74Updated 9 years ago
- Hanzi Converter for Traditional and Simplified Chinese☆188Updated 5 years ago
- Corpus creator for Chinese Wikipedia☆41Updated 4 years ago
- Chinese word segmentation algorithm based on entropy(基于熵,无需语料库的中文分词)☆11Updated 7 years ago
- convert sogou input dict ( .scel file ) to mmseg(coreseek) dict☆96Updated 11 years ago
- THU Chinese Keyphrase Extraction Toolkit☆124Updated 7 years ago
- 一个轻量且功能全面的中文分词器,帮助学生了解分词器的工作原理。MicroTokenizer: A lightweight Chinese tokenizer designed for educational and research purposes. Provides a…☆154Updated 8 months ago
- 绝对有趣的中文发音引擎 funny chinese text to speech enginee☆51Updated 11 years ago
- ☆93Updated last week
- Chinese morphological analysis with Word Segment and POS Tagging data for MeCab☆161Updated 7 years ago
- ☆66Updated 8 years ago
- 中文分词软件基准测试 | Chinese tokenizer benchmark☆24Updated 6 years ago
- Chinese Word Similarity Computation based on HowNet☆27Updated 7 years ago
- A tool for ancient Chinese segmentation.☆53Updated 6 years ago
- This is a corpus of Chinese abbreviation, including negative full forms.☆196Updated 3 years ago
- classic Chinese punctuate experiment with keras using daizhige(殆知阁古代文献藏书) dataset☆35Updated 2 years ago
- 提取中文的偏旁部首和拼音(一些生僻字的拼音没有补全,待优化)☆43Updated 7 years ago
- 下载搜狗、百度、QQ输入法的词库文件的 python 爬虫,可用于构建不同行业的词汇库☆115Updated 7 years ago
- auto generate chinese words in huge text.☆91Updated 10 years ago
- Chinese word segmentation module of LTP☆46Updated 9 years ago
- (WIP) My humble contribution to the democratization of the Chinese NLP technology☆46Updated 6 years ago
- Conceptual Keyboard☆28Updated 2 years ago
- 古典中文語料庫☆285Updated 3 years ago
- 用结巴(Jieba)轻松实现细粒度分词☆16Updated 5 years ago
- This is a pre-trained LSTM model. This model can help you to segment unpunctuated historical Chinese texts. 這是基於 LSTM 的預訓練模型。此模型可幫助您為漢語古文…☆26Updated 3 years ago
- 离线端阅读理解应用 QA for mobile, Android & iPhone☆60Updated 2 years ago
- Chinese stopwords collection☆137Updated 5 years ago
- Somiao Pinyin: Train your own Chinese Input Method with Seq2seq Model 搜喵拼音输入法☆270Updated 5 years ago