ueda-keisuke / CC-CEDICT-MeCab
CC-CEDICT-MeCab is a MeCab dictionary for Chinese (Mandarin) text segmentation
☆11Updated 4 years ago
Alternatives and similar repositories for CC-CEDICT-MeCab:
Users that are interested in CC-CEDICT-MeCab are comparing it to the libraries listed below
- ☆24Updated 4 months ago
- 🦞 Rust library of natural language dictionaries using character-wise double-array tries.☆30Updated 2 months ago
- Japanese tokenizer for rust☆35Updated 5 years ago
- Yada is a yet another double-array trie library aiming for fast search and compact data representation.☆35Updated last year
- A tool for visualizing the internal structures of morphological analyzer Sudachi☆17Updated 2 years ago
- sqlite3 fts5 mecab☆20Updated 5 years ago
- Rakuten MA (Python version)☆22Updated 7 years ago
- python版日本語意味役割付与システム(ASA)☆23Updated 2 years ago
- Japanese synonym library☆53Updated 3 years ago
- Solr / Elasticsearch Synonym mapping file for Japanese web documents using results of NEologd☆39Updated 9 years ago
- Yet another sentence-level tokenizer for the Japanese text☆22Updated 2 years ago
- A small version of UniDic for easy pip installs.☆43Updated 4 years ago
- A Japanese Morphological Analyzer written in pure Rust☆26Updated 5 years ago
- Tokenizer POS-tagger Lemmatizer and Dependency-parser for modern and contemporary Japanese☆35Updated 4 months ago
- A lidera japanese tokenizer wrapper for javascript and typescript☆14Updated 3 years ago
- IPAdic packaged for easy use from Python.☆25Updated 3 years ago
- CaboCha wrapper for Python3☆47Updated 6 years ago
- Japanese text preprocessor for Text-to-Speech applications (OpenJTalk rewrite in rust language)☆39Updated this week
- 全国書誌データから作成した振り仮名のデータセット☆26Updated 3 years ago
- 日本語文法誤り訂正ツール☆28Updated 2 years ago
- Japanese data from the Google UDT 2.0.☆37Updated 4 months ago
- Safe Rust bindings for mecab a part-of-speech and morphological analyzer library☆62Updated last year
- Accommodation Search Dialog Corpus (宿泊施設探索対話コーパス)☆25Updated last year
- A comparison tool of Japanese tokenizers☆121Updated 9 months ago
- AllenNLP integration for Shiba: Japanese CANINE model☆12Updated 3 years ago
- ☆71Updated 2 years ago
- Word List by Semantic Principles (WLSP): “It is a collection of words classified and arranged by their meanings”☆53Updated 4 years ago
- Finding all pairs of similar documents time- and memory-efficiently☆60Updated 2 weeks ago
- 🛥 Vaporetto: Very accelerated pointwise prediction based tokenizer☆235Updated 2 weeks ago
- A paraphrase database for Japanese text simplification☆32Updated 8 years ago