Jieba RsThe Jieba Chinese Word Segmentation Implemented in Rust
Nlp4han中文自然语言处理工具集【断句/分词/词性标注/组块/句法分析/语义分析/NER/N元语法/HMM/代词消解/情感分析/拼写检查】
MonpaMONPA 罔拍是一個提供正體中文斷詞、詞性標註以及命名實體辨識的多任務模型
Lac百度NLP:分词,词性标注,命名实体识别,词重要性
Pyhanlp中文分词 词性标注 命名实体识别 依存句法分析 新词发现 关键词短语提取 自动摘要 文本分类聚类 拼音简繁 自然语言处理
JiaguJiagu深度学习自然语言处理工具 知识图谱关系抽取 中文分词 词性标注 命名实体识别 情感分析 新词发现 关键词 文本摘要 文本聚类
G2pcg2pC: A Context-aware Grapheme-to-Phoneme Conversion module for Chinese
SymspellSymSpell: 1 million times faster spelling correction & fuzzy search through Symmetric Delete spelling correction algorithm
GreedycwsSource code for an ACL2017 paper on Chinese word segmentation
ChinesenlpDatasets, SOTA results of every fields of Chinese NLP
JcsegJcseg is a light weight NLP framework developed with Java. Provide CJK and English segmentation based on MMSEG algorithm, With also keywords extraction, key sentence extraction, summary extraction implemented based on TEXTRANK algorithm. Jcseg had a build-in http server and search modules for the latest lucene,solr,elasticsearch
Pkuseg Pythonpkuseg多领域中文分词工具; The pkuseg toolkit for multi-domain Chinese word segmentation
FrisoHigh performance Chinese tokenizer with both GBK and UTF-8 charset support based on MMSEG algorithm developed by ANSI C. Completely based on modular implementation and can be easily embedded in other programs, like: MySQL, PostgreSQL, PHP, etc.
nlpir-analysis-cn-ictclasLucene/Solr Analyzer Plugin. Support MacOS,Linux x86/64,Windows x86/64. It's a maven project, which allows you change the lucene/solr version. //Maven工程,修改Lucene/Solr版本,以兼容相应版本。
berserkerBerserker - BERt chineSE woRd toKenizER
Cross-Domain-CWSCode for IJCAI 2018 paper "Neural Networks Incorporating Unlabeled and Partially-labeled Data for Cross-domain Chinese Word Segmentation"