搜索资源列表
百度分词词库
- 据说是百度以前用的中文分词词典,希望对大家有一点帮助哈,快下快下-allegedly Baidu before the Chinese word dictionaries, we hope to have a bit of help to Kazakhstan, where fast under fast!
pymmseg-cpp-win32-1.0.1.tar.gz
- 基于python的中文分词程序,易用性高,可以作为接口直接在python程序中使用,Python-based Chinese word segmentation process, ease of use high, can be used as interface directly in the python program to use
Chinese-Segmentation.rar
- 自己编写的中文分词源程序,用vc++编写,附有完整的文档,以及标准的分词数据库,I have written the source code of the Chinese word segmentation, using vc++ to prepare, with complete documentation, as well as sub-standard speech database
中文分词算法
- 本程序使用给出的字典进行学习并对训练语料进行分词处理,采用C语言编写,高效易懂!
RMM
- 基于rmm算法(逆向最大匹配)实现的中文分词系统,具体内容是一个mfc工程文件。-Rmm-based algorithm (reverse maximum matching) to achieve the Chinese word segmentation system, specific content is a mfc project file.
Source_Code
- 中文分词系统分析库,可以对UTF8表述的中文文字进行分词操作, 支持外挂的词库和手动干预(回调方式)的加权处理。 可用于搜索引擎的基础分词开发-Chinese word segmentation system analysis libraries, can be expressed UTF8 Chinese text segmentation operation, support plug and manual intervention in the thesaurus (callback me
splitword
- 自己写的小分词程序,中文分词测试版,仅供参考,谢谢!-Writing their own small sub-term process, English sub-test version of the word, for reference purposes only, thank you!
Codes_and_Application
- 中科院的的分词工具,应该是分中文用的,效率不错-Chinese Academy of Sciences of the sub-word tools, should be used at the Chinese, the efficiency of a good
fenci
- 中文分词 中文搜索引擎中的关键字识辨-Chinese word segmentation in Chinese search engine keywords that identify
wordsegment1
- 中文分词算法,效率很高,使用词典树装搜索进行单词切割,并提供扩充词库的函数!-It s an arithmetc of word segment,it has a very high efficiency!
beiyes
- 贝叶斯网络概率中文分词算法,基于概率的分词算法-Bayesian network probability of Chinese word segmentation algorithm, based on the probability of word segmentation algorithm
src
- 利用lucene编写的一个简单搜索引擎,能够中文分词。-a simple search engine built with lucene.
imdict-chinese-analyzer
- imdict-chinese-analyzer 是 imdict智能词典 的智能中文分词模块,算法基于隐马尔科夫模型(Hidden Markov Model, HMM),是中国科学院计算技术研究所的ictclas中文分词程序的重新实现(基于Java),可以直接为lucene搜索引擎提供简体中文分词支持。-imdict-chinese-analyzer is a smart imdict Chinese Dictionary smart module segmentation algorithm
mmseg4j-1.7.3
- 中文分词技术 MMSeg1.7.3 最新版-Chinese word segmentation technology, the latest version of MMSeg1.7.3
Dictory
- 用B-树作为查找存储结构,对中文单词进行哈希,用最长逆向匹配算法进行中文句段分词,实现中文词典以及分词。中文词典规模为十万八千多词汇。分词的正确率在90 以上。-Use the B-tree as storage structure , and hash the Chinese word while storing or searching. Use the longest reverse matching algorithm to split Chinese sentence to word
IKAnalyzer3.1.1StableAllInOne
- Lucene 中文分词,很好的 可以随便下压,加油-Lucene Java
33753115ktdictseg_v1.0.01
- 中文分词算法,效率很高,使用词典树装搜索进行单词切割,并提供扩充词库的函数-It s an arithmetc of word segment,it has a very high efficiency!
lucene.net_search
- lucene.net 中文分词 分页高月显示
wordseg
- 运用R语言进行中文分词处理,得到词频统计,并绘制词云图直观表示(Chinese word segmentation and word cloud drawing)
Python逆向最大匹配实现-代码-文档-截图等
- 对于文本进行自然语言处理,中文分词。北邮计算机院NLP课的作业。(This is a homework for BUPT student, implementing the word segment of natural language processing.)