搜索资源列表
CDevideSentence
- 用c++写的分词算法,简单,实用,详情看里面的帮助文件!-using c + + to write the sub-term algorithm is simple, practical, inside look at the details of the help files!
SQLET_split
- SQLET分词算法,一个C程序,供大家参考。-SQLET segmentation algorithm, a C program, for your reference.
ShootSearch_path_for_bate_1.0
- 一段关于中文信息分类的分词算法,他实现了信息过滤中的分词,为信息过滤的分类打下基础-section on the classification of information in Chinese word segmentation algorithm, he realized the information filtering the word, Information filtering to lay the foundation for the classification
word_vc
- vc++实现的基于字典的中文分词算法,基于贝耶斯网络的概率中文分词算法,以及文本相似程度比较的算法。
hash结构最大逆向中文分词算法
- 建立hash结构的中文词典,并实现了最大逆向中文分词算法
中文分词
- 一个不错的简单有效的中文分词算法, 可以做为研究中文的一个好例子
中文分词系统
- 一个简单有效中文分词算法, 可以学习研究
VisualC.rar
- 分词算法,是用C++代码实现的,并有实际的运行例子。,Segmentation algorithm is the use of C++ code, and there is a practical example of the operation.
SW_I_WordSegment
- SW-I中文分词算法,MFC程序,在visual studio 2008中调试通过。默认词库为mdb,由于较大未包含在源文件中,请自行下载mdb格式的词典。-SW-I Chinese word segmentation algorithm, MFC procedures, visual studio 2008 in debug through. Default thesaurus for the mdb, as a result of the larger not included in the
OpenCNSegmenter
- 中文分词,可以将中文的句子按照单词进行切分,很优秀的算法,在网络中得到-Chinese word segmentation, Chinese sentence can be carried out in accordance with the word segmentation, it is excellent algorithm, in the network have been
ICTCLAS
- 中科院分词系统VC++版本,在VS2005下编译通过,含有所有源代码,可以保证在中科院算法上,自己添加新的想法,或者对已有的算法进行优化。-Word Segmentation System, Chinese Academy of Sciences VC++ version at compile under VS2005 passed, containing all the source code, algorithms can guarantee at the Chinese Academy o
中文分词算法
- 本程序使用给出的字典进行学习并对训练语料进行分词处理,采用C语言编写,高效易懂!
V1.4
- 由KaiToo搜索开发的一款基于字典的简单中英文分词算法-Search by KaiToo developed a simple dictionary based on English and Chinese word segmentation algorithm
MmFenCi
- 基于MM的分词算法,有兴趣者可以把程序中没有完成的部分继续。-MM sub-word based algorithm, are interested in can not complete the program part to continue.
wordsegment1
- 中文分词算法,效率很高,使用词典树装搜索进行单词切割,并提供扩充词库的函数!-It s an arithmetc of word segment,it has a very high efficiency!
keyword-chouqu
- 基于逆向最大匹配算法的分词及基于HMM模型的词性标注系统,包括了未登录词的识别、数据库的添加等内容。(需要手动修改数据库的路径才可以运行)-Reverse Maximum Matching Algorithm Based on the sub-word HMM-based model and part of speech tagging system, including the unknown word identification, such as the contents of the d
beiyes
- 贝叶斯网络概率中文分词算法,基于概率的分词算法-Bayesian network probability of Chinese word segmentation algorithm, based on the probability of word segmentation algorithm
imdict-chinese-analyzer
- imdict-chinese-analyzer 是 imdict智能词典 的智能中文分词模块,算法基于隐马尔科夫模型(Hidden Markov Model, HMM),是中国科学院计算技术研究所的ictclas中文分词程序的重新实现(基于Java),可以直接为lucene搜索引擎提供简体中文分词支持。-imdict-chinese-analyzer is a smart imdict Chinese Dictionary smart module segmentation algorithm
Dictory
- 用B-树作为查找存储结构,对中文单词进行哈希,用最长逆向匹配算法进行中文句段分词,实现中文词典以及分词。中文词典规模为十万八千多词汇。分词的正确率在90 以上。-Use the B-tree as storage structure , and hash the Chinese word while storing or searching. Use the longest reverse matching algorithm to split Chinese sentence to word
基于mmseg算法的一个轻量级中文分词器
- 基于mmseg算法的一个轻量级开源中文分词器(A lightweight open source Chinese word segmentation based on the mmseg algorithm)