搜索资源列表
SQLET_split
- 另外一个中文分词程序,采用的是可显示的字典,可供大家分析-another Chinese word segmentation procedures, the show is the dictionary for our analysis
segmentor_Perl
- 中文分词算法。Perl语言编写。wordlist.txt为词库。-Chinese Segmentation. Perl language. Wordlist.txt for the thesaurus.
xerdoc
- 这些都是关于中文分词的,一定会对很多人有帮助的!-these are on the Chinese word, and you will help a lot of people!
SplitWord
- 中文分词组件,?形姆执首榧?中文分词组件,-Chinese word components, the Chinese word components, the Chinese word components,
MySeg
- 最短路径法分词程序。将中文句子经过原子切分后生成一个有向无环图,然后使用Dijkstra算法求出由起点到终点的最短路径以实现中文分词。-shortest path method participle procedures. Chinese Sentence will be read after splitting atoms generated a directed acyclic graph. then use the Dijkstra algorithm derived from the s
findkey.c
- 此程序解决的问题:较好的, 并适应短字符串的中文分词算法.根据词库 发现以换行符分隔的众多标题中的 top N 关键字并以此更新词库.是一个分类分词算法 -this procedure to solve the problem : better, and adapt to the short string of Chinese Segmentation. According thesaurus found in the many separate newline heading the to
5271615762
- 中文分词技术 从别的网上摘的 感觉还不错 请大家-Chinese word technology from other online pick feeling quite well please try
ictclas10
- 基于中科院的ICTCLAS实现中文分词系统 开发工具是JAVA.经测试,效果很好.-ICTCLAS based on the realization of the Chinese Academy of Sciences Chinese word segmentation system is the Java development tools. Tested, good results.
framework
- 基于动态规划的中文分词程序,用vc写的,便于扩展。-based on dynamic programming of the Chinese word segmentation procedures using vc write, easy expansion.
chentian.nutch
- 实现了基于词库的nutch中文分词,主要修改了其中的.jj文件等-realized based on the thesaurus nutch Chinese word, the main change of them. Jj documents
chentian.fenci
- 实现了基于词库的nutch中文分词,这一部分是其中的dll文件-realized based on the thesaurus nutch Chinese word, this part is one of the dll file
firtex_beta102_src
- FirteX介绍 功能: 支持增量索引,差量索引,多字段索引,提供了3种前向索引方式; 支持纯文本,HTML,PDF等文件格式; 提供快速中文分词; 从底层到高层,提供了多种索引访问接口,灵活自由地使用索引文件; 提供丰富的检索语法,支持多字段检索,日期范围检索,检索结果自定义排序等。 性能: 在Pentium 4 2.8G 2GRAM的机器上超过200Mb每分钟的索引速度 在近7G的索引文件(100G网页,11G纯文本的索引)上检索,仅使用十几M内存在数毫
ECDictionary
- java树型字典的建立,是中文分词 检索的基础-java tree dictionary for the establishment of the Chinese-based word retrieval
CSW_CN_dll_5.0.0
- CSW中文分词组件_5.0.0 CSW中文分词组件_5-CSW Chinese word segmentation components _5.0.0 CSW Chinese word segmentation components _5
SegmentRAM
- 1.基于luncene倒排索引格式的高性能索引数据访问接口。 2. 实现若干中文分词的的分词算法。-1. Based on luncene inverted index format of high-performance data access interface Index. 2. A number of Chinese word segmentation algorithm of.
3GWS-jni
- 又一个free中文分词 不过需要向作者索要注册码~这不是什么问题吧-also a free Chinese word but it needs to license key-authors claim this is not what it
lunce-splitword
- 一个运用于lunce中的中文分词算法源码,与大家共享-a lunce applied to the Chinese word segmentation algorithm source code and share
sousyinqing
- 搜索引擎技术的研究论文,本文阐述了搜索引擎的基本原理,着重分析了中文分词的设计与实现。-search engine technology research papers, the paper deals with the search engine's basic principles focused on analysis of the Chinese word Design and Implementation.
ICTCLAS(C++)
- 中科院分词,可供直接使用,源于中科院中文分词网-CAS segmentation for direct use, and from the Chinese Academy of the Chinese word segmentation Network
ShootSearch
- 中文分词 ShootSearch中文分词组件.rar-Chinese word segmentation ShootSearch Chinese components. Rar