搜索资源列表
MeCab-Chinese-master
- MeCab的中文分词工具包哦!-MeCab Chinese word Toolkit
CWS_GatedRNN-master
- RNN深度学习的一款中文分词的工具,很支持大家参考的-A Chinese word learning tool RNN depth, very supportive of reference
ICTCLAS2016
- NLPIR/ ICTCLAS 2016系统分词实例-ICTCLAS2016 wordSegment demo
THULAC_lite_java_v1
- 中文文本分词 词频统计,分词,去掉停词。 仅支持UTF-8编码-Chinese text segmentation To get the word frequency, word segmentation, remove stop words. Support only UTF-8 encoding
NLPLibSVM
- libsvm分词训练集的java版本。包括libsvm.jar以及训练集样本-Libsvm version of the Java word segmentation training set. Including libsvm.jar and training set samples
dedesplit
- 本站采用的是一种用最小切分和逆向最大匹配法结合的一种复合分词算法,目前对岐义识别还处于初步研究阶段.-Powered by is a composite word segmentation algorithm with a minimum and maximum matching method combined reverse, the current recognition of ambiguity still in the preliminary research stage.
ICTCLAS
- 中科院的中文分词系统ICTCLAS,eclipse直接引入项目就可以用,已经测试过很好用的。- Chinese Academy of Sciences Chinese word breaker ICTCLAS, eclipse introduced directly into the project can be used, it has been tested very good use.
6_29
- 基于逆向匹配的中文分词算法实现,产生词典和测试数据,分词后具有结果分析功能,计算精确度,召回率,F值, -Chinese word segmentation algorithm based on reverse matching, dictionary and test data, with the result after word analysis function, calculation precision and recall rate, F value,
fenci
- 易语言中文分词,很难得可以分析出词性。3000字每秒-Part of speech analysis module
divide
- 采用正向最大匹配算法实现中文分词,基于matlab2013编程-MATLAB code
CreateIndex
- Lucene.Net+盘古分词做的定时建索引程序,自己网站用的。-A program which create index Sql server byLucene.Net+pangu.dll
WordSegment
- 大家可以试用一下。如果平时使用过程中发现有一些分词不正确,可以到 http://www.easyustc.com/fenci/build_sqlite.php 来修正字典,帮助我们优化字典。同时你也可以在这里下载最新的字典版本。 -You can try. If you usually use process found that some word is not correct, it can be corrected to http://www.easyustc.com/fenci/buil
wordpress-seo
- 什么是中文分词众所周知,英文是以词为单位的,词和词之间是靠空格隔开,而中文是以字为单位,句子中所有的字连起来才能描述一个意思。例如,英文句子I ama student,用中文则为:“我是一个学生”。-What is the Chinese word is known, is based on English words separated by spaces between words and word as a unit, and the Chinese as words, sentences
PORT
- 中国所有港口词库。适用于中文分词及中文港口专名抽取。-port dictionary
ICTCLAS_api
- 用于为指定文本进行分词操作。按照不同的词性进行分词。-Used to specify the text for the operation of word segmentation. According to different parts of speech.
CustomTokenizer
- 针对社交媒体推文的分词工具 针对社交媒体推文的分词工具 针对社交媒体推文的分词工具-Social media tweet s Tokenizer
HanLP-1.2.10.tar
- 汉语自然语言处理,包括分词,词性标注,命名实体,及句法依存-chinese netrual solve
IKAnalyzer2012_u6
- IKAnalyzer2012_u6词库,用于分词,然后进行分析或者抽取关键点研究-IKAnalyzer2012_u6 thesaurus for word, and then analyzed to extract key points or study
kmeansClassifier
- 该程序实现了keans分类,使用IK分词技术实现分词。-The program implements the k means classification, the use of IK word segmentation technology to achieve word segmentation.
OntologyLearning
- 本体学习相关研究 (1)ConcepLearing 从文本中学习本体概念,使用中科院NLPIR进行分词(也有jieba分词版本的),然后在根据统计(互信息/卡方值)和Topic Model(PLSA)的方法进行实验对比。 -Ontology learning research (1) ConcepLearing learning ontology concept the text, use the CAS NLPIR word segmentation (also jieba Wo