Chinese word segmentation is the foundation of NLP.
汉语自动分词是中文信息处理的重要基石。
A fast algorithm for generating Chinese word segmentation digraph was given.
给出了一种汉语分词有向图的快速生成算法。
Overlapping ambiguity is a major type of ambiguity in Chinese word segmentation.
交集型分词歧义是汉语自动分词中的主要歧义类型之一。
Automatic Chinese word segmentation is the basis of Chinese information processing.
汉语自动分词是进行中文信息处理的基础。
Automatic Chinese word segmentation is the basis of Chinese information processing.
中文自动分词是中文信息处理的基础。
A major problem of Chinese word segmentation facing is the New word Identification (NWI).
汉语自动分词面临的一个主要难题就是新词的自动识别。
We use a simple and effective Chinese word segmentation method and compare CLIR performance.
使用了一种简单高效的汉语分词方法,比较了两种检索条件翻译方法的性能。
Transliterated person names identification is the necessary part of Chinese word segmentation.
西方姓名译名的自动识别为汉语自动分词不可或缺的组成部分。
The system of Chinese word segmentation based on machine learning is researched and implemented.
本文研究并实现了基于机器学习的分词系统。
The ACL-SIGHAN sponsored the First International Chinese Word Segmentation in July, 2003 in Japan.
2003年在日本札幌举行了第一届ACL -SIGHAN国际中文分词竞赛。
So, to make the computer capable of handling Chinese text, text must do Chinese word segmentation first.
所以,要使计算机能够处理中文文本,就必须先进行中文分词。
In those texts, we select bigram as feature after Chinese word segmentation, deleting stop word and other process.
在筛选出的文本中,经过分词、去除停用词等处理后,选取二元词串作为特征;
This paper mostly researches Chinese word segmentation, because it is a basal place in intelligent answering system.
它是中文智能答疑系统的一项基本性技术,可以说,没有分词技术任何中文答疑系统都不具有智能性。
The design and implementation of"Chinese Word Segmentation"is the most important part of the search engine technology.
中文分词是搜索引擎中比较重要的部分,本文分析了正向和逆向的最大匹配分词以及基于统计的分词方法。
Index module: first of all, discuss the design method of Chinese word segmentation and choose a word segmentation algorithm.
索引模块中:首先,讨论了中文分词的设计思想,选择了分词的算法。
The speed of Chinese word segmentation is very important for many Chinese NLP systems, such as web search engines based on words.
对于基于词的搜索引擎等中文处理系统,分词速度要求较高。
Aiming at the dissatisfied effect of Chinese word segmentation to Email texts, an improved Maximum Match Based Approach is presented.
针对邮件文本分词效果较差的特点,提出采用一种改进的最大匹配法来进行中文分词的方法。
We analyzed, designed and achieved a module of Chinese word segmentation and Part-Of-Speech Tagging based on Condition Random Fields model.
分析、设计和实现了一个基于条件随机场模型的汉语分词和词性标注模块。
Here we explore SVM for a Chinese word segmentation task, use the context attributes and rule-based attributes as the features for a sample.
本文首次使用SVM方法来完成中文分词的任务,使用上下文窗体属性和基于规则的属性对样本进行刻画。
Combinational ambiguity is a challenging issue in Chinese word segmentation in that its disambiguation depends on the contextual information.
组合型歧义切分字段一直是汉语自动分词的难点,难点在于消歧依赖其上下文语境信息。
The paper introduces the design and implementation of Chinese word segmentation system, which is based on statistic the frequency of the word.
论文介绍了一个基于词频统计的中文分词系统的设计和实现。
Using prefix tree and dynamic programming, this algorithm boosts the speed of Chinese word segmentation and guarantees relatively high precision.
基于前缀树和动态规划,该算法提高了中文分词速度,同时保持了相对较高的分词准确性。
The process of the intelligent scoring is divided into three main steps: Chinese word segmentation, syntactic analysis and similarity computation.
本文将阅卷过程分解为三个主要步骤来进行:中文分词、句法分析和相似度计算。
Search engine technology related to natural language understanding, Chinese word segmentation, artificial intelligence, machine learning and so on.
搜索引擎的技术涉及到自然语言理解、中文分词、人工智能、机器学习等学科。
Chinese word segmentation is a basic research issue on Chinese NLP areas such as information retrieval, machine translation, text correction, and so on.
汉语分词是信息检索、机器翻译、文本校对等中文信息处理重要领域的基础。
Therefore, the primary issue of Chinese information processing, that is, to a sentence to separate words, this is the Chinese word segmentation problem.
因此中文信息处理的首要问题,就是要将句子中一个个词给分离出来,这就是中文分词问题。
As a basic component of Chinese word segmentation system, the dictionary mechanism influences the speed and the efficiency of segmentation significantly.
词典是中文自动分词的基础,分词词典机制的优劣直接影响到中文分词的速度和效率。
This paper analyzes several existing Chinese word segmentation methods, brings out a keywords extraction algorithm which according to the weight formula.
分析现有几种中文分词方法,提出一种关键词抽取算法。
In this paper, the dictionary mechanism is dynamic TRIE tree, and we have designed the Chinese word segmentation dictionary. The dictionary USES less memory.
论文采用动态TRIE索引树的词典机制,设计并实现了汉语分词词典,有效地减少了词典空间。
In this paper, the dictionary mechanism is dynamic TRIE tree, and we have designed the Chinese word segmentation dictionary. The dictionary USES less memory.
论文采用动态TRIE索引树的词典机制,设计并实现了汉语分词词典,有效地减少了词典空间。
应用推荐