中文自动分词是中文信息处理的基础。
Automatic Chinese word segmentation is the basis of Chinese information processing.
特别是对专有名词的处理是中文自动分词中的又一个难点。
Especially, it is very difficult to deal with special noun in Chinese automatic word segmentation.
中文搜索引擎的重点在于中文关键信息提取,其中的难点就是中文自动分词。
The core of Chinese Search Engine is the key content extracting, and the bottleneck is Chinese Word Automatic Segmentation.
词典是中文自动分词的基础,分词词典机制的优劣直接影响到中文分词的速度和效率。
As a basic component of Chinese word segmentation system, the dictionary mechanism influences the speed and the efficiency of segmentation significantly.
汉语自动分词是计算机中文信息处理中的难题,也是文献内容分析中必须解决的关键问题之一。
Chinese automatic segmentation is one of the most difficult problems in computer Chinese information disposal and the key problem that document content analysis must resolve.
汉语自动分词是进行中文信息处理的基础。
Automatic Chinese word segmentation is the basis of Chinese information processing.
汉语自动分词是进行中文信息处理的基础。
Automatic Chinese segmentation is the basis of Chinese information processing.
汉语自动分词是计算机中文信息处理中的难题,也是文献内容分析中必须解决的关键问题之一。
Chinese automatic seg mentation is one of the most difficult problems in computer Chinese information disposal and the key problem that document content analysis must resolve.
汉语自动分词是中文信息处理的首要工作。
Chinese automatic word segmentation is the first work in Chinese information processing.
汉语的自动分词,是计算机中文信息处理领域中一个基础而困难的课题。
Automatic word segmentation for the Chinese language is a fundamental and difficult problem in the field of computer Chinese language information processing.
文章首先构造了自动答疑系统架构,改进了中文分词算法,并利用领域本体库和语句相似度设计了该系统。
In this paper, we first construct the system architecture, improve the Chinese text segmentation algorithm, then, by making use of domain ontology base and sentence similarity, design the system.
现代汉语文本自动分词是中文信息处理的重要基石,为此提供一个通用的分词接口是非常重要的。
Automatic word segmentation of modern Chinese text is the base of Chinese information processing. So a general purpose application interface for word segmentation is important.
汉语自动分词是中文信息处理中的基础课题。
Word segmentation is a basic task of Chinese information processing.
在本文中,我们提出了一种统一的统计语言模型方法用来汉语自动分词和中文命名实体识别,这种方法对基于词的三元语言模型进行了很好的扩展。
In this paper, we extend a word-based trigram modeling to Chinese word segmentation and Chinese named entity recognition, by proposing a unified approach to SLM.
汉语自动分词是中文信息处理的重要基石。
目前学术界主要采用计算机自动分词解决中文文本分词,但是这种方法不能完全解决分词问题,这是因为计算机自动分词不能彻底地解决歧义字段的切分。
And now the most widely used method is automatic segmentation. But this method can't solve the problem thoroughly, because this method can't solve the problem of ambiguous segment.
该方法通过对机器分词语料和人工校对语料的学习,自动获取中文文本的分词校对规则,并应用规则对机器分词结果进行自动校对。
It discusses and analyzes the actuality of Chinese word segmentation, and describes an approach to correcting the Chinese word segmentation automatically based on rules.
该方法通过对机器分词语料和人工校对语料的学习,自动获取中文文本的分词校对规则,并应用规则对机器分词结果进行自动校对。
It discusses and analyzes the actuality of Chinese word segmentation, and describes an approach to correcting the Chinese word segmentation automatically based on rules.
应用推荐