全部 标题 作者
关键词 摘要

OALib Journal期刊
ISSN: 2333-9721
费用:99美元

查看量下载量

相关文章

更多...

Scaling Conditional Random Fields by One-Against-the-Other Decomposition

Keywords: natural language processing,machine learning,conditional random fields,Chinese word segmentation

Full-Text   Cite this paper   Add to My Lib

Abstract:

As a powerful sequence labeling model,conditional random fields (CRFs) have had successful applications in many natural language processing (NLP) tasks.However,the high complexity of CRFs training only allows a very small tag (or label) set,because the training becomes intractable as the tag set enlarges.This paper proposes an improved decomposed training and joint decoding algorithm for CRF learning.Instead of training a single CRF model for all tags,it trains a binary sub-CRF independently for each tag.An optimal tag sequence is then produced by a joint decoding algorithm based on the probabilistic output of all sub-CRFs involved.To test its effectiveness,we apply this approach to tackling Chinese word segmentation (CWS) as a sequence labeling problem.Our evaluation shows that it can reduce the computational cost of this language processing task by 40-50% without any significant performance loss on various large-scale data sets.

Full-Text

Contact Us

service@oalib.com

QQ:3279437679

WhatsApp +8615387084133