Semi-Supervised Learning of Statistical Models for Natural Language Understanding

Natural language understanding is to specify a computational model that maps sentences to their semantic mean representation. In this paper, we propose a novel framework to train the statistical models without using expensive fully annotated data. In particular, the input of our framework is a set o...

Full description

Bibliographic Details
Main Authors: Deyu Zhou, Yulan He
Format: Article
Language:English
Published: Hindawi Limited 2014-01-01
Series:The Scientific World Journal
Online Access:http://dx.doi.org/10.1155/2014/121650
id doaj-15e1a6b7efec4ddb8fb4a15b0ff5df11
record_format Article
spelling doaj-15e1a6b7efec4ddb8fb4a15b0ff5df112020-11-24T21:59:05ZengHindawi LimitedThe Scientific World Journal2356-61401537-744X2014-01-01201410.1155/2014/121650121650Semi-Supervised Learning of Statistical Models for Natural Language UnderstandingDeyu Zhou0Yulan He1School of Computer Science and Engineering, Key Laboratory of Computer Network and Information Integration, Ministry of Education, Southeast University, Nanjing 210096, ChinaSchool of Engineering and Applied Science, Aston University, Birmingham B4 7ET, UKNatural language understanding is to specify a computational model that maps sentences to their semantic mean representation. In this paper, we propose a novel framework to train the statistical models without using expensive fully annotated data. In particular, the input of our framework is a set of sentences labeled with abstract semantic annotations. These annotations encode the underlying embedded semantic structural relations without explicit word/semantic tag alignment. The proposed framework can automatically induce derivation rules that map sentences to their semantic meaning representations. The learning framework is applied on two statistical models, the conditional random fields (CRFs) and the hidden Markov support vector machines (HM-SVMs). Our experimental results on the DARPA communicator data show that both CRFs and HM-SVMs outperform the baseline approach, previously proposed hidden vector state (HVS) model which is also trained on abstract semantic annotations. In addition, the proposed framework shows superior performance than two other baseline approaches, a hybrid framework combining HVS and HM-SVMs and discriminative training of HVS, with a relative error reduction rate of about 25% and 15% being achieved in F-measure.http://dx.doi.org/10.1155/2014/121650
collection DOAJ
language English
format Article
sources DOAJ
author Deyu Zhou
Yulan He
spellingShingle Deyu Zhou
Yulan He
Semi-Supervised Learning of Statistical Models for Natural Language Understanding
The Scientific World Journal
author_facet Deyu Zhou
Yulan He
author_sort Deyu Zhou
title Semi-Supervised Learning of Statistical Models for Natural Language Understanding
title_short Semi-Supervised Learning of Statistical Models for Natural Language Understanding
title_full Semi-Supervised Learning of Statistical Models for Natural Language Understanding
title_fullStr Semi-Supervised Learning of Statistical Models for Natural Language Understanding
title_full_unstemmed Semi-Supervised Learning of Statistical Models for Natural Language Understanding
title_sort semi-supervised learning of statistical models for natural language understanding
publisher Hindawi Limited
series The Scientific World Journal
issn 2356-6140
1537-744X
publishDate 2014-01-01
description Natural language understanding is to specify a computational model that maps sentences to their semantic mean representation. In this paper, we propose a novel framework to train the statistical models without using expensive fully annotated data. In particular, the input of our framework is a set of sentences labeled with abstract semantic annotations. These annotations encode the underlying embedded semantic structural relations without explicit word/semantic tag alignment. The proposed framework can automatically induce derivation rules that map sentences to their semantic meaning representations. The learning framework is applied on two statistical models, the conditional random fields (CRFs) and the hidden Markov support vector machines (HM-SVMs). Our experimental results on the DARPA communicator data show that both CRFs and HM-SVMs outperform the baseline approach, previously proposed hidden vector state (HVS) model which is also trained on abstract semantic annotations. In addition, the proposed framework shows superior performance than two other baseline approaches, a hybrid framework combining HVS and HM-SVMs and discriminative training of HVS, with a relative error reduction rate of about 25% and 15% being achieved in F-measure.
url http://dx.doi.org/10.1155/2014/121650
work_keys_str_mv AT deyuzhou semisupervisedlearningofstatisticalmodelsfornaturallanguageunderstanding
AT yulanhe semisupervisedlearningofstatisticalmodelsfornaturallanguageunderstanding
_version_ 1725849303822368768