Joint Recognition of Intent and Semantic Slot Filling Combining Multiple Constraints

Spoken language understanding (SLU) consists of two sub-tasks, which are intent detection and semantic slot filling. Although the existing joint modeling methods realize the sharing of model parameters and apply the result of intent detection to semantic slot filling, the dependency before and after...

Full description

Bibliographic Details
Main Author: HOU Lixian, LI Yanling, LIN Min, LI Chengcheng
Format: Article
Language:zho
Published: Journal of Computer Engineering and Applications Beijing Co., Ltd., Science Press 2020-09-01
Series:Jisuanji kexue yu tansuo
Subjects:
Online Access:http://fcst.ceaj.org/CN/abstract/abstract2359.shtml
id doaj-7e7c6e71b3054c3da22bf44e29096208
record_format Article
spelling doaj-7e7c6e71b3054c3da22bf44e290962082021-08-10T07:31:04ZzhoJournal of Computer Engineering and Applications Beijing Co., Ltd., Science PressJisuanji kexue yu tansuo1673-94182020-09-011491545155310.3778/j.issn.1673-9418.1909009Joint Recognition of Intent and Semantic Slot Filling Combining Multiple ConstraintsHOU Lixian, LI Yanling, LIN Min, LI Chengcheng0College of Computer Science and Technology, Inner Mongolia Normal University, Hohhot 010022, ChinaSpoken language understanding (SLU) consists of two sub-tasks, which are intent detection and semantic slot filling. Although the existing joint modeling methods realize the sharing of model parameters and apply the result of intent detection to semantic slot filling, the dependency before and after annotation is not considered for semantic slot filling task. A joint recognition model based on bidirectional long- short term memory (BLSTM) is adopted. After the hidden layer state is obtained by BLSTM, the attention mechanism is added to the two tasks respectively, and the result of intent detection is applied to the semantic slot filling by slot-gated mechanism. Considering the dependency before and after annotation, conditional random field (CRF) model is added into the semantic slot filling task to make the annotation result more accurate. Experimental data select the query in the field of flight information, the accuracy of the intent detection is 93.20% and F1 score of semantic slot filling is 99.28%. The performance of the model is verified on the SMP Chinese human-machine dialogue technology evaluation dataset. The results prove that the method is superior to other joint recognition models.http://fcst.ceaj.org/CN/abstract/abstract2359.shtmljoint modelingintent detectionsemantic slot fillingattention mechanismslot-gated mechanismconditional random field (crf)
collection DOAJ
language zho
format Article
sources DOAJ
author HOU Lixian, LI Yanling, LIN Min, LI Chengcheng
spellingShingle HOU Lixian, LI Yanling, LIN Min, LI Chengcheng
Joint Recognition of Intent and Semantic Slot Filling Combining Multiple Constraints
Jisuanji kexue yu tansuo
joint modeling
intent detection
semantic slot filling
attention mechanism
slot-gated mechanism
conditional random field (crf)
author_facet HOU Lixian, LI Yanling, LIN Min, LI Chengcheng
author_sort HOU Lixian, LI Yanling, LIN Min, LI Chengcheng
title Joint Recognition of Intent and Semantic Slot Filling Combining Multiple Constraints
title_short Joint Recognition of Intent and Semantic Slot Filling Combining Multiple Constraints
title_full Joint Recognition of Intent and Semantic Slot Filling Combining Multiple Constraints
title_fullStr Joint Recognition of Intent and Semantic Slot Filling Combining Multiple Constraints
title_full_unstemmed Joint Recognition of Intent and Semantic Slot Filling Combining Multiple Constraints
title_sort joint recognition of intent and semantic slot filling combining multiple constraints
publisher Journal of Computer Engineering and Applications Beijing Co., Ltd., Science Press
series Jisuanji kexue yu tansuo
issn 1673-9418
publishDate 2020-09-01
description Spoken language understanding (SLU) consists of two sub-tasks, which are intent detection and semantic slot filling. Although the existing joint modeling methods realize the sharing of model parameters and apply the result of intent detection to semantic slot filling, the dependency before and after annotation is not considered for semantic slot filling task. A joint recognition model based on bidirectional long- short term memory (BLSTM) is adopted. After the hidden layer state is obtained by BLSTM, the attention mechanism is added to the two tasks respectively, and the result of intent detection is applied to the semantic slot filling by slot-gated mechanism. Considering the dependency before and after annotation, conditional random field (CRF) model is added into the semantic slot filling task to make the annotation result more accurate. Experimental data select the query in the field of flight information, the accuracy of the intent detection is 93.20% and F1 score of semantic slot filling is 99.28%. The performance of the model is verified on the SMP Chinese human-machine dialogue technology evaluation dataset. The results prove that the method is superior to other joint recognition models.
topic joint modeling
intent detection
semantic slot filling
attention mechanism
slot-gated mechanism
conditional random field (crf)
url http://fcst.ceaj.org/CN/abstract/abstract2359.shtml
work_keys_str_mv AT houlixianliyanlinglinminlichengcheng jointrecognitionofintentandsemanticslotfillingcombiningmultipleconstraints
_version_ 1721212472749195264