Joint Recognition of Intent and Semantic Slot Filling Combining Multiple Constraints
Spoken language understanding (SLU) consists of two sub-tasks, which are intent detection and semantic slot filling. Although the existing joint modeling methods realize the sharing of model parameters and apply the result of intent detection to semantic slot filling, the dependency before and after...
Main Author: | |
---|---|
Format: | Article |
Language: | zho |
Published: |
Journal of Computer Engineering and Applications Beijing Co., Ltd., Science Press
2020-09-01
|
Series: | Jisuanji kexue yu tansuo |
Subjects: | |
Online Access: | http://fcst.ceaj.org/CN/abstract/abstract2359.shtml |
id |
doaj-7e7c6e71b3054c3da22bf44e29096208 |
---|---|
record_format |
Article |
spelling |
doaj-7e7c6e71b3054c3da22bf44e290962082021-08-10T07:31:04ZzhoJournal of Computer Engineering and Applications Beijing Co., Ltd., Science PressJisuanji kexue yu tansuo1673-94182020-09-011491545155310.3778/j.issn.1673-9418.1909009Joint Recognition of Intent and Semantic Slot Filling Combining Multiple ConstraintsHOU Lixian, LI Yanling, LIN Min, LI Chengcheng0College of Computer Science and Technology, Inner Mongolia Normal University, Hohhot 010022, ChinaSpoken language understanding (SLU) consists of two sub-tasks, which are intent detection and semantic slot filling. Although the existing joint modeling methods realize the sharing of model parameters and apply the result of intent detection to semantic slot filling, the dependency before and after annotation is not considered for semantic slot filling task. A joint recognition model based on bidirectional long- short term memory (BLSTM) is adopted. After the hidden layer state is obtained by BLSTM, the attention mechanism is added to the two tasks respectively, and the result of intent detection is applied to the semantic slot filling by slot-gated mechanism. Considering the dependency before and after annotation, conditional random field (CRF) model is added into the semantic slot filling task to make the annotation result more accurate. Experimental data select the query in the field of flight information, the accuracy of the intent detection is 93.20% and F1 score of semantic slot filling is 99.28%. The performance of the model is verified on the SMP Chinese human-machine dialogue technology evaluation dataset. The results prove that the method is superior to other joint recognition models.http://fcst.ceaj.org/CN/abstract/abstract2359.shtmljoint modelingintent detectionsemantic slot fillingattention mechanismslot-gated mechanismconditional random field (crf) |
collection |
DOAJ |
language |
zho |
format |
Article |
sources |
DOAJ |
author |
HOU Lixian, LI Yanling, LIN Min, LI Chengcheng |
spellingShingle |
HOU Lixian, LI Yanling, LIN Min, LI Chengcheng Joint Recognition of Intent and Semantic Slot Filling Combining Multiple Constraints Jisuanji kexue yu tansuo joint modeling intent detection semantic slot filling attention mechanism slot-gated mechanism conditional random field (crf) |
author_facet |
HOU Lixian, LI Yanling, LIN Min, LI Chengcheng |
author_sort |
HOU Lixian, LI Yanling, LIN Min, LI Chengcheng |
title |
Joint Recognition of Intent and Semantic Slot Filling Combining Multiple Constraints |
title_short |
Joint Recognition of Intent and Semantic Slot Filling Combining Multiple Constraints |
title_full |
Joint Recognition of Intent and Semantic Slot Filling Combining Multiple Constraints |
title_fullStr |
Joint Recognition of Intent and Semantic Slot Filling Combining Multiple Constraints |
title_full_unstemmed |
Joint Recognition of Intent and Semantic Slot Filling Combining Multiple Constraints |
title_sort |
joint recognition of intent and semantic slot filling combining multiple constraints |
publisher |
Journal of Computer Engineering and Applications Beijing Co., Ltd., Science Press |
series |
Jisuanji kexue yu tansuo |
issn |
1673-9418 |
publishDate |
2020-09-01 |
description |
Spoken language understanding (SLU) consists of two sub-tasks, which are intent detection and semantic slot filling. Although the existing joint modeling methods realize the sharing of model parameters and apply the result of intent detection to semantic slot filling, the dependency before and after annotation is not considered for semantic slot filling task. A joint recognition model based on bidirectional long- short term memory (BLSTM) is adopted. After the hidden layer state is obtained by BLSTM, the attention mechanism is added to the two tasks respectively, and the result of intent detection is applied to the semantic slot filling by slot-gated mechanism. Considering the dependency before and after annotation, conditional random field (CRF) model is added into the semantic slot filling task to make the annotation result more accurate. Experimental data select the query in the field of flight information, the accuracy of the intent detection is 93.20% and F1 score of semantic slot filling is 99.28%. The performance of the model is verified on the SMP Chinese human-machine dialogue technology evaluation dataset. The results prove that the method is superior to other joint recognition models. |
topic |
joint modeling intent detection semantic slot filling attention mechanism slot-gated mechanism conditional random field (crf) |
url |
http://fcst.ceaj.org/CN/abstract/abstract2359.shtml |
work_keys_str_mv |
AT houlixianliyanlinglinminlichengcheng jointrecognitionofintentandsemanticslotfillingcombiningmultipleconstraints |
_version_ |
1721212472749195264 |