Accesso libero

Research on Joint Modeling of Intent Detection and Slot Filling

INFORMAZIONI SU QUESTO ARTICOLO

Cita

Introduction

The Natural-language understanding module is the core part of task dialogue, which aims to transform user input into structured language. [1] The main task is to intentionally detect and supplement these two types of sub marriages. The former aims to roughly understand the intention of the target discourse. And determining the category they belong to is usually seen as sentence level text classification work. The latter translates the intention of the target discourse into specific instructions, namely. e. Identifying key semantic information contained in user discourse is considered as a character-level sequence annotation task [2].

Earlier, intent detection and filling tasks were independently modeled, commonly known as assembly line route [3]. The approach of the pipeline ignores the correlation between the two tasks, resulting in the problem that the intention detection results are difficult to match with the slot filling results and the problem of error propagation [4]. The joint modeling of the two is a method with better performance at present. A hot topic in recent years has been the linking of intent detection and socket execution tasks into a common model. Starting from these three types, this paper will analyze the correlation between Intent detection and slot filling tasks. And analyze the impact of the individual task on the performance of joint pattern discovery.

Related Work

Intention detection can be seen as a sentence-level classification task. The traditional method is to obtain important physical information from text through n-grams[9], but this method is limited to simple sentences. Traditional machine learning algorithms, such as SVM [10] and Adapost [11], train models by labeling certain data. Deep learning methods are also more effective in intention detection tasks such as CNN, RCNN, LSTM, and FastTextFor filling in spaces, it is usually considered as a character hierarchy marking task. The traditional method is based on the Conditional Combination Field (CRF) architecture and has strong sequence labeling ability, but only applicable to relatively small datasets. [13] In addition, in the “time slot filling” task, deep learning methods are also superior to traditional models, such as CNN based [14] and RNN based [15].

People have found that traditional methods often overlook the strong relationship between two tasks, leading to the problem of error propagation. Therefore, scholars began to study the way of joint modeling and became the mainstream. It is broadly classified into three types: intent to plus one way synthesis, plus to intent one way synthesis, double synthesis model. GanniTur et al. [4] proposed to use RNNs to learn the joint work. The models consider the correlation between the two tasks by sharing parameters. One joint model guides intent detection and slot filling tasks through intent or slot information display. For example, Goo et al. [4] combined the loss functions of the two tasks for optimization, and used the gating mechanism as a special gate function to model intent detection and slot, the relationship between fillings. Li et al. [16] proposed an intention enhancement gating mechanism to mine the semantic association between slots and intentions. Qin et al.[17] used the stack propagation framework to directly input the word level intention detection information into the slot filling. While the two-way joint model models the two tasks in two directions, Wang et al. [18] proposed Bi-Model to consider the cross influence between intent and slot.

Although the joint model of these two tasks, namely intention detection and filling gaps, has made significant progress. However, the relationship between the two in joint modeling and the extent to which each task affects the overall model validation performance still need to be studied and tested. Let's compare based on the joint model proposed by Qin et al.[8] for intention detection and filling in gaps.

Model

As shown in the overall structure in the figure 1.This model includes an encoder module, a bidirectional relationship module, an intention and time channel, and a decoding module, among which a bilateral relationship module, an intention and channel Time includes the attention layer of intention and spatial labels, the attention layer of intention and spatial interaction, and the feeding network layer.

Figure 1.

The Overall Structure of the Model.

Encoder Module

This module uses contextual semantic attributes as input to subsequent sub modules through Bi-LSTM, as shown in Figure 2. When encoding user text is transformed into an input sequence after a pre-training layer, and X is input to the Bi-LSTM layer to take advantage of temporal features in word sequences to obtain contextual semantic information. For the input sequence, each position i in the sequence has an LSTM to learn it from both positive and negative directions, respectively, if the hidden layer state of the LSTM output at position i positive is hi \overrightarrow {{h_i}} , the positive LSTM outputs the hidden layer state as hi \overleftarrow {{h_i}} . The forward and reverse results of LSTM are also combined to obtain the hidden layer state of each vector after encoding. BilSTM is as follows: H={h1,h2,,hn} H = \{ {{\rm{h}}_1},{{\rm{h}}_2}, \ldots ,{{\rm{h}}_{\rm{n}}}\} hi=[hi,hi] {h_i} = \left[ {\overrightarrow {{h_i}} ,\overleftarrow {{h_i}} } \right]

Figure 2.

Encoder.

Intent and Slot Joint Module
Intent and slot labeling attention layer

The intent label and the slot label are given different levels of attention to obtain explicit intent representation and slot label representation, which are used for the subsequent direct interaction of the input co-interactive attention layer. In particular, the parameters of slot filling decoder and intent detection decoder layer are used as slot embedding matrix and intent embedding matrix (and the number of slots and intent tags are recorded). Use as query and askey and value to obtain intent and slot attention representation: A=softmax(HWv) A = softmax \left( {H{W^v}} \right)

Enter the embedded state obtained from the partition coding module into the intent focus layer and the anecdote focus layer to obtain the semantic information of the intent or anecdote and to obtain the intent annotation focus or anecdote representation, as shown in Figure 3. Hv=H+AWv {H_v} = H + A{W^v}

Figure 3.

Label Attention Layer.

Intent and Slot Interaction Attention Layer

The intention representation and slot representation obtained by marking the attention layer capture the semantic information of the intention and slot respectively, and further explore in the collaborative interactive attention layer through and, as shown in Figure 4, to realize the two-way connection between the two tasks. Matrices and mappings, as well as matrices using different linear projections.

The intention of updating, displaying, and combining corresponding slot data is considered as a total weight query for keywords, values, and results, which is a normal layer with Dong intentionally obtains new expressions of intent from the attention layer of the tag, in order to avoid overfitting and reduce errors. Similarly, in order to improve the impression of time channels and integrate corresponding intention information, they are treated as queries, keywords, and values, resulting in a parity sum. Weight, using intention expression to obtain new intention expressions from the attention label layer to the normal layer. CI=softmax(QIKSTdk)VS {{\rm{C}}_{\rm{I}}} = softmax \left( {{{{{\rm{Q}}_{\rm{I}}}{{\rm{K}}_{\rm{S}}}^{\rm{T}}} \over {\sqrt {{{\rm{d}}_{\rm{k}}}} }}} \right){{\rm{V}}_{\rm{S}}} HI=LN(HI+CI) {{\rm{H}}_{\rm{I}}}^\prime = LN\left( {{{\rm{H}}_{\rm{I}}} + {{\rm{C}}_{\rm{I}}}} \right) CS=softmax(QSKITdk)VI {{\rm{C}}_{\rm{S}}} = softmax \left( {{{{{\rm{Q}}_{\rm{S}}}{{\rm{K}}_{\rm{I}}}^{\rm{T}}} \over {\sqrt {{{\rm{d}}_{\rm{k}}}} }}} \right){{\rm{V}}_{\rm{I}}} HS=LN(HS+CS) {H_S}^\prime = LN\left( {{H_S} + {C_S}} \right)

Figure 4.

Intent and Slot Interaction Attention layer.

The above is the case when the intention is associated with time slots in two directions, and the unidirectional relationship is also similar. By treating the questionnaire as a keyword and representing the intention of the time period, one can obtain the intention to express a one-way relationship between the time period. The one-way relationship between time intervals is similar by treating them as questionnaires and keywords as values.

Feed-forward network layer

The intention and interval data are implicitly fused through the feed network layer, as shown in Figure 5.Firstly, the intention display and interval display of connection updates, including intention and spatial information, are connected through layers. Feed the network and ultimately obtain the latest intention display and interval display through the “layer normalization” function. HS=LN(HS+CS) {H_S}^\prime = LN\left( {{H_S} + {C_S}} \right) HIS=HIHS {H_{IS}} = {H_I}^\prime\; \oplus {H_S}^\prime

Figure 5.

Feed-forward Network Layer.

Decoder

In order to have sufficient interaction between slot and intent detection tasks, a network with multi-layer stacked cooperative interactive attention is applied. After stacking the L layer, the final updated slot and intent representation are obtained, as shown in Figure 6. A maximum pooling operation is applied to obtain the representation C of the sentence as the input of intent detection: H^I(L)=(H^(I,1)(L),H^(I,2)(L),,H^(I,n)(L)) \hat H_I^{\left( L \right)} = \left( {\hat H_{\left( {I,1} \right)}^{\left( L \right)},\hat H_{\left( {I,2} \right)}^{\left( L \right)}, \ldots ,\;\hat H_{\left( {I,n} \right)}^{\left( L \right)}} \right) H^S(L)=(H^(S,1)(L),H^(S,2)(L),,H^(S,n)(L)) {\rm{\hat H}}_{\rm{S}}^{\left( {\rm{L}} \right)} = \left( {{\rm{\hat H}}_{\left( {{\rm{S}},1} \right)}^{\left( {\rm{L}} \right)},{\rm{\hat H}}_{\left( {{\rm{S}},2} \right)}^{\left( {\rm{L}} \right)}, \ldots ,\;{\rm{\hat H}}_{\left( {{\rm{S,n}}} \right)}^{\left( {\rm{L}} \right)}} \right) y^I=softmax(WIc+bs) {\hat y^I} = softmax\left( {{W^I}c + bs} \right) oI=argmax(yI) {o^I} = argmax\left( {{y^I}} \right)

Here ŷI denotes the output intent distribution, oI denotes the intent label, and WI denotes the trainable parameters of the model.

Figure 6.

Decoder.

Here, conditional random fields are used to model the slot scale dependence between adjacent character, i.e: OS=WSH^S(L)+bs {O_S} = {W^S}\hat H_S^{\left( L \right)} + bs P(y^|Os)=i=1expf(yi1,yi,Os)yi=1expf(yi1,yi,Os) P(\hat y|{O_s}) = {{\sum\nolimits_{i = 1} {exp\;f\left( {{y_{i - 1}},{y_i},{O_s}} \right)} } \over {\sum\nolimits_{y^\prime} {\sum\nolimits_{i = 1} {exp\;f\left( {{y_{i - 1}}^\prime,{y_i}^\prime,{O_s}} \right)} } }}

Here, the transition score is calculated from to represent the prediction label sequence.

Experiment
Experimental setup and evaluation index

This experiment was conducted using the Chinese CAIS dataset, and Liu et al. [7] used the dataset to collect the voices of Chinese artificial intelligence (CAIS) speakers. And marked with slot door and intention label, exercise kit (7995 sentences), inspection kit (994 sentences) and Test suite. (Including 1024 sentences) Separate from intention allocation.

The following models are used for experiments: Slot-Gated [4], Stack-Propagation, SF-ID [5], CM-Net [6], and Co-Interactive Transformer [8]. Among them, slot gate control and stack communication are unidirectional common mode, SF-ID network, and CM network. And the interaction converter is a bidirectional joint format. At the same time, precision (acc) was used in the experiment to evaluate intention detection work, and F1 value was used to evaluate the work of filling gaps. And evaluate the overall performance of the model using sentence accuracy (sent_acc), defined as follows: acc=i=1b{1IDi*=IDi0IDi*IDib acc = {{\mathop \sum \nolimits_{i = 1}^b \left\{ {\matrix{ 1 \hfill & {ID_i^* = I{D_i}} \hfill \cr 0 \hfill & {ID_i^* \ne I{D_i}} \hfill \cr } } \right.} \over b} F1=2×i=1b|SFiSFi*|i=1b|SFi|+i=1b|SFi*| {F_1} = {{2 \times \sum\nolimits_{i = 1}^b {\left| {S{F_i} \cap SF_i^*} \right|} } \over {\sum\nolimits_{i = 1}^b {\left| {S{F_i}} \right| + \sum\nolimits_{i = 1}^b {\left| {SF_i^*} \right|} } }}

Identify the Headings Experimental results and analysis

Table 1 shows the results of some mainstream models on the data set CAS. The models used in experiment 1 and experiment 2 are one-way combined models, and the models used in experiment 3–6 are two-way combined models. The data show that the performance of the two-way joint model is higher than that of the one-way joint model, but the performance of some one-way joint models is not excluded from the two-way joint model.

Model Results of Chinese Dataset CAIS

Experiment Model CAIS
Slot F1 Acc Sent-acc
1 Sloted Gated 81.8a 94.3 80.5
2 Stack-Propagation 87.8 94.7 84.7
3 SF-ID Network 84.9 94.5 82.4
4 CM-Net 86.2 94.6 84.6
5 Co-Interactive Transformer 88.6 95.2 86.2
6 Our Model 89.2 96.3 87.9

The one-way joint model slot gated used in experiment 1 learned the relationship between intent and slot attention vector by learning the relationship between intent and slot attention vector, but the information acquisition of intent is limited. The one-way joint model Stack Propagation used in Experiment 2 uses the word-level intention detection mechanism, uses the output of intention detection as the input of the slot filling task, and directly uses the information of intention to guide the slot filling task to improve the performance of the model, making the performance of the model higher than that of the two-way joint models SF-ID network and cm net used in experiments 3 and 4. However, the performance of Stack Propagation model is still limited by the limited filling capacity of Intent guide slot. The Co-Interactive Transformer model used in Experiment 5 is a relatively advanced two-way correlation model in recent years, and the two have achieved good results by establishing two-way connections in two related tasks to consider cross influence. The model in Experiment 6 was optimized based on Experiment 5. In order to establish bidirectional connections between intention and time, and improve model performance.

Validity check of each module of the model presented in this document. The experimental results of one-way joint modeling of slot, one-way joint modeling of slot to intention detection and two-way joint modeling of slot by intention detection are compared. The results are shown in Table 2. Although it is impossible to see which model performs better in the two-way joint model, the evaluation indexes of the two-way joint model are higher than those of the two-way joint model. This shows that the performance of bidirectional modeling is better than the performance of unidirectional collaborative modeling in joint modeling of intent detection and slot filling.

Results of Model in Chinese Dataset CAIS

Model CAIS
Slot F1 Acc Sent-acc
Intent➡Slot 88.4 95.8 85.5
Slot➡Intent 88.8 95.6 85.8
Our Model 89.2 96.3 87.9
Intent Detection and Slot Filling Analysis

To explore the extent to which Intent detection and slot filling tasks affect the overall performance of the model, using the results of CAIS data set on each model, using a discount plot to visualize the experimental detection results on each model, as shown in Figure 7. The horizontal coordinates of the folding diagram represent the individual model, and the vertical coordinates represent the correct rate of each category (sentence level, intention, and slot). Analysis shows that each model can accurately identify intent labels and slot labels corresponding to most examples. All three types of valid identification numbers are intendedslots Sentence level, therefore slot filling work affects the overall detection accuracy of the joint model.

Figure 7.

Sentence-level Analysis

Conclusions

In order to investigate the relationship between the two tasks of intention detection and space filling and the degree of influence on the joint model, a three-dimensional analysis of the joint model of intention detection and space filling based on attentional mechanisms was carried out, i.e. from intention detection to space extraction and then to filling.

Unidirectional modeling from slot filling to intention detection, and bidirectional modeling from intention detection to slot filling. The experimental results show that there are three evaluation results for F1 value. The intention accuracy and overall accuracy of bidirectional modeling for detecting and filling intentions are higher than those of joint modeling. Unidirectional. In addition, in both unidirectional and bidirectional common modes, slot filling has a greater impact on the overall detection performance of the model.

eISSN:
2470-8038
Lingua:
Inglese
Frequenza di pubblicazione:
4 volte all'anno
Argomenti della rivista:
Computer Sciences, other