Abstract |
: |
Spoken Language Understanding (SLU) is a part of the Spoken dialogue system (SDS) and it assists to achieve the user's goal. SLU maps user utterance to a logical structure that can be easily understood by the computer. SLU accomplish this task by using deep learning models such as Convolution Neural Network (CNN) from machine learning. These machine-learning models suffer from an Internal Covariant Shift (ICS). Due to ICS, the learning efficiency of the deep model has reduced which in turn reduces the learning efficiency of SLU. An ICS in SLU can be reduced by techniques proposed in machine learning. The techniques are batch normalization, cosine normalization, group normalization, layer normalization, and instance normalization. The work in the paper extends these techniques to a deep model in SLU to mitigate the ICS and to enhance the performance of SLU. The results show improvement in SLU performance with normalization. The speed of training is also improved with normalization. The evaluation parameters are F-score, accuracy, and detection rate. When the SLU is trained with less training data, SLU with the instance normalization displayed good results. SLU with instance normalization displayed minimum variation in the learning curve. SLU with batch normalization displayed better accuracy, detection rate, and F-score for both area and price slot. In terms of training time, SLU with cosine normalization and batch normalization required the least time. |