Redirigiendo al acceso original de articulo en 20 segundos...
Inicio  /  Applied Sciences  /  Vol: 12 Par: 3 (2022)  /  Artículo
ARTÍCULO
TITULO

Fine-Tuning BERT Models for Intent Recognition Using a Frequency Cut-Off Strategy for Domain-Specific Vocabulary Extension

Fernando Fernández-Martínez    
Cristina Luna-Jiménez    
Ricardo Kleinlein    
David Griol    
Zoraida Callejas and Juan Manuel Montero    

Resumen

Intent recognition is a key component of any task-oriented conversational system. The intent recognizer can be used first to classify the user?s utterance into one of several predefined classes (intents) that help to understand the user?s current goal. Then, the most adequate response can be provided accordingly. Intent recognizers also often appear as a form of joint models for performing the natural language understanding and dialog management tasks together as a single process, thus simplifying the set of problems that a conversational system must solve. This happens to be especially true for frequently asked question (FAQ) conversational systems. In this work, we first present an exploratory analysis in which different deep learning (DL) models for intent detection and classification were evaluated. In particular, we experimentally compare and analyze conventional recurrent neural networks (RNN) and state-of-the-art transformer models. Our experiments confirmed that best performance is achieved by using transformers. Specifically, best performance was achieved by fine-tuning the so-called BETO model (a Spanish pretrained bidirectional encoder representations from transformers (BERT) model from the Universidad de Chile) in our intent detection task. Then, as the main contribution of the paper, we analyze the effect of inserting unseen domain words to extend the vocabulary of the model as part of the fine-tuning or domain-adaptation process. Particularly, a very simple word frequency cut-off strategy is experimentally shown to be a suitable method for driving the vocabulary learning decisions over unseen words. The results of our analysis show that the proposed method helps to effectively extend the original vocabulary of the pretrained models. We validated our approach with a selection of the corpus acquired with the Hispabot-Covid19 system obtaining satisfactory results.

 Artículos similares

       
 
Huidong Tang, Sayaka Kamei and Yasuhiko Morimoto    
Text classification is widely studied in natural language processing (NLP). Deep learning models, including large pre-trained models like BERT and DistilBERT, have achieved impressive results in text classification tasks. However, these models? robustnes... ver más
Revista: Algorithms

 
Marios Koniaris, Dimitris Galanis, Eugenia Giannini and Panayiotis Tsanakas    
The increasing amount of legal information available online is overwhelming for both citizens and legal professionals, making it difficult and time-consuming to find relevant information and keep up with the latest legal developments. Automatic text summ... ver más
Revista: Information

 
Mamatjan Abdurxit, Turdi Tohti and Askar Hamdulla    
Biomedical entity linking is an important research problem for many downstream tasks, such as biomedical intelligent question answering, information retrieval, and information extraction. Biomedical entity linking is the task of mapping mentions in medic... ver más
Revista: Applied Sciences

 
Mukul Jaggi, Priyanka Mandal, Shreya Narang, Usman Naseem and Matloob Khushi    
Stock price prediction can be made more efficient by considering the price fluctuations and understanding people?s sentiments. A limited number of models understand financial jargon or have labelled datasets concerning stock price change. To overcome thi... ver más