Digital Library[ Search Result ]
Noise Injection for Natural Language Sentence Generation from Knowledge Base
http://doi.org/10.5626/JOK.2020.47.10.965
Generating a natural language sentence from Knowledge base is an operation of entering a triple in the Knowledge base to generate triple information, which is a natural language sentence containing the relationship between the entities. To solve the task of generating sentences from triples using a deep neural network, learning data consisting of many pairs of triples and natural language sentences are required. However, it is difficult to learn the model because the learning data composed in Korean is not yet released. To solve the deficiency of learning data, this paper proposes an unsupervised learning method that extracts keywords based on Korean Wikipedia sentence data and generates learning data using a noise injection technique. To evaluate the proposed method, we used gold-standard dataset produced by triples and sentence pairs. Consequently, the proposed noise injection method showed superior performances over normal unsupervised learning on various evaluation metrics including automatic and human evaluations.
Sentence Generation from Knowledge Base Triples Using Attention Mechanism Encoder-decoder
http://doi.org/10.5626/JOK.2019.46.9.934
In this paper, we have investigated the generation of natural language sentences by using Knowledge Base Triples data with a structured structure. In order to generate a sentence that expresses the triple, a LSTM (Long Short-term Memory Network) encoder-decoder structure is used along with an Attention Mechanism. The BLEU score and ROUGE score for the test data were 42.264 (BLEU-1), 32.441 (BLEU-2), 26.820 (BLEU-3), 24.446 (BLEU-4), and 47.341 and 0.8% (based on BLEU-1) for the data comparison model. In addition, the average of the top 10 test data BLEU scores was recorded as 99.393 (BLEU-1).
Regularizing Korean Conversational Model by Applying Denoising Mechanism
Tae-Hyeong Kim, Yunseok Noh, Seong-Bae Park, Se-Yeong Park
http://doi.org/10.5626/JOK.2018.45.6.572
A conversation system is a system that responds appropriately to input utterances. Recently, the sequence-to-sequence framework has been widely used as a conversation-learning model. However, the conversation model learned in such a way often generates a safe and dull response that does not provide appropriate information or sophisticated meaning. In addition, this model is also useless for input utterances appearing in various forms, such as with changed ending words or changed word order. To solve these problems, we propose a denoising response generation model applying a denoising mechanism. By injecting noise into original input, the proposed method creates a model that will stochastically experience new input made up of items that were not included in the original data during the training process. This data augmentation effect regularizes the model and allows the realization of a robust model. We evaluate our model using 90k input utterances-responses from Korean conversation pair data. The proposed model achieves better results compared to a baseline model on both ROUGE F1 score and qualitative evaluations by human annotators.
Search

Journal of KIISE
- ISSN : 2383-630X(Print)
- ISSN : 2383-6296(Electronic)
- KCI Accredited Journal
Editorial Office
- Tel. +82-2-588-9240
- Fax. +82-2-521-1352
- E-mail. chwoo@kiise.or.kr