Informative Language Encoding by Variational Autoencoders Using Transformer
- Title
- Informative Language Encoding by Variational Autoencoders Using Transformer
- Author
- 이기천
- Keywords
- natural language processing; transformer; variational autoencoder; text mining
- Issue Date
- 2022-08
- Publisher
- MDPI
- Citation
- APPLIED SCIENCES-BASEL, v. 12, NO. 16, article no. 7968, Page. 1-16
- Abstract
- In natural language processing (NLP), Transformer is widely used and has reached the state-of-the-art level in numerous NLP tasks such as language modeling, summarization, and classification. Moreover, a variational autoencoder (VAE) is an efficient generative model in representation learning, combining deep learning with statistical inference in encoded representations. However, the use of VAE in natural language processing often brings forth practical difficulties such as a posterior collapse, also known as Kullback–Leibler (KL) vanishing. To mitigate this problem, while taking advantage of the parallelization of language data processing, we propose a new language representation model as the integration of two seemingly different deep learning models, which is a Transformer model solely coupled with a variational autoencoder. We compare the proposed model with previous works, such as a VAE connected with a recurrent neural network (RNN). Our experiments with four real-life datasets show that implementation with KL annealing mitigates posterior collapses. The results also show that the proposed Transformer model outperforms RNN-based models in reconstruction and representation learning, and that the encoded representations of the proposed model are more informative than other tested models.
- URI
- https://www.mdpi.com/2076-3417/12/16/7968https://repository.hanyang.ac.kr/handle/20.500.11754/178075
- ISSN
- 2076-3417;2076-3417
- DOI
- 10.3390/app12167968
- Appears in Collections:
- COLLEGE OF ENGINEERING[S](공과대학) > INDUSTRIAL ENGINEERING(산업공학과) > Articles
- Files in This Item:
- Informative Language Encoding by Variational Autoencoders Using Transformer.pdfDownload
- Export
- RIS (EndNote)
- XLS (Excel)
- XML