USING MULTI-LAYER LSTMS FOR QUESTION RETRIEVAL | Huế | TNU Journal of Science and Technology

USING MULTI-LAYER LSTMS FOR QUESTION RETRIEVAL

About this article

Received: 01/04/22                Revised: 26/05/22                Published: 27/05/22

Authors

Luong Thi Minh Hue Email to author, TNU - University of Information and Communication Technology

Abstract


Question retrieval is one of the important problems in the Community Question Answering system. The biggest challenge of this problem is the lexical gap between the words and phrases of the first and second question. Although there are many studies applied to this problem, the exploitation of multi-layer LSTM model has not been tested on this problem. In this paper, we exploit a multi-layer LSTM model applied to the problem of finding similar questions for the purpose of exploiting hidden semantics of sentences. The multi-layer LSTM model is capable of synthesizing semantics by multiple layers and exploits hidden semantics through many layers. Our model learned the semantics of sentences and improved the performance of finding question. The results show that the model with 3 layers gives the best results compared to the original LSTM model and other multi-layer models on the 2017 semeval dataset for the problem of finding similar questions.

Keywords


LSTM; NLP; Deep learning; CQA; Multi-layerLSTM

References


[1] G. Zhou, Y. Chen, D. Zeng, and J. Zhao, “Towards faster and better retrieval models for question search,” In Proceedings of the 22nd ACM International Conference on Information Knowledge Management, CIKM13, New York, NY, USA. Association for Computing Machinery, 2013, pp. 2139-2148.

[2] G. Zhou, T. He, J. Zhao, and P. Hu, “Learning continuous word embedding with metadata for question retrieval in community question answering,” CIKM13, vol. 01, pp. 250-259, 2015.

[3] L. Cai, G. Zhou, K. Liu, and J. Zhao, “Learning the latent topics for question retrieval in community QA,” In Proceedings of 5th International Joint Conference on Natural Language Processing, Chiang Mai, Thailand, November. Asian Federation of Natural Language Processing, 2011, pp. 273-281.

[4] W. Wu, X. Sun, and H. Wang, “Question condensing networks for answer selection in community question answering,” In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Melbourne, Australia, July. Association for Computational Linguistics, 2018, pp. 1746-1755.

[5] Y. Tay, A. T. Luu, and S. C. Hui, “Enabling efficient question answer retrieval via hyperbolic neural networks,” CoRR, pp. 265-274, 2017, doi: abs/1707.07847.

[6] S. Robertson, S. Walker, S. Jones, M. M. HancockBeaulieu, and M. Gatford, “Okapi at trec 3,” In Overview of the Third Text REtrieval Conference (TREC-3), January, 1995.

[7] X. Cao, G. Cong, B. Cui, C. S. Jensen, and C. Zhang, “The use of categorization information in language models for question retrieval,” In Proceedings of the 18th ACM Conference on Information and Knowledge Management, CIKM ’09, New York, NY, USA. Association for Computing Machinery, 2019, pp. 265-274.

[8] D. M. Blei, A. Y. Ng, and M. I. Jordan, “Latent dirichlet allocation,” In T. G. Dietterich, S. Becker, and Z. Ghahramani, editors, Advances in Neural Information Processing Systems 14, pp. 601-608. MIT Press, 2002.

[9] P. Nakov, D. Hoogeveen, L. Màrquez, A. Moschitti, H. Mubarak, T. Baldwin, and K. Verspoor, “SemEval-2017 task 3: Community question answering,” In Proceedings of the 11th International Workshop on Semantic Evaluation (SemEval-2017), Vancouver, Canada, August. Association for Computational Linguistics, 2017, pp. 27-48.

[10] S. Filice, G. Da San Martino, and A. Moschitti, “KeLP at SemEval-2017 task 3: Learning pairwise patterns in community question answering,” In Proceedings of the 11th International Workshop on Semantic Evaluation (SemEval-2017), Vancouver, Canada, August. Association for Computational Linguistics, 2017, pp. 326-333.

[11] M. Tan, B. Xiang, and B. Zhou, “LSTM-based Deep Learning Models for non-factoid answer selection,” 2015. [Online]. Available: https://arxiv.org/abs/1511.04108. [Accessed May 2021].

[12] D. Britz, A. Goldie, M.-T. Luong, and Q. Le, “Massive Exploration of Neural Machine Translation Architectures,” In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, Copenhagen, Denmark. Association for Computational Linguistics, 2017, pp. 1442-1451.




DOI: https://doi.org/10.34238/tnu-jst.5799

Refbacks

  • There are currently no refbacks.
TNU Journal of Science and Technology
Rooms 408, 409 - Administration Building - Thai Nguyen University
Tan Thinh Ward - Thai Nguyen City
Phone: (+84) 208 3840 288 - E-mail: jst@tnu.edu.vn
Based on Open Journal Systems
©2018 All Rights Reserved