Searching for just a few words should be enough to get started. If you need to make more complex queries, use the tips below to guide you.
Article type: Research Article
Authors: Fan, Yuweia | Shi, Leia; b; * | Yuan, Lua; *
Affiliations: [a] State Key Laboratory of Media Convergence and Communication, Communication University of China, Beijing, China | [b] Guangxi Key Laboratory of Trusted Software, Guilin University of Electronic Technology, Guilin, China
Correspondence: [*] Corresponding author. Lei Shi, State Key Laboratory of Media Convergence and Communication, Communication University of China, Beijing, China. E-mail: leikyshi@qq.com and Lu Yuan, State Key Laboratory of Media Convergence and Communication, Communication University of China, Beijing, China. E-mail: luyuan.bigdata@cuc.edu.cn.
Abstract: In the present day, online users are incentivized to engage in short text-based communication. These short texts harbor a significant amount of implicit information, including opinions, topics, and emotions, which are of notable value for both exploration and analysis. By alleviating the sparsity in short texts, topic models can be used to discover topics from large collections of short texts. While there is a large body of surveys focused on topic modeling, but only a few of them have focused on the short texts. This paper presents a comprehensive overview of topic modeling methods for short texts from a novel perspective. Firstly, it discusses short text probabilistic topic models and outlines the directions in which they can be improved. Secondly, it explores short text neural topic models, which can be categorized into three groups based on their underlying structures. In addition, this paper provides a detailed investigation of embedding methods in topic modeling. Moreover, various applications and corresponding works are surveyed, with a focus on short texts. The commonly used public corpora and evaluation indicators for topic modeling are also summarized. Finally, the advantages and disadvantages of short text topic modeling are discussed in detail, and future research directions are proposed.
Keywords: Short text, probabilistic topic model, neural topic model, word embeddings, deep learning
DOI: 10.3233/JIFS-223834
Journal: Journal of Intelligent & Fuzzy Systems, vol. 45, no. 2, pp. 1971-1990, 2023
IOS Press, Inc.
6751 Tepper Drive
Clifton, VA 20124
USA
Tel: +1 703 830 6300
Fax: +1 703 830 2300
sales@iospress.com
For editorial issues, like the status of your submitted paper or proposals, write to editorial@iospress.nl
IOS Press
Nieuwe Hemweg 6B
1013 BG Amsterdam
The Netherlands
Tel: +31 20 688 3355
Fax: +31 20 687 0091
info@iospress.nl
For editorial issues, permissions, book requests, submissions and proceedings, contact the Amsterdam office info@iospress.nl
Inspirees International (China Office)
Ciyunsi Beili 207(CapitaLand), Bld 1, 7-901
100025, Beijing
China
Free service line: 400 661 8717
Fax: +86 10 8446 7947
china@iospress.cn
For editorial issues, like the status of your submitted paper or proposals, write to editorial@iospress.nl
如果您在出版方面需要帮助或有任何建, 件至: editorial@iospress.nl