Skip to yearly menu bar Skip to main content


Poster
in
Workshop: The 4th Workshop on practical ML for Developing Countries: learning under limited/low resource settings

S2VNTM: SEMI-SUPERVISED VMF NEURAL TOPIC MODELING

Weijie Xu · Jay Desai · Srinivasan Sengamedu · Xiaoyu Jiang · Francis Iannacci


Abstract:

Language model based methods are powerful techniques for text classification. However, the models have several shortcomings. (1) It is difficult to integrate human knowledge such as keywords. (2) It needs a lot of resources to train the models. (3) It relied on large text data to pretrain. In this paper, we propose Semi- Supervised vMF Neural Topic Modeling (S2vNTM) to overcome these difficulties. S2vNTM takes a few seed keywords as input for topics. S2vNTM leverages the pattern of keywords to identify potential topics, as well as optimize the quality of topics’ keywords sets. Across a variety of datasets, S2vNTM outperforms existing semi-supervised topic modeling methods in classification accuracy with limited keywords provided. S2vNTM is at least twice as fast as baselines.

Chat is not available.