Topic Modeling with Transformers for Sentence-Level Using Coronavirus Corpus

Dublin Core

Title

Topic Modeling with Transformers for Sentence-Level Using Coronavirus Corpus

Subject

Topic Model
Sentence-Level
Machine Learning
LDA
BERT
BERTopic

Description

A Topic Model is a class of generative probabilistic models which has gained widespread use in computer science in recent years, especially in the field of text mining and information retrieval. Since it was first proposed, it has received a large amount of attention and general interest among scientists in many research areas. It allows us to discover the mix of hidden or "latent" subjects that differs from one document to another in a given corpus. But since topic modeling usually requires the prior definition of some parameters - above all the number of topics k to be discovered -, model evaluation is decisive to identify an "optimal" set of parameters for the specific data.  Latent Dirichlet allocation (LDA) and Bidirectional Encoder Representations from Transformers Topic (BerTopic) are the two most popular topic modeling techniques. LDA uses a probabilistic approach whereas BerTopic uses transformers (BERT embeddings) and class-based TF-IDF to create dense clusters.

Creator

Mifrah, Sara
Benlahmar, El Habib

Source

International Journal of Interactive Mobile Technologies (iJIM); Vol. 16 No. 17 (2022); pp. 50-59
1865-7923

Publisher

International Association of Online Engineering (IAOE), Vienna, Austria

Date

2022-09-08

Rights

Copyright (c) 2022 Sara Mifrah
https://creativecommons.org/licenses/by/4.0

Relation

Format

application/pdf

Language

eng

Type

info:eu-repo/semantics/article
info:eu-repo/semantics/publishedVersion
Peer-reviewed Article

Identifier

Citation

Sara Mifrah and El Benlahmar Habib, Topic Modeling with Transformers for Sentence-Level Using Coronavirus Corpus, International Association of Online Engineering (IAOE), Vienna, Austria, 2022, accessed November 6, 2024, https://igi.indrastra.com/items/show/2370

Social Bookmarking