I'm dealing with topic modeling for short text and have come across three models that focus on the same: The biterm topic model (BTM), the word network topic model (WNTM) and the latent-feature LDA (LF-LDA).
I know that for conventional LDA (I have implemented it using the R package topicmodels), the unstructured shape of text documents is converted to a computer-readable format via the construction of a Document-Term matrix (DTM).
I'm wondering if the above mentioned models use a similar way for implementation, especially if they also create a matrix that is similar to DTM. Does anyone know that? Unfortunately I couldn't find that information by reading the original papers.
Thank you in advance!
topicmodels
,lda
, ortext2vec
, each using slightly different sampling/estimation algorithms. – Manuel Bickel