Publication result detail

Learning Document Embeddings Along With Their Uncertainties

KESIRAJU, S.; PLCHOT, O.; BURGET, L.; GANGASHETTY, S.

Original Title

Learning Document Embeddings Along With Their Uncertainties

English Title

Learning Document Embeddings Along With Their Uncertainties

Type

WoS Article

Original Abstract

Majority of the text modeling techniques yield onlypoint-estimates of document embeddings and lack in capturingthe uncertainty of the estimates. These uncertainties give a notionof how well the embeddings represent a document. We presentBayesian subspace multinomial model (Bayesian SMM), a generativelog-linear model that learns to represent documents in theform of Gaussian distributions, thereby encoding the uncertaintyin its covariance. Additionally, in the proposed Bayesian SMM,we address a commonly encountered problem of intractabilitythat appears during variational inference in mixed-logit models.We also present a generative Gaussian linear classifier for topicidentification that exploits the uncertainty in document embeddings.Our intrinsic evaluation using perplexity measure showsthat the proposed Bayesian SMM fits the unseen test data betteras compared to the state-of-the-art neural variational documentmodel on (Fisher) speech and (20Newsgroups) text corpora. Ourtopic identification experiments showthat the proposed systems arerobust to over-fitting on unseen test data. The topic ID results showthat the proposedmodel outperforms state-of-the-art unsupervisedtopic models and achieve comparable results to the state-of-the-artfully supervised discriminative models.

English abstract

Majority of the text modeling techniques yield onlypoint-estimates of document embeddings and lack in capturingthe uncertainty of the estimates. These uncertainties give a notionof how well the embeddings represent a document. We presentBayesian subspace multinomial model (Bayesian SMM), a generativelog-linear model that learns to represent documents in theform of Gaussian distributions, thereby encoding the uncertaintyin its covariance. Additionally, in the proposed Bayesian SMM,we address a commonly encountered problem of intractabilitythat appears during variational inference in mixed-logit models.We also present a generative Gaussian linear classifier for topicidentification that exploits the uncertainty in document embeddings.Our intrinsic evaluation using perplexity measure showsthat the proposed Bayesian SMM fits the unseen test data betteras compared to the state-of-the-art neural variational documentmodel on (Fisher) speech and (20Newsgroups) text corpora. Ourtopic identification experiments showthat the proposed systems arerobust to over-fitting on unseen test data. The topic ID results showthat the proposedmodel outperforms state-of-the-art unsupervisedtopic models and achieve comparable results to the state-of-the-artfully supervised discriminative models.

Keywords

Bayesian methods, embeddings, topic identification.

Key words in English

Bayesian methods, embeddings, topic identification.

Authors

KESIRAJU, S.; PLCHOT, O.; BURGET, L.; GANGASHETTY, S.

RIV year

2021

Released

27.07.2020

ISBN

2329-9290

Periodical

IEEE-ACM Transactions on Audio Speech and Language Processing

Volume

2020

Number

28

State

United States of America

Pages from

2319

Pages to

2332

Pages count

14

URL

BibTex

@article{BUT168164,
  author="Santosh {Kesiraju} and Oldřich {Plchot} and Lukáš {Burget} and Suryakanth V {Gangashetty}",
  title="Learning Document Embeddings Along With Their Uncertainties",
  journal="IEEE-ACM Transactions on Audio Speech and Language Processing",
  year="2020",
  volume="2020",
  number="28",
  pages="2319--2332",
  doi="10.1109/TASLP.2020.3012062",
  issn="2329-9290",
  url="https://ieeexplore.ieee.org/document/9149686"
}

Documents