To Örebro University

oru.seÖrebro University Publications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Differentiable Sampling of Categorical Distributions Using the CatLog-Derivative Trick
Department of Computer Science, KU Leuven, Belgium.ORCID iD: 0000-0003-3136-0634
Department of Computer Science, KU Leuven, Belgium.ORCID iD: 0000-0003-3136-0634
Örebro University, School of Science and Technology. (Machine Perception Interaction Lab)ORCID iD: 0000-0001-5834-0188
2023 (English)In: Proceedings of the Conference on Neural Information Processing Systems, Neural Information Processing Systems Foundation Inc. , 2023, Vol. 36Conference paper, Published paper (Refereed)
Abstract [en]

Categorical random variables can faithfully represent the discrete and uncertain aspects of data as part of a discrete latent variable model. Learning in such models necessitates taking gradients with respect to the parameters of the categorical probability distributions, which is often intractable due to their combinatorial nature. A popular technique to estimate these otherwise intractable gradients is the Log-Derivative trick. This trick forms the basis of the well-known REINFORCE gradient estimator and its many extensions. While the Log-Derivative trick allows us to differentiate through samples drawn from categorical distributions, it does not take into account the discrete nature of the distribution itself. Our first contribution addresses this shortcoming by introducing the CatLog-Derivative trick– a variation of the Log-Derivative trick tailored towards categorical distributions. Secondly, we use the CatLog-Derivative trick to introduce IndeCateR, a novel and unbiased gradient estimator for the important case of products of independent categorical distributions with provably lower variance than REINFORCE. Thirdly, we empirically show that IndeCateR can be efficiently implemented and that its gradient estimates have significantly lower bias and variance for the same number of samples compared to the state of the art.

Place, publisher, year, edition, pages
Neural Information Processing Systems Foundation Inc. , 2023. Vol. 36
Series
Advances in Neural Information Processing Systems, ISSN 1049-5258 ; 36
National Category
Computer Sciences
Identifiers
URN: urn:nbn:se:oru:diva-110789ISI: 001229751901005Scopus ID: 2-s2.0-85191144370OAI: oai:DiVA.org:oru-110789DiVA, id: diva2:1828643
Conference
37th Annual Conference on Neural Information Processing Systems (NeurIPS 2023), New Orleans, Louisiana, USA, December 10-16, 2023
Funder
Wallenberg AI, Autonomous Systems and Software Program (WASP)EU, Horizon 2020, 952215
Note

This research received funding from the Flemish Government (AI Research Program), from the Flanders Research Foundation (FWO) under project G097720N and under EOS project No. 30992574, from the KU Leuven Research Fund (C14/18/062) and TAILOR, a project from the EU Horizon 2020 research and innovation programme under GA No. 952215. It is also supported by the Wallenberg AI, Autonomous Systems and Software Program (WASP) funded by the Knut and Alice Wallenberg-Foundation.

Available from: 2024-01-17 Created: 2024-01-17 Last updated: 2024-08-02Bibliographically approved

Open Access in DiVA

Differentiable Sampling of Categorical Distributions Using the CatLog-Derivative Trick(1417 kB)145 downloads
File information
File name FULLTEXT01.pdfFile size 1417 kBChecksum SHA-512
ed41708f783fdabaa9ca9cbcd12b29746841d4cde6712634b4cea1433e6d7c5604868ccc23af218c9cfa7de557f344f32ba650e191babbae94e3b3020fda4cc3
Type fulltextMimetype application/pdf

Scopus

Authority records

Zuidberg dos Martires, Pedro

Search in DiVA

By author/editor
De Smet, LennertSansone, EmanueleZuidberg dos Martires, Pedro
By organisation
School of Science and Technology
Computer Sciences

Search outside of DiVA

GoogleGoogle Scholar
Total: 145 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

urn-nbn

Altmetric score

urn-nbn
Total: 600 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf