To Örebro University

oru.seÖrebro University Publications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Modeling PU learning using probabilistic logic programming
Department of Computer Science, Leuven.AI, KU Leuven, Leuven, Belgium.
Örebro University, School of Science and Technology. Department of Computer Science, Leuven.AI, KU Leuven, Leuven, Belgium. (AASS)ORCID iD: 0000-0002-6860-6303
Department of Computer Science, Leuven.AI, KU Leuven, Leuven, Belgium.
2024 (English)In: Machine Learning, ISSN 0885-6125, E-ISSN 1573-0565, Vol. 113, no 3, p. 1351-1372Article in journal (Refereed) Published
Abstract [en]

The goal of learning from positive and unlabeled (PU) examples is to learn a classifier that predicts the posterior class probability. The challenge is that the available labels in the data are determined by (1) the true class, and (2) the labeling mechanism that selects which positive examples get labeled, where often certain examples have a higher probability to be selected than others. Incorrectly assuming an unbiased labeling mechanism leads to learning a biased classifier. Yet, this is what most existing methods do. A handful of methods makes more realistic assumptions, but they are either so general that it is impossible to distinguish between the effects of the true classification and of the labeling mechanism, or too restrictive to correctly model the real situation, or require knowledge that is typically unavailable. This paper studies how to formulate and integrate more realistic assumptions for learning better classifiers, by exploiting the strengths of probabilistic logic programming (PLP). Concretely, (1) we propose PU ProbLog: a PLP-based general method that allows to (partially) model the labeling mechanism. (2) We show that our method generalizes existing methods, in the sense that it can model the same assumptions. (3) Thanks to the use of PLP, our method supports also PU learning in relational domains. (4) Our empirical analysis shows that partially modeling the labeling bias, improves the learned classifiers.

Place, publisher, year, edition, pages
Springer, 2024. Vol. 113, no 3, p. 1351-1372
Keywords [en]
Modeling, Positive unlabeled learning, Probabilistic logic programming, Unidentifiability, Weak supervision
National Category
Computer Sciences
Identifiers
URN: urn:nbn:se:oru:diva-110216DOI: 10.1007/s10994-023-06461-3ISI: 001114982400001Scopus ID: 2-s2.0-85178334835OAI: oai:DiVA.org:oru-110216DiVA, id: diva2:1819467
Funder
EU, Horizon 2020Wallenberg AI, Autonomous Systems and Software Program (WASP)Knut and Alice Wallenberg Foundation
Note

This research received funding from the Flemish Government (AI Research Program), from the Research Foundation-Flanders under the Data-driven logistics project (FWO-S007318N), from the KU Leuven Research Fund (C14/18/062) and TAILOR, a project from the EU Horizon 2020 research and innovation programme under GA No 952215. LDR is also supported by the Wallenberg AI, Autonomous Systems and Software Program (WASP) funded by the Knut and Alice Wallenberg-Foundation.

Available from: 2023-12-14 Created: 2023-12-14 Last updated: 2024-07-24Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

De Raedt, Luc

Search in DiVA

By author/editor
De Raedt, Luc
By organisation
School of Science and Technology
In the same journal
Machine Learning
Computer Sciences

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 27 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf