Bayesian active learning for maximal information gain on model parameters

Publikation: Bidrag til bog/antologi/rapportKonferencebidrag i proceedingsForskningfagfællebedømt

Dokumenter

  • Fulltext

    Accepteret manuskript, 4,39 MB, PDF-dokument

The fact that machine learning models, despite their advancements, are still trained on randomly gathered data is proof that a lasting solution to the problem of optimal data gathering has not yet been found. In this paper, we investigate whether a Bayesian approach to the classification problem can provide assumptions under which one is guaranteed to perform at least as good as random sampling. For a logistic regression model, we show that maximal expected information gain on model parameters is a promising criterion for selecting samples, assuming that our classification model is well-matched to the data. Our derived criterion is closely related to the maximum model change. We experiment with data sets which satisfy this assumption to varying degrees to see how sensitive our performance is to the violation of our assumption in practice.

OriginalsprogEngelsk
TitelProceedings of ICPR 2020 - 25th International Conference on Pattern Recognition
Antal sider8
ForlagIEEE
Publikationsdato2020
Sider10524-10531
Artikelnummer9411962
ISBN (Elektronisk)9781728188089
DOI
StatusUdgivet - 2020
Begivenhed25th International Conference on Pattern Recognition, ICPR 2020 - Virtual, Milan, Italien
Varighed: 10 jan. 202115 jan. 2021

Konference

Konference25th International Conference on Pattern Recognition, ICPR 2020
LandItalien
ByVirtual, Milan
Periode10/01/202115/01/2021

Antal downloads er baseret på statistik fra Google Scholar og www.ku.dk


Ingen data tilgængelig

ID: 286999036