Supervised scale-regularized linear convolutionary filters

Publikation: Bidrag til bog/antologi/rapportKonferencebidrag i proceedingsForskningfagfællebedømt

Standard

Supervised scale-regularized linear convolutionary filters. / Loog, Marco; Lauze, Francois Bernard.

Proceedings of BMVC 2017. British Machine Vision Conference, 2017.

Publikation: Bidrag til bog/antologi/rapportKonferencebidrag i proceedingsForskningfagfællebedømt

Harvard

Loog, M & Lauze, FB 2017, Supervised scale-regularized linear convolutionary filters. i Proceedings of BMVC 2017. British Machine Vision Conference, British Machine Vision Conference 2017, London, Storbritannien, 04/09/2017. <https://bmvc2017.london/proceedings/>

APA

Loog, M., & Lauze, F. B. (2017). Supervised scale-regularized linear convolutionary filters. I Proceedings of BMVC 2017 British Machine Vision Conference. https://bmvc2017.london/proceedings/

Vancouver

Loog M, Lauze FB. Supervised scale-regularized linear convolutionary filters. I Proceedings of BMVC 2017. British Machine Vision Conference. 2017

Author

Loog, Marco ; Lauze, Francois Bernard. / Supervised scale-regularized linear convolutionary filters. Proceedings of BMVC 2017. British Machine Vision Conference, 2017.

Bibtex

@inproceedings{4356660fb2d34f1caa5d9adf34710c21,
title = "Supervised scale-regularized linear convolutionary filters",
abstract = "We start by demonstrating that an elementary learning task—learning a linear filterfrom training data by means of regression—can be solved very efficiently for featurespaces of very high dimensionality. In a second step, firstly, acknowledging that suchhigh-dimensional learning tasks typically benefit from some form of regularization and,secondly, arguing that the problem of scale has not been taken care of in a very satis-factory manner, we come to a combined resolution of both of these shortcomings byproposing a technique that we coin scale regularization. This regularization problem canalso be solved relatively efficient. All in all, the idea is to properly control the scale of atrained filter, which we solve by introducing a specific regularization term into the overallobjective function. We demonstrate, on an artificial filter learning problem, the capabil-ities of our basic filter. In particular, we demonstrate that it clearly outperforms the defacto standard Tikhonov regularization, which is the one employed in ridge regression orWiener filtering.",
author = "Marco Loog and Lauze, {Francois Bernard}",
year = "2017",
month = jul,
language = "English",
booktitle = "Proceedings of BMVC 2017",
publisher = "British Machine Vision Conference",
note = "null ; Conference date: 04-09-2017 Through 07-09-2017",
url = "https://bmvc2017.london/",

}

RIS

TY - GEN

T1 - Supervised scale-regularized linear convolutionary filters

AU - Loog, Marco

AU - Lauze, Francois Bernard

PY - 2017/7

Y1 - 2017/7

N2 - We start by demonstrating that an elementary learning task—learning a linear filterfrom training data by means of regression—can be solved very efficiently for featurespaces of very high dimensionality. In a second step, firstly, acknowledging that suchhigh-dimensional learning tasks typically benefit from some form of regularization and,secondly, arguing that the problem of scale has not been taken care of in a very satis-factory manner, we come to a combined resolution of both of these shortcomings byproposing a technique that we coin scale regularization. This regularization problem canalso be solved relatively efficient. All in all, the idea is to properly control the scale of atrained filter, which we solve by introducing a specific regularization term into the overallobjective function. We demonstrate, on an artificial filter learning problem, the capabil-ities of our basic filter. In particular, we demonstrate that it clearly outperforms the defacto standard Tikhonov regularization, which is the one employed in ridge regression orWiener filtering.

AB - We start by demonstrating that an elementary learning task—learning a linear filterfrom training data by means of regression—can be solved very efficiently for featurespaces of very high dimensionality. In a second step, firstly, acknowledging that suchhigh-dimensional learning tasks typically benefit from some form of regularization and,secondly, arguing that the problem of scale has not been taken care of in a very satis-factory manner, we come to a combined resolution of both of these shortcomings byproposing a technique that we coin scale regularization. This regularization problem canalso be solved relatively efficient. All in all, the idea is to properly control the scale of atrained filter, which we solve by introducing a specific regularization term into the overallobjective function. We demonstrate, on an artificial filter learning problem, the capabil-ities of our basic filter. In particular, we demonstrate that it clearly outperforms the defacto standard Tikhonov regularization, which is the one employed in ridge regression orWiener filtering.

M3 - Article in proceedings

BT - Proceedings of BMVC 2017

PB - British Machine Vision Conference

Y2 - 4 September 2017 through 7 September 2017

ER -

ID: 183735818