Artículo
On semi-supervised learning
Fecha de publicación:
12/2020
Editorial:
Springer
Revista:
Test
ISSN:
1133-0686
Idioma:
Inglés
Tipo de recurso:
Artículo publicado
Clasificación temática:
Resumen
Major efforts have been made, mostly in the machine learning literature, to construct good predictors combining unlabelled and labelled data. These methods are known as semi-supervised. They deal with the problem of how to take advantage, if possible, of a huge amount of unlabelled data to perform classification in situations where there are few labelled data. This is not always feasible: it depends on the possibility to infer the labels from the unlabelled data distribution. Nevertheless, several algorithms have been proposed recently. In this work, we present a new method that, under almost necessary conditions, attains asymptotically the performance of the best theoretical rule when the size of the unlabelled sample goes to infinity, even if the size of the labelled sample remains fixed. Its performance and computational time are assessed through simulations and in the well- known “Isolet” real data of phonemes, where a strong dependence on the choice of the initial training sample is shown. The main focus of this work is to elucidate when and why semi-supervised learning works in the asymptotic regime described above. The set of necessary assumptions, although reasonable, show that semi-parametric methods only attain consistency for very well-conditioned problems.
Palabras clave:
CONSISTENCY
,
SEMI-SUPERVISED LEARNING
,
SMALL TRAINING SAMPLE
Archivos asociados
Licencia
Identificadores
Colecciones
Articulos (IC)
Articulos de INSTITUTO DE CALCULO
Articulos de INSTITUTO DE CALCULO
Citación
Cholaquidis, A.; Fraiman, R.; Sued, Raquel Mariela; On semi-supervised learning; Springer; Test; 29; 4; 12-2020; 914-937
Compartir
Altmétricas