An Optimized k-NN Approach for Classification on Imbalanced Datasets with Missing Data
Research output: Chapter in Book/Report/Conference proceeding › Conference contribution › Scientific › peer-review
Standard
An Optimized k-NN Approach for Classification on Imbalanced Datasets with Missing Data. / Ozan, Ezgi Can; Riabchenko, Ekaterina; Kiranyaz, Serkan; Gabbouj, Moncef.
Advances in Intelligent Data Analysis XV: 15th International Symposium, IDA 2016, Stockholm, Sweden, October 13-15, 2016, Proceedings. Springer, 2016. p. 387-392 (Lecture Notes in Computer Science; Vol. 9897).Research output: Chapter in Book/Report/Conference proceeding › Conference contribution › Scientific › peer-review
Harvard
APA
Vancouver
Author
Bibtex - Download
}
RIS (suitable for import to EndNote) - Download
TY - GEN
T1 - An Optimized k-NN Approach for Classification on Imbalanced Datasets with Missing Data
AU - Ozan, Ezgi Can
AU - Riabchenko, Ekaterina
AU - Kiranyaz, Serkan
AU - Gabbouj, Moncef
N1 - EXT="Kiranyaz, Serkan"
PY - 2016
Y1 - 2016
N2 - In this paper, we describe our solution for the machine learning prediction challenge in IDA 2016. For the given problem of 2-class classification on an imbalanced dataset with missing data, we first develop an imputation method based on k-NN to estimate the missing values. Then we define a tailored representation for the given problem as an optimization scheme, which consists of learned distance and voting weights for k-NN classification. The proposed solution performs better in terms of the given challenge metric compared to the traditional classification methods such as SVM, AdaBoost or Random Forests.
AB - In this paper, we describe our solution for the machine learning prediction challenge in IDA 2016. For the given problem of 2-class classification on an imbalanced dataset with missing data, we first develop an imputation method based on k-NN to estimate the missing values. Then we define a tailored representation for the given problem as an optimization scheme, which consists of learned distance and voting weights for k-NN classification. The proposed solution performs better in terms of the given challenge metric compared to the traditional classification methods such as SVM, AdaBoost or Random Forests.
U2 - 10.1007/978-3-319-46349-0_34
DO - 10.1007/978-3-319-46349-0_34
M3 - Conference contribution
SN - 978-3-319-46348-3
T3 - Lecture Notes in Computer Science
SP - 387
EP - 392
BT - Advances in Intelligent Data Analysis XV
PB - Springer
ER -