TY - GEN
T1 - Bandit multiclass linear classification
T2 - 36th International Conference on Machine Learning, ICML 2019
AU - Beygelzimer, Alina
AU - Pál, Dávid
AU - Szörényi, Balázs
AU - Thiruvenkatachari, Devanathan
AU - Wei, Chen Yu
AU - Zhang, Chicheng
N1 - Publisher Copyright:
© 36th International Conference on Machine Learning, ICML 2019. All rights reserved.
PY - 2019
Y1 - 2019
N2 - We study the problem of efficient online multi-class linear classification with bandit feedback, where all examples belong to one of K classes and lie in the (¿-dimensional Euclidean space. Previous works have left open the challenge of designing efficient algorithms with finite mistake bounds when the data is linearly separable by a margin 7. In this work, we take a first step towards this problem. We consider two notions of linear separability, strong and weak. 1. Under the strong linear separability condition, we design an efficient algorithm that achieves a near-optimal mistake bound of o(K/γ2). 2. Under the more challenging weak linear separability condition, we design an efficient algorithm with a mistake bound of min(2õ(K log 2(1/γ)),2õ(√1/γ log K)).1 Our algorithm is based on kernel Perceptron and is inspired by the work of Klivans & Serve-dio (2008) on improperly learning intersection of halfspaces.
AB - We study the problem of efficient online multi-class linear classification with bandit feedback, where all examples belong to one of K classes and lie in the (¿-dimensional Euclidean space. Previous works have left open the challenge of designing efficient algorithms with finite mistake bounds when the data is linearly separable by a margin 7. In this work, we take a first step towards this problem. We consider two notions of linear separability, strong and weak. 1. Under the strong linear separability condition, we design an efficient algorithm that achieves a near-optimal mistake bound of o(K/γ2). 2. Under the more challenging weak linear separability condition, we design an efficient algorithm with a mistake bound of min(2õ(K log 2(1/γ)),2õ(√1/γ log K)).1 Our algorithm is based on kernel Perceptron and is inspired by the work of Klivans & Serve-dio (2008) on improperly learning intersection of halfspaces.
UR - http://www.scopus.com/inward/record.url?scp=85077954380&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85077954380&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85077954380
T3 - 36th International Conference on Machine Learning, ICML 2019
SP - 975
EP - 1011
BT - 36th International Conference on Machine Learning, ICML 2019
PB - International Machine Learning Society (IMLS)
Y2 - 9 June 2019 through 15 June 2019
ER -