Improved Algorithms for Efficient Active Learning Halfspaces with Massart and Tsybakov Noise

Chicheng Zhang, Yinan Li

Research output: Contribution to journalConference articlepeer-review

8 Scopus citations

Abstract

We give a computationally-efficient PAC active learning algorithm for d-dimensional homogeneous halfspaces that can tolerate Massart noise (Massart and Nédélec, 2006) and Tsybakov noise (Tsybakov, 2004). Specialized to the η-Massart noise setting, our algorithm achieves an information-theoretically near-optimal label complexity of (Equation presented) under a wide range of unlabeled data distributions (specifically, the family of “structured distributions” defined in Diakonikolas et al. (2020a)). Under the more challenging Tsybakov noise condition, we identify two subfamilies of noise conditions, under which our efficient algorithm provides label complexity guarantees strictly lower than passive learning algorithms.

Original languageEnglish (US)
Pages (from-to)4526-4527
Number of pages2
JournalProceedings of Machine Learning Research
Volume134
StatePublished - 2021
Event34th Conference on Learning Theory, COLT 2021 - Boulder, United States
Duration: Aug 15 2021Aug 19 2021

Keywords

  • Active learning
  • halfspaces
  • noise-tolerant learning

ASJC Scopus subject areas

  • Artificial Intelligence
  • Software
  • Control and Systems Engineering
  • Statistics and Probability

Fingerprint

Dive into the research topics of 'Improved Algorithms for Efficient Active Learning Halfspaces with Massart and Tsybakov Noise'. Together they form a unique fingerprint.

Cite this