Skip to main navigation Skip to search Skip to main content

Learning Functional Distributions with Private Labels

  • Changlong Wu
  • , Yifan Wang
  • , Ananth Grama
  • , Wojciech Szpankowski

Research output: Contribution to journalConference articlepeer-review

Abstract

We study the problem of learning functional distributions in the presence of noise. A functional is a map from the space of features to distributions over a set of labels, and is often assumed to belong to a known class of hypotheses F. Features are generated by a general random process and labels are sampled independently from feature-dependent distributions. In privacy sensitive applications, labels are passed through a noisy kernel. We consider online learning, where at each time step, a predictor attempts to predict the actual (label) distribution given only the features and noisy labels in prior steps. The performance of the predictor is measured by the expected KL-risk that compares the predicted distributions to the underlying truth. We show that the minimax expected KL-risk is of order Θ̃(pT log |F|) for finite hypothesis class F and any non-trivial noise level. We then extend this result to general infinite classes via the concept of stochastic sequential covering and provide matching lower and upper bounds for a wide range of natural classes.

Original languageEnglish (US)
Pages (from-to)37728-37744
Number of pages17
JournalProceedings of Machine Learning Research
Volume202
StatePublished - 2023
Externally publishedYes
Event40th International Conference on Machine Learning, ICML 2023 - Honolulu, United States
Duration: Jul 23 2023Jul 29 2023

ASJC Scopus subject areas

  • Artificial Intelligence
  • Software
  • Control and Systems Engineering
  • Statistics and Probability

Fingerprint

Dive into the research topics of 'Learning Functional Distributions with Private Labels'. Together they form a unique fingerprint.

Cite this