Document Type

Technical Report

Publication Date

1992-09-22

Filename

WUCS-92-33.pdf

DOI:

10.7936/K7PC30QS

Technical Report Number

WUCS-92-33

Abstract

In this paper, we study the sample complexity of weak learning. That is, we ask how much data must be collected from an unknown distribution in order to extract a small but significant advantage in prediction. We show that it is important to distinguish between those learning algorithms that output deterministic hypotheses and those that output randomized hypotheses. We prove that in the weak learning model, any algorithm using deterministic hypotheses to weakly learn a class of Vapnik-Chervonenkis dimension d(n) requires (omega) (square root of d(n) examples.

Comments

Permanent URL: http://dx.doi.org/10.7936/K7PC30QS

Share

COinS