BOUNDING THE VAPNIK-CHERVONENKIS DIMENSION OF CONCEPT CLASSES PARAMETERIZED BY REAL NUMBERS

被引:89
作者
GOLDBERG, PW [1 ]
JERRUM, MR [1 ]
机构
[1] UNIV EDINBURGH, DEPT COMP SCI, EDINBURGH EH9 3JZ, MIDLOTHIAN, SCOTLAND
关键词
CONCEPT LEARNING; INFORMATION THEORY; VAPNIK-CHERVONENKIS DIMENSION; MILNORS THEOREM;
D O I
10.1007/BF00993408
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The Vapnik-Chervonenkis (V-C) dimension is an important combinatorial tool in the analysis of learning problems in the PAC framework. For polynomial learnability, we seek upper bounds on the V-C dimension that are polynomial in the syntactic complexity of concepts. Such upper bounds are automatic for discrete concept classes, but hitherto little has been known about what general conditions guarantee polynomial bounds on V-C dimension for classes in which concepts and examples are represented by tuples of real numbers. In this paper, we show that for two general kinds of concept class the V-C dimension is polynomially bounded in the number of real numbers used to define a problem instance. One is classes where the criterion for membership of an instance in a concept can be expressed as a formula (in the first-order theory of the reals) with fixed quantification depth and exponentially-bounded length, whose atomic predicates are polynomial inequalities of exponentially-bounded degree. The other is classes where containment of an instance in a concept is testable in polynomial time, assuming we may compute standard arithmetic operations on reals exactly in constant time. Our results show that in the continuous case, as in the discrete, the real barrier to efficient learning in the Occam sense is complexity-theoretic and not information-theoretic. We present examples to show how these results apply to concept classes defined by geometrical figures and neural nets, and derive polynomial bounds on the V-C dimension for these classes.
引用
收藏
页码:131 / 148
页数:18
相关论文
共 27 条
  • [1] ANTHONY M, 1992, COMPUTATIONAL LEARNI
  • [2] ATTALAH MJ, 1993, INFORMATION PROCESSI, V17, P207
  • [3] What Size Net Gives Valid Generalization?
    Baum, Eric B.
    Haussler, David
    [J]. NEURAL COMPUTATION, 1989, 1 (01) : 151 - 160
  • [4] BENDAVID S, 1993, 6TH P ANN ACM C COMP, P327
  • [5] OCCAM RAZOR
    BLUMER, A
    EHRENFEUCHT, A
    HAUSSLER, D
    WARMUTH, MK
    [J]. INFORMATION PROCESSING LETTERS, 1987, 24 (06) : 377 - 380
  • [6] CENTRAL LIMIT-THEOREMS FOR EMPIRICAL MEASURES
    DUDLEY, RM
    [J]. ANNALS OF PROBABILITY, 1978, 6 (06) : 899 - 929
  • [7] A GENERAL LOWER BOUND ON THE NUMBER OF EXAMPLES NEEDED FOR LEARNING
    EHRENFEUCHT, A
    HAUSSLER, D
    KEARNS, M
    VALIANT, L
    [J]. INFORMATION AND COMPUTATION, 1989, 82 (03) : 247 - 261
  • [8] GOLDBERG P, 1992, THESIS U EDINBURGH
  • [9] Hall M., 1967, COMBINATORIAL THEORY
  • [10] Haussler D., 1988, 29th Annual Symposium on Foundations of Computer Science (IEEE Cat. No.88CH2652-6), P100, DOI 10.1109/SFCS.1988.21928