PAUTOMAC: a probabilistic automata and hidden Markov models learning competition

被引:28
作者
Verwer, Sicco [1 ]
Eyraud, Remi [2 ]
de la Higuera, Colin [3 ]
机构
[1] Radboud Univ Nijmegen, Inst Comp & Informat Sci, NL-6525 ED Nijmegen, Netherlands
[2] Univ Aix Marseille, Lab Informat Fondamentale Marseille, QARMA Team, Marseille, France
[3] Univ Nantes, TALN Team, Lab Informat Nantes Atlantique, Nantes 1, France
关键词
Grammatical inference; Probabilistic automata; Hidden Markov models; Programming competition; FINITE-STATE MACHINES; GRAMMATICAL INFERENCE; PAC-LEARNABILITY; LANGUAGES; DISTRIBUTIONS; MULTIPLICITY; RULE;
D O I
10.1007/s10994-013-5409-9
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Approximating distributions over strings is a hard learning problem. Typical techniques involve using finite state machines as models and attempting to learn these; these machines can either be hand built and then have their weights estimated, or built by grammatical inference techniques: the structure and the weights are then learned simultaneously. The Probabilistic Automata learning Competition (PAUTOMAC), run in 2012, was the first grammatical inference challenge that allowed the comparison between these methods and algorithms. Its main goal was to provide an overview of the state-of-the-art techniques for this hard learning problem. Both artificial data and real data were presented and contestants were to try to estimate the probabilities of strings. The purpose of this paper is to describe some of the technical and intrinsic challenges such a competition has to face, to give a broad state of the art concerning both the problems dealing with learning grammars and finite state machines and the relevant literature. This paper also provides the results of the competition and a brief description and analysis of the different approaches the main participants used.
引用
收藏
页码:129 / 154
页数:26
相关论文
共 82 条
[1]  
ABE N, 1992, MACH LEARN, V9, P205, DOI 10.1007/BF00992677
[2]  
ANGLUIN D, 1988, DCSRR614 YAL U
[3]  
[Anonymous], THESIS U CALIFORNIA
[4]  
[Anonymous], 2007, The Minimum Description Length Principle
[5]  
[Anonymous], P 11 INT C GRAMM INF
[6]  
[Anonymous], 2012, J MACHINE LEARNING R
[7]  
Bailly R., 2011, JMLR WORKSH C P P AS, V20, P147
[8]  
Bailly R., 2009, P 26 ANN INT C MACH, P33, DOI [10.1145/ 1553374.1553379, DOI 10.1145/1553374.1553379, 10.1145/1553374.1553379]
[9]  
Balle Borja, 2012, P MACHINE LEARNING R, V21, P34
[10]  
Baum L. E., 1972, Inequalities, V3, P1