Filters








13,779 Hits in 2.9 sec

Some Classes of Regular Languages Identifiable in the Limit from Positive Data [chapter]

François Denis, Aurélien Lemay, Alain Terlutte
2002 Lecture Notes in Computer Science  
Inference with Multinomial Tests p. 149 Learning Languages with Help p. 161 Incremental Learning of Context Free Grammars p. 174 Estimating Grammar Parameters Using Bounded Memory p. 185 Stochastic  ...  Languages p. 283 Software Descriptions The EMILE 4.1 Grammar Induction Toolbox p. 293 Software for Analysing Recurrent Neural Nets That Learn to Predict Non-regular Languages p. 296 A Framework  ... 
doi:10.1007/3-540-45790-9_6 fatcat:nmlknwqoyfbybhb6rpomqrn7qy

Page 1778 of Mathematical Reviews Vol. 41, Issue 6 [page]

1971 Mathematical Reviews  
A fuzzy automaton behaves in a deterministic fashion. However, it has many properties similar to those of stochastic automata. Its application as a model of learning systems is discussed.  ...  T. 9670 Linear regular languages. I. Acta Cybernet’ 1 (1969), 3-12.  ... 

Ten Open Problems in Grammatical Inference [chapter]

Colin de la Higuera
2006 Lecture Notes in Computer Science  
They cover the areas of polynomial learning models, learning from ordered alphabets, learning deterministic Pomdps, learning negotiation processes, learning from context-free background knowledge.  ...  Work with Henning Fernau on polynomial learning is where the ideas in section 3 come from. Philippe Jaillon gave me the initial ideas for the negotiation problem in section 9. Discussions with  ...  Testing equivalence of regular deterministic distributions Learning stochastic languages is an important topic in grammatical inference.  ... 
doi:10.1007/11872436_4 fatcat:5snm4lpumbhw5e7ffh5vwxzhum

A bibliographical study of grammatical inference

Colin de la Higuera
2005 Pattern Recognition  
The field of grammatical inference (also known as grammar induction) is transversal to a number of research areas including machine learning, formal language theory, syntactic and structural pattern recognition  ...  Therefore, learning a stochastic automaton involves a modification of bias from what has been presented before: even if the underlying language is regular, the distribution may not be.  ...  The theory The main focus of research in the field of grammatical inference has been set on learning regular grammars or deterministic finite automata (DFA).  ... 
doi:10.1016/j.patcog.2005.01.003 fatcat:62qwskiqcvddjobakbdshwebqq

Page 1554 of Mathematical Reviews Vol. 57, Issue 4 [page]

1979 Mathematical Reviews  
S. 57 + 11976 Learning with stochastic automata and stochastic languages. With discussion. Computer oriented learning processes (Proc. NATO Advanced Study Inst., Bonas, 1974), pp. 69-107.  ...  This paper contains a survey of learning algorithms for stochastic automata and inference procedures for stochastic grammars.  ... 

Learning Deterministic Regular Expressions for the Inference of Schemas from XML Data

Geert Jan Bex, Wouter Gelade, Frank Neven, Stijn Vansummeren
2010 ACM Transactions on the Web  
Unfortunately, there is no algorithm capable of learning the complete class of deterministic regular expressions from positive examples only, as we will show.  ...  Inferring an appropriate DTD or XML Schema Definition (XSD) for a given collection of XML documents essentially reduces to learning deterministic regular expressions from sets of positive example words  ...  Since deterministic regular expressions like a * define infinite languages, and since every non-empty finite language can be defined by a deterministic expression (as we show in the full version of this  ... 
doi:10.1145/1841909.1841911 fatcat:hre7agfyuzhudl3xwlaxxd4xg4

Learning deterministic regular expressions for the inference of schemas from XML data

Geert Jan Bex, Wouter Gelade, Frank Neven, Stijn Vansummeren
2008 Proceeding of the 17th international conference on World Wide Web - WWW '08  
Unfortunately, there is no algorithm capable of learning the complete class of deterministic regular expressions from positive examples only, as we will show.  ...  Inferring an appropriate DTD or XML Schema Definition (XSD) for a given collection of XML documents essentially reduces to learning deterministic regular expressions from sets of positive example words  ...  Since deterministic regular expressions like a * define infinite languages, and since every non-empty finite language can be defined by a deterministic expression (as we show in the full version of this  ... 
doi:10.1145/1367497.1367609 dblp:conf/www/BexGNV08 fatcat:bqs6npqyi5eaxlzv5n3ktgb6di

Stochastic

Hareesh Bahuleyan, Lili Mou, Hao Zhou, Olga Vechtomova
2019 Proceedings of the 2019 Conference of the North  
In this paper, we propose to use the Wasserstein autoencoder (WAE) for probabilistic sentence generation, where the encoder could be either stochastic or deterministic.  ...  the stochasticity of the encoder.  ...  D and S refer to the deterministic and stochastic encoders, respectively. ↑/↓ The larger/lower, the better.  ... 
doi:10.18653/v1/n19-1411 dblp:conf/naacl/BahuleyanMZV19 fatcat:y5cetx4ct5hdjcr6x4ibeequ4y

Mungojerrie: Reinforcement Learning of Linear-Time Objectives [article]

Ernst Moritz Hahn, Mateo Perez, Sven Schewe, Fabio Somenzi, Ashutosh Trivedi, Dominik Wojtczak
2021 arXiv   pre-print
Mungojerrie (https://plv.colorado.edu/mungojerrie/) is a tool for testing reward schemes for ω-regular objectives on finite models.  ...  An alternative to this manual programming, akin to programming directly in assembly, is to specify the objective in a formal language and have it "compiled" to a reward scheme.  ...  A natural choice for this language is Linear Temporal Logic (LTL) [22, 27] , or more generally, ω-regular languages [26] . ω-regular languages describe infinite sequences.  ... 
arXiv:2106.09161v2 fatcat:k7jvqed2wzebfbthfg2zwhxzo4

Sparseout: Controlling Sparsity in Deep Networks [chapter]

Najeeb Khan, Ian Stavness
2019 Lecture Notes in Computer Science  
Sparsity is a potentially important property of neural networks, but is not explicitly controlled by Dropout-based regularization.  ...  Sparseout provides a way to investigate sparsity in state-of-the-art deep learning models. Source code for Sparseout could be found at .  ...  Stochastic regularization has become the standard practice in training deep learning models and have outperformed deterministic regularization methods on many tasks.  ... 
doi:10.1007/978-3-030-18305-9_24 fatcat:6kc7pdzm2zhxnc5mxutdiogu3y

Bayesian Layers: A Module for Neural Network Uncertainty [article]

Dustin Tran and Michael W. Dusenberry and Mark van der Wilk and Danijar Hafner
2019 arXiv   pre-print
This enables composition via a unified abstraction over deterministic and stochastic functions and allows for scalability via the underlying system.  ...  Finally, we show how Bayesian Layers can be used within the Edward2 probabilistic programming language for probabilistic programs with stochastic processes.  ...  ., 2017 ) ( Figure 6 ); or an auto-encoder with stochastic encoders and decoders (Figure 7) . 3 Signature To implement stochastic output layers, we perform deterministic computations given a tensordimensional  ... 
arXiv:1812.03973v3 fatcat:oxsckegvezcfljz25nlz4cfn54

Learning deterministic regular grammars from stochastic samples in polynomial time

Rafael C. Carrasco, Jose Oncina
1999 RAIRO - Theoretical Informatics and Applications  
In this paper, the identification of stochastic regular languages is addressed.  ...  For this purpose, we propose a class of algorithms which allow for the identification of the structure of the minimal stochastic automaton generating the language.  ...  Every stochastic deterministic regular grammar G defines a stochastic deterministic regular language (SDRL), L G , through the probabilities p(w|L G ) = p(S ⇒ w).  ... 
doi:10.1051/ita:1999102 fatcat:rw2vcb2qtnfo7ma5cns3dffuum

Learning Probabilistic Residual Finite State Automata [chapter]

Yann Esposito, Aurélien Lemay, François Denis, Pierre Dupont
2002 Lecture Notes in Computer Science  
We prove that there are more languages generated by PRFA than by Probabilistic Deterministic Finite Automata (PDFA).  ...  We show that this class can be characterized by a simple intrinsic property of the stochastic languages they generate (the set of residual languages is finitely generated) and that it admits canonical  ...  It consists of all stochastic languages generated by probabilistic finite automata. Also, the class of stochastic deterministic regular languages on Σ is denoted by L P DF A (Σ).  ... 
doi:10.1007/3-540-45790-9_7 fatcat:go6ap7enwjhnflnwve4ayalj74

Probabilistic finite-state machines - part II

E. Vidal, F. Thollard, C. de la Higuera, F. Casacuberta, R.C. Carrasco
2005 IEEE Transactions on Pattern Analysis and Machine Intelligence  
This is a direct consequence of the fact that every regular language is the support of at least one stochastic regular language, and there are regular languages which are not k-testable.  ...  APPENDIX A.1 Proof of Theorem 3Theorem 3 (Stochastic morphism theorem). Let AE be a finite alphabet and D be a stochastic regular language on AE ? .  ... 
doi:10.1109/tpami.2005.148 pmid:16013757 fatcat:vaoopt4ypzffzpv53pxx2hodpy

Learning Languages with Help [chapter]

Christopher Kermorvant, Colin de la Higuera
2002 Lecture Notes in Computer Science  
We propose a general setting to deal with these cases and provide algorithms that can learn deterministic finite automata in these conditions.  ...  Grammatical inference consists in learning formal grammars for unknown languages when given learning data. Classically this data is raw: strings that belong to the language or that do not.  ...  An interesting alternative is to consider the hypothesis that not only is the language regular, but that the distribution also is. In such a case one needs to learn a Stochastic Finite Automaton.  ... 
doi:10.1007/3-540-45790-9_13 fatcat:adbjrahojfejxieqmglom6vnay
« Previous Showing results 1 — 15 out of 13,779 results