Structure and its impact for recognition

Publications

Authors Title
S. Basovník, F. Mráz Learning Limited Context Restarting Automata by Genetic Algorithms (Technical Report)
Technical report, 2011, Charles University, Faculty of Mathematics and Physics, Prague
Abstract: We propose a genetic algorithm for learning restricted variants of restarting automata from positive and negative samples. Experiments comparing the proposed genetic algorithm to algorithms RPNI and LARS on sample languages indicate that the new algorithm is able to infer a target language even from a small set of samples.
Links: BibTeX, Fulltext
P. Černo, F. Mráz Delta-Clearing Restarting Automata and CFL
Proceedings of the DLT 2011 15th International Conference on Developments in Language Theory (Milano, Italy), Springer, Berlin, 2011, LNCS, Vol. 6795, 153-164.
Abstract: Delta-clearing restarting automata represent a new restricted model of restarting automata which, based on a limited context, can either delete a substring of the current content of its tape or replace a substring by a special auxiliary symbol Delta, which cannot be overwritten anymore, but it can be deleted later. The main result of this paper consists in proving that besides their limited operations, Delta-clearing restarting automata recognize all context-free languages.
Links: BibTeX, Presentation, SpringerLink
P. Černo, F. Mráz Delta-Clearing Restarting Automata and CFL (Technical Report)
Technical report, 2011, Charles University, Faculty of Mathematics and Physics, Prague
Abstract: See above
Links: BibTeX, Fulltext
I. Mrázová, Z. Reitermanová A New Sensitivity-Based Feature Selection Technique for Feed-Forward Neural Networks That Improves Generalization (Technical Report)
Technical report, 2011, Charles University, Faculty of Mathematics and Physics, Prague
Abstract: Multi-layer neural networks of the back-propagation type became already a well-established tool used successfully in various application areas. Efficient solutions to complex tasks currently dealt with obviously require sufficient generalization capabilities of the formed networks and an easy interpretation of their function. For this reason, we will introduce here a new feature selection technique called SCGSIR inspired by the fast method of scaled conjugate gradients (SCG) and sensitivity analysis.
Enforced internal knowledge representation supports an easy interpretation of the formed network structure. Network sensitivity inhibited during training impacts successful pruning of input neurons and optimization of network structure, too. Experiments performed so far on the problem of binary addition and on real data obtained from the World Bank yield promising results outperforming reference techniques when considering both their ability to find networks with and optimum architectures and generalization capabilities of the trained networks.
Links: BibTeX, Fulltext
D. Pardubská, M. Plátek, F. Otto Parallel communicating grammar systems with regular control and skeleton preserving FRR-automata
Theoretical Computer Science 412 (2011) 458–477.
Abstract: Parallel communicating grammar systems with regular control (RPCGS, for short) are introduced, which are obtained from returning regular parallel communicating grammar systems by restricting the derivations that are executed in parallel by the various components through a regular control language. For the class of languages that are generated by RPCGSs with constant communication complexity we derive a characterisation in terms of a restricted type of freely rewriting restarting automaton. From this characterisation we obtain that these languages are semi-linear, and that for RPCGSs with constant communication complexity, the centralised variant has the same generative power as the non-centralised variant.
Links: BibTeX, ScienceDirect

Last updated: Sunday, 03/09/14