Research Output per year

### Abstract

Bayesian algorithms pose a limit to the performance learning algorithms can achieve. Natural selection should guide the evolution of information processing systems towards those limits. What can we learn from this evolution and what properties do the intermediate stages have? While this question is too general to permit any answer, progress can be made by restricting the class of information processing systems under study. We present analytical and numerical results for the evolution of on-line algorithms for learning from examples for neural network classifiers, which might include or not a hidden layer. The analytical results are obtained by solving a variational problem to determine the learning algorithm that leads to maximum generalization ability. Simulations using evolutionary programming, for programs that implement learning algorithms, confirm and expand the results. The principal result is not just that the evolution is towards a Bayesian limit. Indeed it is essentially reached. In addition we find that evolution is driven by the discovery of useful structures or combinations of variables and operators. In different runs the temporal order of the discovery of such combinations is unique. The main result is that combinations that signal the surprise brought by an example arise always before combinations that serve to gauge the performance of the learning algorithm. This latter structures can be used to implement annealing schedules. The temporal ordering can be understood analytically as well by doing the functional optimization in restricted functional spaces. We also show that there is data suggesting that the appearance of these traits also follows the same temporal ordering in biological systems.

Original language | English |
---|---|

Title of host publication | Bayesian inference and maximum entropy methods in science and engineering |

Editors | Ali Mohammad-Djafari |

Publisher | AIP |

Pages | 203-210 |

Number of pages | 8 |

ISBN (Print) | 978-0-7354-0371-6 |

DOIs | |

Publication status | Published - 29 Dec 2006 |

Event | Bayesian inference and maximum entropy methods In science and engineering - Paris, France Duration: 8 Jul 2006 → 13 Jul 2006 |

### Publication series

Name | AIP conference proceedings |
---|---|

Publisher | AIP |

Volume | 872 |

ISSN (Print) | 0094-243X |

ISSN (Electronic) | 1551-7616 |

### Conference

Conference | Bayesian inference and maximum entropy methods In science and engineering |
---|---|

Country | France |

City | Paris |

Period | 8/07/06 → 13/07/06 |

## Fingerprint Dive into the research topics of 'The evolution of learning systems: to Bayes or not to be'. Together they form a unique fingerprint.

## Research Output

- 1 Conference contribution

## Online learning in discrete hidden Markov models

Alamino, R. C. & Caticha, N., 29 Dec 2006,*Bayesian inference and maximum entropy methods In science and engineering.*Mohammad-Djafari, A. (ed.). AIP, p. 187-194 8 p. (AIP conference proceedings; vol. 872).

Research output: Chapter in Book/Report/Conference proceeding › Conference contribution

## Cite this

*Bayesian inference and maximum entropy methods in science and engineering*(pp. 203-210). (AIP conference proceedings; Vol. 872). AIP. https://doi.org/10.1063/1.2423276