Online learning in Bayesian spiking neurons

Levin Kuhlmann, Michael Hauser-Raspe, Jonathan Manton, David B. Grayden, Jonathan Tapson, André van Schaik

    Research output: Chapter in Book / Conference PaperConference Paperpeer-review

    1 Citation (Scopus)

    Abstract

    Bayesian Spiking Neurons (BSNs) provide a probabilistic interpretation of how neurons can perform inference and learning. Learning in a single BSN can be formulated as an online maximum-likelihood expectation-maximisation (ML-EM) algorithm. This form of learning is quite slow. Here, an alternative to this learning algorithm, called Fast Learning (FL), is presented. The FL algorithm is shown to have acceptable convergence performance when compared to the ML-EM algorithm. Moreover, for our implementations the FL algorithm is approximately 25 times faster than the ML-EM algorithm. Although only approximate, the FL algorithm therefore makes learning in hierarchical BSN networks much more tractable.
    Original languageEnglish
    Title of host publicationProceedings of the 2012 International Joint Conference on Neural Networks (IJCNN): 10-15 June 2012, Brisbane, Qld.
    PublisherIEEE
    Pages1-6
    Number of pages6
    ISBN (Print)9781467314909
    DOIs
    Publication statusPublished - 2012
    EventInternational Joint Conference on Neural Networks -
    Duration: 10 Jun 2012 → …

    Conference

    ConferenceInternational Joint Conference on Neural Networks
    Period10/06/12 → …

    Keywords

    • Bayesian spiking neurons
    • algorithms
    • learning
    • mathematical models
    • neurons
    • online learning

    Fingerprint

    Dive into the research topics of 'Online learning in Bayesian spiking neurons'. Together they form a unique fingerprint.

    Cite this