Neural Information Processing: Course homepage 2016-2017

This is a course for MSc level students. It runs in semester 2.

There has a recent surge of computational architectures inspired by how the brain works, and unlike previously, they actually perform well on certain tasks. At the same time advanced computational analysis is increasingly used to analyse and understand neural data. This course covers both these developments. After describing rigorous ways to describe neural activity mathematically, and introducing methods how high-dimensional neural activity patterns can be represented and modelled, we present a number of the architectures recently used to do tasks like image understanding, memory and cognition, as well as some brain inspired hardware implementations.

The background needed to successfully take this course is a good grounding in mathematics, particularly with regard to probability and statistics, vectors and matrices. The mathematical level required is similar to that which would be obtained by students who did not have significant difficulties with the courses Mathematics for Informatics 1-4 taken in the first two years of the Informatics undergraduate syllabus. The Neural Computation (NC) course is a helpful but not necessary prerequisite, as biological realism is not such an important objective as in the NC course. Machine learning courses (LfD, PMR, IAML) will be also useful preparations.

Instructors: Mark van Rossum and Matthias Hennig.
Course Tutor: TBA


Monday and Thursday 9.00-9.50. David Hume LG.11
No lecture on Monday Jan 16th. The first lecture is on Thursday Jan 19th.

Course Outline

Class papers

Click this link. This list might be updated.


There will be two assessed assignments worth in total 25%. There will be an exam worth 75%.

First assignment . Deadline March 25th, 2017.

Second assignment, Deadline April 4th, 2017

Old assignments for practice: A1:12-13, A1:11-12, A1:10-11, A1:09-10, A1:08-09


See also references in the lecture notes.
Theoretical Neuroscience by P Dayan and L F Abbott (MIT Press 2001) is recommended reading, see also the list of errata.
Neuronal Dynamics by Wulfram Gerstner, Werner M. Kistler, Richard Naud and Liam Paninski. Full version online.
Natural Image Statistics by Aapo Hyvarinen, Jarmo Hurri, and Patrik O. Hoyer. Full version online.
Information Theory, Inference and Learning Algorithms by David MacKay \item Introduction To The Theory Of Neural Computation, Volume I, by John Hertz.

Week-by-week listing (approximate)

Week 1

Lectures: Introduction: lecture slides (2x2) , lecture slides (single page)

Week 2

Neural encoding: lecture slides (2x2) , lecture slides (single page)
Neural decoding: lecture slides, lecture slides (single page)

Week 3

Lectures: Information theory: lecture slides, lecture slides (single page)

Week 4

Retina coding: lecture slides, lecture slides (single page)

Week 5

Higher order statistics : lecture slides, lecture slides (single page)
Networks for visual processing: lecture slides, lecture slides (single page)

Week 6

Break week. No lectures.

Week 5

Maximum entropy models: lecture slides, mean constraint derivation.

Week 7

Maximum entropy models, contd.: lecture slides

Week 8

Hopfield model and Boltzmann machines: lecture slides

Week 9

Lectures Restricted Boltzmann machines: lecture slides

Week 10

Week 11

Class papers

Week 12


Solutions to homework, MvR's part

This page is maintained by Mark van Rossum .

Home : Teaching : Courses 

Informatics Forum, 10 Crichton Street, Edinburgh, EH8 9AB, Scotland, UK
Tel: +44 131 651 5661, Fax: +44 131 651 1426, E-mail:
Please contact our webadmin with any comments or corrections. Logging and Cookies
Unless explicitly stated otherwise, all material is copyright © The University of Edinburgh