Navigation auf uzh.ch
Type: | Lecture with Exercises |
ECTS: | 6 points |
Lectures: | Fridays, 08:30-12:00 |
Venue: | BIN 2.A.01 |
Lecturers: | Prof. Rolf Pfeifer |
Target audience: | Recommended for MSc students. The course is interdisciplinary; it is also targeted at students from other fields than computer science, e.g. economics, biology, natural sciences, and psychology. |
Language: | English |
Assessment: | Written exam. Date: Friday, 14.06.2013, 8:00 - 9:30am |
Assistants: | Nico Schmidt, Matthias Weyland |
Systematic introduction to neural networks, biological foundations; important network classes and learning algorithms; supervised models (perceptrons, adalines, multi-layer perceptrons), support-vector machines, echo-state networks, non-supervised networks (competitive, Kohonen, Hebb), recurrent networks (Hopfield, CTRNNs - continuous-time recurrent neural networks), spiking neural networks, spike-time dependent plasticity, applications. Special consideration will be given to neural networks embedded in adaptive systems having to interact with the real world, such as embodied systems (in particular robots). Cooperation of neural control, morphology, materials, and environment. Evolutionary approaches to designing autonomous systems; interaction of learning and evolution. Network theory applied to brain networks; motifs.
Additional case studies will be discussed to deepen the understanding of neural networks, e.g. Neural interfacing - coupling neural systems with technology (in particular robotic devices), neural imaging studies, Distributed Adaptive Control (DAC), neural gas and DRNNs - Dynamically Rearranging Neural Networks (neuro-modulator-based networks), neural network models of memory.
This is an elementary, interdisciplinary introduction to neural networks, suited not only for computer scientists, but also for economists, biologists, psychologists, etc.
If you wish, you can do the exercises in groups of two - please hand in only one task sheet. You can also do them on your own if you prefer.
The page will be subjected to changes over time. Please stay up-to-date by checking it periodically.
List of topics relevant for exam (PDF, 69 KB)
List of topics relevant for the final exam.
The final exam will be Friday, 14 June, 2013, from 8.00 to 9.30am in the lecture room 2.A.01.
In order to attend the final exam, you must achieve 50% out of the total possible points of all the exercises (Task sheets 1-4) together (not each).
It will not be an open-book exam, so you are not allowed to use your books and notes. However, you don't have to learn any formulas by heart, as we will provide you with a sheet containing all formulas (but you do have to know which one applies to which type of Neural Net). No laptops or cell phones.
Please bring along:
We wish you a lot of success.
Date | Topic | Lecture | Exercises |
---|---|---|---|
22 February | Introduction, Linear Algebra | For an intuitive, 50min introduction to artificial neural networks, please consult this video which was recorded in the context of the ShanghAI Lectures. All the points raised in this video will be taken up again and discussed in more detail later in the class. Please also consult the pdf of the slides for this lecture. Neural networks require relatively little prior knowledge in mathematics, just some linear algebra and a bit of elementary calculus. For those who are not confident about their linear algebra skills, we will provide an introductory tutorial - including a set of exercises during the first lecture on 22 February, starting approx at 10.00h. If you are confident that you already master basic linear algebra, you don't need to attend this tutorial. |
|
1 march | Supervised models | Perceptron, Adaline, delta-rule | |
8 March | No lecture | Special event: Robots on Tour www.robotsontour.com | |
15 March | Supervised models | Back-propagation: examples, properties; Error surfaces, Momentum term, Other improvements; N-fold cross-validation, VC dimension | (Due date: 12 April) |
22 March | Supervised models | Cascade correlation, Suport vector machines (SVMs) | |
29 March | No lecture | ||
5 April | No lecture | ||
12 April | Supervised models | Cascade correlation, Suport vector machines (SVMs) | hand in Task Sheet 1 (Due date: 26 April) |
19 April | Recurrent neural networks | Hopfield nets, Stocastic Models, CTRNNs (Continuous Time Recurrent Neural Network) | |
26 April | Hybrid models | Guest lecture: Naveen Kuppuswamy: Reservoir computing | hand in Task Sheet 2; Videos of Lectures by Andrew Ng on SVMs: 1, 2, 3 |
3 May | Unsupervised models | Nico Schmidt on Hebbian Learning, PCA, Oja's rule, Sanger's rule | |
10 May | Biologically more plausible models | Guest lecture: Pascal Kaufmann: Basic neurophysiology, Spiking neurons, Cyborgs, Lamprey experiment, Brain imaging |
Simulator (ZIP, 315 KB) |
17 May | Unsupervised models | Competitive learning, SOM, Kohonen-algorithm, Extended Kohonen map (robot arm), Adaptive light compass. | it is OK to hand in Task Sheet 3 on 17 May! |
24 May | Application of recurrent networks | Morphological Computation, Evolutionary Robotics, Co-evolution of morphology and control | hand in Task Sheet 4 |
31 May | Wrap-up | Wrap-up session, questions, final discussion | |
14 Jun | Exam |
Some links on the internet that are useful for the understanding of the course