This is the first of two workshops on reproducibility in science, during which participants are introduced to concepts of FAIR and open science. After discussing the definition of and need for FAIR science, participants are walked through tutorials on installing and using Github and Docker, the powerful, open-source tools for versioning and publishing code and software, respectively.
This lesson breaks down the principles of Bayesian inference and how it relates to cognitive processes and functions like learning and perception. It is then explained how cognitive models can be built using Bayesian statistics in order to investigate how our brains interface with their environment.
This lesson corresponds to slides 1-64 in the PDF below.
Optimization for machine learning - Day 02 lecture of the Foundations of Machine Learning in Python course.
High-Performance Computing and Analytics Lab, University of Bonn
Linear Algebra for Machine Learning - Day 03 lecture of the Foundations of Machine Learning in Python course.
High-Performance Computing and Analytics Lab, University of Bonn
Support Vector Machines - Day 06 lecture of the Foundations of Machine Learning in Python course.
High-Performance Computing and Analytics Lab, University of Bonn
Decision Trees and Random Forests - Day 07 lecture of the Foundations of Machine Learning in Python course.
High-Performance Computing and Analytics Lab, University of Bonn
Clustering and Density Estimation - Day 08 lecture of the Foundations of Machine Learning in Python course.
High-Performance Computing and Analytics Lab, University of Bonn
Dimensionality Reduction - Day 09 lecture of the Foundations of Machine Learning in Python course.
High-Performance Computing and Analytics Lab, University of Bonn
Introduction to Neural Networks - Day 10 lecture of the Foundations of Machine Learning in Python course.
High-Performance Computing and Analytics Lab, University of Bonn
Introduction to Convolutional Neural Networks - Day 11 lecture of the Foundations of Machine Learning in Python course.
High-Performance Computing and Analytics Lab, University of Bonn
Initialization, Optimization, and Regularization - Day 12 lecture of the Foundations of Machine Learning in Python course.
High-Performance Computing and Analytics Lab, University of Bonn
U-Nets for medical Image-Segmentation - Day 13 lecture of the Foundations of Machine Learning in Python course.
High-Performance Computing and Analytics Lab, University of Bonn
Sequence Processing - Day 15 lecture of the Foundations of Machine Learning in Python course.
High-Performance Computing and Analytics Lab, University of Bonn
This lesson briefly goes over the outline of the Neuroscience for Machine Learners course.
This lesson delves into the the structure of one of the brain's most elemental computational units, the neuron, and how said structure influences computational neural network models.
In this lesson you will learn how machine learners and neuroscientists construct abstract computational models based on various neurophysiological signalling properties.
This lesson goes over the basic mechanisms of neural synapses, the space between neurons where signals may be transmitted.
While the previous lesson in the Neuro4ML course dealt with the mechanisms involved in individual synapses, this lesson discusses how synapses and their neurons' firing patterns may change over time.
This lesson describes spike timing-dependent plasticity (STDP), a biological process that adjusts the strength of connections between neurons in the brain, and how one can implement or mimic this process in a computational model. You will also find links for practical exercises at the bottom of this page.
In this lesson, you will learn about some of the many methods to train spiking neural networks (SNNs) with either no attempt to use gradients, or only use gradients in a limited or constrained way.