This lecture describes how to build research workflows, including a demonstrate using DataJoint Elements to build data pipelines.
This lesson gives a primer to project management in a scientific context, with a particular neuroinformatic case study.
In this lesson, you will hear about the current challenges regarding data management, as well as policies and resources aimed to address them.
This lesson provides an overview of how to manage relationships in a research context, while highlighting the need for effective communication at various levels.
This lecture covers different perspectives on the study of the mental, focusing on the difference between Mind and Brain.
This lesson briefly goes over the outline of the Neuroscience for Machine Learners course.
This lecture focuses on where and how Jupyter notebooks can be used most effectively for education.
JupyterHub is a simple, highly extensible, multi-user system for managing per-user Jupyter Notebook servers, designed for research groups or classes. This lecture covers deploying JupyterHub on a single server, as well as deploying with Docker using GitHub for authentication.
This lesson introduces the practical usage of The Virtual Brain (TVB) in its graphical user interface and via python scripts. In the graphical user interface, you are guided through its data repository, simulator, phase plane exploration tool, connectivity editor, stimulus generator, and the provided analyses. The implemented iPython notebooks of TVB are presented, and since they are public, can be used for further exploration of TVB.
This lesson provides a brief overview of the Python programming language, with an emphasis on tools relevant to data scientists.
This tutorial covers the fundamentals of collaborating with Git and GitHub.
This lesson is a general overview of overarching concepts in neuroinformatics research, with a particular focus on clinical approaches to defining, measuring, studying, diagnosing, and treating various brain disorders. Also described are the complex, multi-level nature of brain disorders and the data associated with them, from genes and individual cells up to cortical microcircuits and whole-brain network dynamics. Given the heterogeneity of brain disorders and their underlying mechanisms, this lesson lays out a case for multiscale neuroscience data integration.
This lesson breaks down the principles of Bayesian inference and how it relates to cognitive processes and functions like learning and perception. It is then explained how cognitive models can be built using Bayesian statistics in order to investigate how our brains interface with their environment.
This lesson corresponds to slides 1-64 in the PDF below.
Whereas the previous two lessons described the biophysical and signalling properties of individual neurons, this lesson describes properties of those units when part of larger networks.
This lesson goes over some examples of how machine learners and computational neuroscientists go about designing and building neural network models inspired by biological brain systems.
Neuronify is an educational tool meant to create intuition for how neurons and neural networks behave. You can use it to combine neurons with different connections, just like the ones we have in our brain, and explore how changes on single cells lead to behavioral changes in important networks. Neuronify is based on an integrate-and-fire model of neurons. This is one of the simplest models of neurons that exist. It focuses on the spike timing of a neuron and ignores the details of the action potential dynamics. These neurons are modeled as simple RC circuits. When the membrane potential is above a certain threshold, a spike is generated and the voltage is reset to its resting potential. This spike then signals other neurons through its synapses.
Neuronify aims to provide a low entry point to simulation-based neuroscience.
This lesson gives an in-depth introduction of ethics in the field of artificial intelligence, particularly in the context of its impact on humans and public interest. As the healthcare sector becomes increasingly affected by the implementation of ever stronger AI algorithms, this lecture covers key interests which must be protected going forward, including privacy, consent, human autonomy, inclusiveness, and equity.
This lesson describes a definitional framework for fairness and health equity in the age of the algorithm. While acknowledging the impressive capability of machine learning to positively affect health equity, this talk outlines potential (and actual) pitfalls which come with such powerful tools, ultimately making the case for collaborative, interdisciplinary, and transparent science as a way to operationalize fairness in health equity.
Introduction of the Foundations of Machine Learning in Python course - Day 01.
High-Performance Computing and Analytics Lab, University of Bonn
This lesson gives a brief introduction to the course Neuroscience for Machine Learners (Neuro4ML).