Whereas the previous two lessons described the biophysical and signalling properties of individual neurons, this lesson describes properties of those units when part of larger networks.
This lesson goes over some examples of how machine learners and computational neuroscientists go about designing and building neural network models inspired by biological brain systems.
This lecture and tutorial focuses on measuring human functional brain networks, as well as how to account for inherent variability within those networks.
This lecture presents an overview of functional brain parcellations, as well as a set of tutorials on bootstrap agregation of stable clusters (BASC) for fMRI brain parcellation.
Neuronify is an educational tool meant to create intuition for how neurons and neural networks behave. You can use it to combine neurons with different connections, just like the ones we have in our brain, and explore how changes on single cells lead to behavioral changes in important networks. Neuronify is based on an integrate-and-fire model of neurons. This is one of the simplest models of neurons that exist. It focuses on the spike timing of a neuron and ignores the details of the action potential dynamics. These neurons are modeled as simple RC circuits. When the membrane potential is above a certain threshold, a spike is generated and the voltage is reset to its resting potential. This spike then signals other neurons through its synapses.
Neuronify aims to provide a low entry point to simulation-based neuroscience.
This lesson provides an introduction the International Neuroinformatics Coordinating Facility (INCF), its mission towards FAIR neuroscience, and future directions.
This talk describes the NIH-funded SPARC Data Structure, and how this project navigates ontology development while keeping in mind the FAIR science principles.
This lesson consists of a brief discussion around this sessions previous talks.
This is the third and final lecture of this course on neuroinformatics infrastructure for handling sensitive data.
In this lecture, you will learn about virtual research environments (VREs) and their technical limitations, (i.e., a computing platform and the software stack behind it) and the security measures which should be considered during implementation.
This lesson consists of a panel discussion, wrapping up the INCF Neuroinformatics Assembly 2023 workshop Research Workflows for Collaborative Neuroscience.
This brief talk outlines the obstacles and opportunities involved in striving for more open and reproducible publishing, highlighting the need for investment in the technical and governance sectors of FAIR data and software.
This brief video provides a welcome and short introduction to the outline of the INCF Short Course in Neuroinformatics, held Seattle, Washington in October 2023, in coordination with the West Big Data Hub and the University of Washington.
This opening lecture from INCF's Short Course in Neuroinformatics provides an overview of the field of neuroinformatics itself, as well as laying out an argument for the necessity for developing more sophisticated approaches towards FAIR data management principles in neuroscience.
This lesson continues from part one of the lecture Ontologies, Databases, and Standards, diving deeper into a description of ontologies and knowledg graphs.
This lesson aims to define computational neuroscience in general terms, while providing specific examples of highly successful computational neuroscience projects.
This lecture covers a wide range of aspects regarding neuroinformatics and data governance, describing both their historical developments and current trajectories. Particular tools, platforms, and standards to make your research more FAIR are also discussed.
This lesson gives an in-depth description of scientific workflows, from study inception and planning to dissemination of results.
Introduction of the Foundations of Machine Learning in Python course - Day 01.
High-Performance Computing and Analytics Lab, University of Bonn
Optimization for machine learning - Day 02 lecture of the Foundations of Machine Learning in Python course.
High-Performance Computing and Analytics Lab, University of Bonn