This lecture presents the Medical Informatics Platform's data federation in epilepsy.
This lecture aims to help researchers, students, and health care professionals understand the place for neuroinformatics in the patient journey using the exemplar of an epilepsy patient.
In this lesson, the simulation of a virtual epileptic patient is presented as an example of advanced brain simulation as a translational approach to deliver improved clinical results. You will learn about the fundamentals of epilepsy, as well as the concepts underlying epilepsy simulation. By using an iPython notebook, the detailed process of this approach is explained step by step. In the end, you are able to perform simple epilepsy simulations your own.
This talk introduces data sharing initiatives in Epilepsy, particularly across Europe.
In this lesson, you will learn in more detail about neuromorphic computing, that is, non-standard computational architectures that mimic some aspect of the way the brain works.
This video provides a very quick introduction to some of the neuromorphic sensing devices, and how they offer unique, low-power applications.
This opening lecture from INCF's Short Course in Neuroinformatics provides an overview of the field of neuroinformatics itself, as well as laying out an argument for the necessity for developing more sophisticated approaches towards FAIR data management principles in neuroscience.
This lesson continues from part one of the lecture Ontologies, Databases, and Standards, diving deeper into a description of ontologies and knowledg graphs.
This lesson aims to define computational neuroscience in general terms, while providing specific examples of highly successful computational neuroscience projects.
This lecture covers a wide range of aspects regarding neuroinformatics and data governance, describing both their historical developments and current trajectories. Particular tools, platforms, and standards to make your research more FAIR are also discussed.
Presented by the OHBM OpenScienceSIG, this lesson covers how containers can be useful for running the same software on different platforms and sharing analysis pipelines with other researchers.
Serving as good refresher, this lesson explains the maths and logic concepts that are important for programmers to understand, including sets, propositional logic, conditional statements, and more.
This compilation is courtesy of freeCodeCamp.
This lesson provides a useful refresher which will facilitate the use of Matlab, Octave, and various matrix-manipulation and machine-learning software.
This lesson was created by RootMath.
In this lecture, you will learn about current methods, approaches, and challenges to studying human neuroanatomy, particularly through the lense of neuroimaging data such as fMRI and diffusion tensor imaging (DTI).
This module covers some basic anatomy such as the brain’s major divisions (brainstem, cerebellum, cerebrum), the cerebral lobes (frontal, temporal, parietal, and occipital), the central and peripheral nervous systems, theories of cognition, and brain orientation terms.
This lecture provides an introduction to the application of genetic testing in neurodevelopmental disorders.
This lesson describes the fundamentals of genomics, from central dogma to design and implementation of GWAS, to the computation, analysis, and interpretation of polygenic risk scores.
This lesson is an overview of transcriptomics, from fundamental concepts of the central dogma and RNA sequencing at the single-cell level, to how genetic expression underlies diversity in cell phenotypes.
This lesson contains both a lecture and a tutorial component. The lecture (0:00-20:03 of YouTube video) discusses both the need for intersectional approaches in healthcare as well as the impact of neglecting intersectionality in patient populations. The lecture is followed by a practical tutorial in both Python and R on how to assess intersectional bias in datasets. Links to relevant code and data are found below.
This lecture provides an overview of successful open-access projects aimed at describing complex neuroscientific models, and makes a case for expanded use of resources in support of reproducibility and validation of models against experimental data.