This tutorial walks users through the creation and visualization of activation flat maps from fMRI datasets.
This tutorial demonstrates to users the conventional preprocessing steps when working with BOLD signal datasets from fMRI.
In this tutorial, users will learn how to create a trial-averaged BOLD response and store it in a matrix in MATLAB.
This tutorial teaches users how to create animations of BOLD responses over time, to allow researchers and clinicians to visualize time-course activity patterns.
This tutorial demonstrates how to use MATLAB to create event-related BOLD time courses from fMRI datasets.
In this tutorial, users learn how to compute and visualize a t-test on experimental condition differences.
You will learn about working with calcium imaging data, including image processing to remove background "blur," identifying cells based on thresholded spatial contiguity, time series filtering, and principal components analysis (PCA). The MATLAB code shows data animations, capabilities of the image processing toolbox, and PCA.
You will learn about working with calcium imaging data, including image processing to remove background "blur," identifying cells based on thresholded spatial contiguity, time series filtering, and principal components analysis (PCA). The MATLAB code shows data animations, capabilities of the image processing toolbox, and PCA.
You will learn about working with calcium imaging data, including image processing to remove background "blur," identifying cells based on thresholded spatial contiguity, time series filtering, and principal components analysis (PCA). The MATLAB code shows data animations, capabilities of the image processing toolbox, and PCA.
You will learn about working with calcium imaging data, including image processing to remove background "blur," identifying cells based on thresholded spatial contiguity, time series filtering, and principal components analysis (PCA). The MATLAB code shows data animations, capabilities of the image processing toolbox, and PCA.
You will learn about working with calcium imaging data, including image processing to remove background "blur," identifying cells based on thresholded spatial contiguity, time series filtering, and principal components analysis (PCA). The MATLAB code shows data animations, capabilities of the image processing toolbox, and PCA.
You will learn about working with calcium imaging data, including image processing to remove background "blur," identifying cells based on thresholded spatial contiguity, time series filtering, and principal components analysis (PCA). The MATLAB code shows data animations, capabilities of the image processing toolbox, and PCA.
You will learn about working with calcium imaging data, including image processing to remove background "blur," identifying cells based on thresholded spatial contiguity, time series filtering, and principal components analysis (PCA). The MATLAB code shows data animations, capabilities of the image processing toolbox, and PCA.
This is a tutorial introducing participants to the basics of RNA-sequencing data and how to analyze its features using Seurat.
This tutorial was part of the 2018 Neurohackademy, a 2-week hands-on summer institute in neuroimaging and data science held at the University of Washington eScience Institute.
This lecture on multi-scale entropy by Jil Meier is part of the TVB Node 10 series, a 4 day workshop dedicated to learning about The Virtual Brain, brain imaging, brain simulation, personalised brain models, TVB use cases, etc. TVB is a full brain simulation platform.
This lecture on modeling epilepsy using TVB by Julie Courtiol is part of the TVB Node 10 series, a 4 day workshop dedicated to learning about The Virtual Brain, brain imaging, brain simulation, personalised brain models, TVB use cases, etc. TVB is a full brain simulation platform.
DAQCORD is a framework for the design, documentation and reporting of data curation methods in order to advance the scientific rigour, reproducibility and analysis of the data. This lecture covers the rationale for developing the framework, the process in which the framework was developed, and ends with a presentation of the framework. While the driving use case for DAQCORD was clinical traumatic brain injury research, the framework is applicable to clinical studies in other domains of clinical neuroscience research.
In this session the Medical Informatics Platform (MIP) federated analytics is presented. The current and future analytical tools implemented in the MIP will be detailed along with the constructs, tools, processes, and restrictions that formulate the solution provided. MIP is a platform providing advanced federated analytics for diagnosis and research in clinical neuroscience research. It is targeting clinicians, clinical scientists and clinical data scientists. It is designed to help adopt advanced analytics, explore harmonized medical data of neuroimaging, neurophysiological and medical records as well as research cohort datasets, without transferring original clinical data. It can be perceived as a virtual database that seamlessly presents aggregated data from distributed sources, provides access and analyze imaging and clinical data, securely stored in hospitals, research archives and public databases. It leverages and re-uses decentralized patient data and research cohort datasets, without transferring original data. Integrated statistical analysis tools and machine learning algorithms are exposed over harmonized, federated medical data.
The Medical Informatics Platform (MIP) is a platform providing federated analytics for diagnosis and research in clinical neuroscience research. The federated analytics is possible thanks to a distributed engine that executes computations and transfers information between the members of the federation (hospital nodes). In this talk the speaker will describe the process of designing and implementing new analytical tools, i.e. statistical and machine learning algorithms. Mr. Sakellariou will further describe the environment in which these federated algorithms run, the challenges and the available tools, the principles that guide its design and the followed general methodology for each new algorithm. One of the most important challenges which are faced is to design these tools in a way that does not compromise the privacy of the clinical data involved. The speaker will show how to address the main questions when designing such algorithms: how to decompose and distribute the computations and what kind of information to exchange between nodes, in order to comply with the privacy constraint mentioned above. Finally, also the subject of validating these federated algorithms will be briefly touched.