In this lesson, you will learn about hardware for computing for non-ICT specialists.
This lecture covers different perspectives on the study of the mental, focusing on the difference between Mind and Brain.
This lecture provides a history of data management, recent developments data management, and a brief description of scientific data management.
Computer arithmetic is necessarily performed using approximations to the real numbers they are intended to represent, and consequently it is possible for the discrepancies between the actual solution and the approximate solutions to diverge, i.e. to become increasingly different. This lecture focuses on how this happens and techniques for reducing the effects of these phenomena and discuss systems which are chaotic.
This lecture will addresses what it means for a problem to have a computable solution, methods for combining computability results to analyse more complicated problems, and finally look in detail at one particular problem which has no computable solution: the halting problem.
This lecture focuses on computational complexity, a concept which lies at the heart of computer science thinking. In short, it is a way to quickly gauge an approximation to the computational resource required to perform a task.
JupyterHub is a simple, highly extensible, multi-user system for managing per-user Jupyter Notebook servers, designed for research groups or classes. This lecture covers deploying JupyterHub on a single server, as well as deploying with Docker using GitHub for authentication.
This tutorial is part 1 of 2. It aims to provide viewers with an understanding of the fundamentals of R tool. Note: parts 1 and 2 of this tutorial are part of the same YouTube video; part 1 ends at 17:42.
This tutorial is part 2 of 2. It aims to provide viewers with an understanding of the fundamentals of R tool. Note: parts 1 and 2 of this tutorial are the same YouTube video. The portion related to this tutorial begins at 17:43.
The Virtual Brain (TVB) is an open-source, multi-scale, multi-modal brain simulation platform. In this lesson, you get introduced to brain simulation in general and to TVB in particular. This lesson also presents the newest approaches for clinical applications of TVB - that is, for stroke, epilepsy, brain tumors, and Alzheimer’s disease - and show how brain simulation can improve diagnostics, therapy, and understanding of neurological disease.
This lesson explains the mathematics of neural mass models and their integration to a coupled network. You will also learn about bifurcation analysis, an important technique in the understanding of non-linear systems and as a fundamental method in the design of brain simulations. Lastly, the application of the described mathematics is demonstrated in the exploration of brain stimulation regimes.
In this lesson, the simulation of a virtual epileptic patient is presented as an example of advanced brain simulation as a translational approach to deliver improved clinical results. You will learn about the fundamentals of epilepsy, as well as the concepts underlying epilepsy simulation. By using an iPython notebook, the detailed process of this approach is explained step by step. In the end, you are able to perform simple epilepsy simulations your own.
This lesson introduces the practical usage of The Virtual Brain (TVB) in its graphical user interface and via python scripts. In the graphical user interface, you are guided through its data repository, simulator, phase plane exploration tool, connectivity editor, stimulus generator, and the provided analyses. The implemented iPython notebooks of TVB are presented, and since they are public, can be used for further exploration of TVB.
This lesson provides a comprehensive introduction to the command line and 50 popular Linux commands. This is a long introduction (nearly 5 hours), but well worth it if you are going to spend a good part of your career working from a terminal, which is likely if you are interested in flexibility, power, and reproducibility in neuroscience research. This lesson is courtesy of freeCodeCamp.
This book was written with the goal of introducing researchers and students in a variety of research fields to the intersection of data science and neuroimaging. This book reflects our own experience of doing research at the intersection of data science and neuroimaging and it is based on our experience working with students and collaborators who come from a variety of backgrounds and have a variety of reasons for wanting to use data science approaches in their work. The tools and ideas that we chose to write about are all tools and ideas that we have used in some way in our own research. Many of them are tools that we use on a daily basis in our work. This was important to us for a few reasons: the first is that we want to teach people things that we ourselves find useful. Second, it allowed us to write the book with a focus on solving specific analysis tasks. For example, in many of the chapters you will see that we walk you through ideas while implementing them in code, and with data. We believe that this is a good way to learn about data analysis, because it provides a connecting thread from scientific questions through the data and its representation to implementing specific answers to these questions. Finally, we find these ideas compelling and fruitful. That’s why we were drawn to them in the first place. We hope that our enthusiasm about the ideas and tools described in this book will be infectious enough to convince the readers of their value.
This talk presents state-of-the-art methods for ensuring data privacy with a particular focus on medical data sharing across multiple organizations.
This lecture talks about the usage of knowledge graphs in hospitals and related challenges of semantic interoperability.