This lecture covers different perspectives on the study of the mental, focusing on the difference between Mind and Brain.
This lecture focuses on where and how Jupyter notebooks can be used most effectively for education.
JupyterHub is a simple, highly extensible, multi-user system for managing per-user Jupyter Notebook servers, designed for research groups or classes. This lecture covers deploying JupyterHub on a single server, as well as deploying with Docker using GitHub for authentication.
The Virtual Brain (TVB) is an open-source, multi-scale, multi-modal brain simulation platform. In this lesson, you get introduced to brain simulation in general and to TVB in particular. This lesson also presents the newest approaches for clinical applications of TVB - that is, for stroke, epilepsy, brain tumors, and Alzheimer’s disease - and show how brain simulation can improve diagnostics, therapy, and understanding of neurological disease.
This lesson explains the mathematics of neural mass models and their integration to a coupled network. You will also learn about bifurcation analysis, an important technique in the understanding of non-linear systems and as a fundamental method in the design of brain simulations. Lastly, the application of the described mathematics is demonstrated in the exploration of brain stimulation regimes.
In this lesson, the simulation of a virtual epileptic patient is presented as an example of advanced brain simulation as a translational approach to deliver improved clinical results. You will learn about the fundamentals of epilepsy, as well as the concepts underlying epilepsy simulation. By using an iPython notebook, the detailed process of this approach is explained step by step. In the end, you are able to perform simple epilepsy simulations your own.
This lesson introduces the practical usage of The Virtual Brain (TVB) in its graphical user interface and via python scripts. In the graphical user interface, you are guided through its data repository, simulator, phase plane exploration tool, connectivity editor, stimulus generator, and the provided analyses. The implemented iPython notebooks of TVB are presented, and since they are public, can be used for further exploration of TVB.
This lesson provides a brief overview of the Python programming language, with an emphasis on tools relevant to data scientists.
This lecture covers FAIR atlases, including their background and construction, as well as how they can be created in line with the FAIR principles.
This lesson contains both a lecture and a tutorial component. The lecture (0:00-20:03 of YouTube video) discusses both the need for intersectional approaches in healthcare as well as the impact of neglecting intersectionality in patient populations. The lecture is followed by a practical tutorial in both Python and R on how to assess intersectional bias in datasets. Links to relevant code and data are found below.
This lecture discusses what defines an integrative approach regarding research and methods, including various study designs and models which are appropriate choices when attempting to bridge data domains; a necessity when whole-person modelling.
This lecture covers an Introduction to neuron anatomy and signaling, and different types of models, including the Hodgkin-Huxley model.
This lesson gives an introductory presentation on how data science can help with scientific reproducibility.
This talk highlights a set of platform technologies, software, and data collections that close and shorten the feedback cycle in research.
This lesson gives an in-depth introduction of ethics in the field of artificial intelligence, particularly in the context of its impact on humans and public interest. As the healthcare sector becomes increasingly affected by the implementation of ever stronger AI algorithms, this lecture covers key interests which must be protected going forward, including privacy, consent, human autonomy, inclusiveness, and equity.
This lesson describes a definitional framework for fairness and health equity in the age of the algorithm. While acknowledging the impressive capability of machine learning to positively affect health equity, this talk outlines potential (and actual) pitfalls which come with such powerful tools, ultimately making the case for collaborative, interdisciplinary, and transparent science as a way to operationalize fairness in health equity.
This lesson contains the first part of the lecture Data Science and Reproducibility. You will learn about the development of data science and what the term currently encompasses, as well as how neuroscience and data science intersect.
This lecture presents selected theories of ethics as applied to questions raised by the Human Brain Project.
The HBP as an ICT flagship project crucially relies on ICT and will contribute important input into the development of new computing principles and artefacts. Individuals working on the HBP should therefore be aware of the long history of ethical issues discussed in computing. This lessson provides an overview of the most widely discussed ethical issues in computing and demonstrate that privacy and data protection are by no means the only issue worth worrying about.
This lecture explores two questions regarding the ethics of robot development and use. Firstly, the increasingly urgent question of the ethical use of robots: are there particular applications of robots that should be proscribed, in eldercare, or surveillance, or combat? Secondly, the talk deals with the longer-term question of whether intelligent robots themselves could or should be ethical.