This lesson continues with the second workshop on reproducible science, focusing on additional open source tools for researchers and data scientists, such as the R programming language for data science, as well as associated tools like RStudio and R Markdown. Additionally, users are introduced to Python and iPython notebooks, Google Colab, and are given hands-on tutorials on how to create a Binder environment, as well as various containers in Docker and Singularity.
This lesson gives an in-depth introduction of ethics in the field of artificial intelligence, particularly in the context of its impact on humans and public interest. As the healthcare sector becomes increasingly affected by the implementation of ever stronger AI algorithms, this lecture covers key interests which must be protected going forward, including privacy, consent, human autonomy, inclusiveness, and equity.
This lesson describes a definitional framework for fairness and health equity in the age of the algorithm. While acknowledging the impressive capability of machine learning to positively affect health equity, this talk outlines potential (and actual) pitfalls which come with such powerful tools, ultimately making the case for collaborative, interdisciplinary, and transparent science as a way to operationalize fairness in health equity.
In this lesson, while learning about the need for increased large-scale collaborative science that is transparent in nature, users also are given a tutorial on using Synapse for facilitating reusable and reproducible research.
This lecture provides an introduction to Plato’s concept of rationality and Aristotle’s concept of empiricism, and the enduring discussion between rationalism and empiricism to this day.
This lecture goes into further detail about the hard problem of developing a scientific discipline for subjective consciousness.
This lecture covers a lot of post-war developments in the science of the mind, focusing first on the cognitive revolution, and concluding with living machines.
This lesson gives a brief introduction to the course Neuroscience for Machine Learners (Neuro4ML).
This lesson covers the history of neuroscience and machine learning, and the story of how these two seemingly disparate fields are increasingly merging.
In this lesson, you will learn about the current challenges facing the integration of machine learning and neuroscience.
This lecture provides an overview of depression (epidemiology and course of the disorder), clinical presentation, somatic co-morbidity, and treatment options.
In this lesson, you will learn about how genetics can contribute to our understanding of psychiatric phenotypes.
This lesson gives an introductory presentation on how data science can help with scientific reproducibility.
This lesson discusses FAIR principles and methods currently in development for assessing FAIRness.
This lesson provides a brief overview of the Python programming language, with an emphasis on tools relevant to data scientists.
The lecture provides an overview of the core skills and practical solutions required to practice reproducible research.
This lecture covers FAIR atlases, including their background and construction, as well as how they can be created in line with the FAIR principles.
This lecture covers multiple aspects of FAIR neuroscience data: what makes it unique, the challenges to making it FAIR, the importance of overcoming these challenges, and how data governance comes into play.
This lecture covers how to make modeling workflows FAIR by working through a practical example, dissecting the steps within the workflow, and detailing the tools and resources used at each step.
This lecture provides guidance on the ethical considerations the clinical neuroimaging community faces when applying the FAIR principles to their research.