In this lecture, you will learn about virtual research environments (VREs) and their technical limitations, (i.e., a computing platform and the software stack behind it) and the security measures which should be considered during implementation.
In this workshop talk, you will receive a tour of the Code Ocean ScienceOps Platform, a centralized cloud workspace for all teams.
This talk describes approaches to maintaining integrated workflows and data management schema, taking advantage of the many open source, collaborative platforms already existing.
This lesson consists of a panel discussion, wrapping up the INCF Neuroinformatics Assembly 2023 workshop Research Workflows for Collaborative Neuroscience.
This brief talk outlines the obstacles and opportunities involved in striving for more open and reproducible publishing, highlighting the need for investment in the technical and governance sectors of FAIR data and software.
This talk gives an overview of the complicated nature of sharing of neuroscientific data in an environment of numerous and often conflicting legal systems around the world.
This lecture provides a history of data management, recent developments data management, and a brief description of scientific data management.
This talk provides an overview of the FAIR-aligned efforts of MATLAB and MathWorks, from the technological building blocks to the open science coordination involved in facilitating greater transparency and efficiency in neuroscience and neuroinformatics.
This brief video provides a welcome and short introduction to the outline of the INCF Short Course in Neuroinformatics, held Seattle, Washington in October 2023, in coordination with the West Big Data Hub and the University of Washington.
This opening lecture from INCF's Short Course in Neuroinformatics provides an overview of the field of neuroinformatics itself, as well as laying out an argument for the necessity for developing more sophisticated approaches towards FAIR data management principles in neuroscience.
This lesson provides a thorough description of neuroimaging development over time, both conceptually and technologically. You will learn about the fundamentals of imaging techniques such as MRI and PET, as well as how the resultant data may be used to generate novel data visualization schemas.
This lesson contains the first part of the lecture Data Science and Reproducibility. You will learn about the development of data science and what the term currently encompasses, as well as how neuroscience and data science intersect.
In this second part of the lecture Data Science and Reproducibility, you will learn how to apply the awareness of the intersection between neuroscience and data science (discussed in part one) to an understanding of the current reproducibility crisis in biomedical science and neuroscience.
This lesson aims to define computational neuroscience in general terms, while providing specific examples of highly successful computational neuroscience projects.
This lecture covers a wide range of aspects regarding neuroinformatics and data governance, describing both their historical developments and current trajectories. Particular tools, platforms, and standards to make your research more FAIR are also discussed.
This lesson gives an in-depth description of scientific workflows, from study inception and planning to dissemination of results.
Computer arithmetic is necessarily performed using approximations to the real numbers they are intended to represent, and consequently it is possible for the discrepancies between the actual solution and the approximate solutions to diverge, i.e. to become increasingly different. This lecture focuses on how this happens and techniques for reducing the effects of these phenomena and discuss systems which are chaotic.
This lecture will addresses what it means for a problem to have a computable solution, methods for combining computability results to analyse more complicated problems, and finally look in detail at one particular problem which has no computable solution: the halting problem.
This lecture focuses on computational complexity, a concept which lies at the heart of computer science thinking. In short, it is a way to quickly gauge an approximation to the computational resource required to perform a task.