This lesson provides a hands-on, Jupyter-notebook-based tutorial to apply machine learning in Python to brain-imaging data.
In this lecture, attendees will learn about the opportunities and challenges associated with Recurrent Neural Networks (RNNs), which, when trained with machine learning techniques on cognitive tasks, have become a widely accepted tool for neuroscientists.
This lesson gives an overview of past and present neurocomputing approaches and hybrid analog/digital circuits that directly emulate the properties of neurons and synapses.
Presentation of the Brian neural simulator, where models are defined directly by their mathematical equations and code is automatically generated for each specific target.
This lesson covers Big O notation, a mathematical notation that describes the limiting behavior of a function as it tends towards a certain value or infinity, proving useful for data scientists who want to evaluate their algorithms' efficiency.
Serving as good refresher, this lesson explains the maths and logic concepts that are important for programmers to understand, including sets, propositional logic, conditional statements, and more.
This compilation is courtesy of freeCodeCamp.
This lesson provides a useful refresher which will facilitate the use of Matlab, Octave, and various matrix-manipulation and machine-learning software.
This lesson was created by RootMath.
This lecture covers a lot of post-war developments in the science of the mind, focusing first on the cognitive revolution, and concluding with living machines.
This lecture provides an overview of depression (epidemiology and course of the disorder), clinical presentation, somatic co-morbidity, and treatment options.
This lesson is part 1 of 2 of a tutorial on statistical models for neural data.