This lecture provides an introduction to optogenetics, a biological technique to control the activity of neurons or other cell types with light.
This primer on optogenetics primer discusses how to manipulate neuronal populations with light at millisecond resolution and offers possible applications such as curing the blind and "playing the piano" with cortical neurons.
This lecture covers the history of behaviorism and the ultimate challenge to behaviorism.
In this lesson, you will learn how to utilize various features and tools included in the EBRAINS platform, particularly focusing on rodent brain atlases and how to incorporate them into your analyses.
This talk describes how to use DataLad for your data management and curation techniques when dealing with animal datasets, which often contain several disparate types of data, including MRI, microscopy, histology, electrocorticography, and behavioral measurements.
In this short talk you will learn about The Neural System Laboratory, which aims to develop and implement new technologies for analysis of brain architecture, connectivity, and brain-wide gene and molecular level organization.
In this lesson, you will learn about the connectome, the collective system of neural pathways in an organism, with a closer look at the neurons, synapses, and connections of particular species.
This lesson introduces the practical exercises which accompany the previous lessons on animal and human connectomes in the brain and nervous system.
In this lecture, attendees will learn how Mutant Mouse Resource and Research Center (MMRRC) archives, cryopreserves, and distributes scientifically valuable genetically engineered mouse strains and mouse ES cell lines for the genetics and biomedical research community.
This lecture discusses how to standardize electrophysiology data organization to move towards being more FAIR.
This is the Introductory Module to the Deep Learning Course at CDS, a course that covered the latest techniques in deep learning and representation learning, focusing on supervised and unsupervised deep learning, embedding methods, metric learning, convolutional and recurrent nets, with applications to computer vision, natural language understanding, and speech recognition.
This module covers the concepts of gradient descent and the backpropagation algorithm and is a part of the Deep Learning Course at NYU's Center for Data Science.
This lecture covers concepts associated with neural nets, including rotation and squashing, and is a part of the Deep Learning Course at New York University's Center for Data Science (CDS).
This lesson provides a detailed description of some of the modules and architectures involved in the development of neural networks.
This lecture covers the concept of neural nets training (tools, classification with neural nets, and PyTorch implementation) and is a part of the Deep Learning Course at NYU's Center for Data Science.
This lecture covers the concept of parameter sharing: recurrent and convolutional nets and is a part of the Deep Learning Course at NYU's Center for Data Science.
This lecture covers the concept of convolutional nets in practice and is a part of the Deep Learning Course at NYU's Center for Data Science.
This lecture discusses the concept of natural signals properties and the convolutional nets in practice and is a part of the Deep Learning Course at NYU's Center for Data Science.
This lecture covers the concept of recurrent neural networks: vanilla and gated (LSTM) and is a part of the Deep Learning Course at NYU's Center for Data Science.
This lecture is a foundationational lecture for the concept of energy-based models with a particular focus on the joint embedding method and latent variable energy-based models (LV-EBMs) and is a part of the Deep Learning Course at NYU's Center for Data Science.