Skip to main content

This lesson briefly goes over the outline of the Neuroscience for Machine Learners course. 

Difficulty level: Intermediate
Duration: 3:05
Speaker: : Dan Goodman

This lesson goes over the basic mechanisms of neural synapses, the space between neurons where signals may be transmitted. 

Difficulty level: Intermediate
Duration: 7:03
Speaker: : Marcus Ghosh

While the previous lesson in the Neuro4ML course dealt with the mechanisms involved in individual synapses, this lesson discusses how synapses and their neurons' firing patterns may change over time. 

Difficulty level: Intermediate
Duration: 4:48
Speaker: : Marcus Ghosh

This lesson introduces some practical exercises which accompany the Synapses and Networks portion of this Neuroscience for Machine Learners course. 

Difficulty level: Intermediate
Duration: 3:51
Speaker: : Dan Goodman

This lesson describes spike timing-dependent plasticity (STDP), a biological process that adjusts the strength of connections between neurons in the brain, and how one can implement or mimic this process in a computational model. You will also find links for practical exercises at the bottom of this page. 

Difficulty level: Intermediate
Duration: 12:50
Speaker: : Dan Goodman

As the previous lesson of this course described how researchers acquire neural data, this lesson will discuss how to go about interpreting and analysing the data. 

Difficulty level: Intermediate
Duration: 9:24
Speaker: : Marcus Ghosh

In this lesson you will learn about the motivation behind manipulating neural activity, and what forms that may take in various experimental designs. 

Difficulty level: Intermediate
Duration: 8:42
Speaker: : Marcus Ghosh

In this lesson, you will learn about one particular aspect of decision making: reaction times. In other words, how long does it take to take a decision based on a stream of information arriving continuously over time?

Difficulty level: Intermediate
Duration: 6:01
Speaker: : Dan Goodman

In this lesson, you will hear about some of the open issues in the field of neuroscience, as well as a discussion about whether neuroscience works, and how can we know?

Difficulty level: Intermediate
Duration: 6:54
Speaker: : Marcus Ghosh

This lesson discusses a gripping neuroscientific question: why have neurons developed the discrete action potential, or spike, as a principle method of communication? 

Difficulty level: Intermediate
Duration: 9:34
Speaker: : Dan Goodman

This tutorial covers the fundamentals of collaborating with Git and GitHub.

Difficulty level: Intermediate
Duration: 2:15:50
Speaker: : Elizabeth DuPre
Course:

 

Panel discussion by leading scientists, engineers and philosophers discuss what brain-computer interfaces are and the unique scientific and ethical challenges they pose. hosted by Lynne Malcolm from ABC Radio National's All in the Mind program and features:

  • Dr Hannah Maslen, Deputy Director, Oxford Uehiro Centre for Practical Ethics, University of Oxford
  • Prof. Eric Racine, Director, Pragmatic Health Ethics Research Unity, Montreal Institute of Clinical Research
  • Prof Jeffrey Rosenfeld, Director, Monash Institute of Medical Engineering, Monash University
  • Dr Isabell Kiral-Kornek, AI and Life Sciences Researcher, IBM Research
  • A/Prof Adrian Carter, Neuroethics Program Coordinator, ARC Centre of Excellence for Integrative Brain Function

 

Difficulty level: Intermediate
Duration: 1:14:34
Course:

 

Panel of experts discuss the virtues and risks of our digital health data being captured and used by others in the age of Facebook, metadata retention laws, Cambridge Analytica and a rapidly evolving neuroscience. The discussion was moderated by Jon Faine, ABC Radio presenter. The panelists were:

  • Mr Sven Bluemmel, Victorian Information Commissioner
  • Prof Judy Illes, Neuroethics Canada, University of British Columbia, Order of Canada
  • Prof Mark Andrejevic, Professor of Media Studies, Monash University
  • Ms Vrinda Edan, Chief Operating Officer, Victorian Mental Illness Awareness Council


 

 

Difficulty level: Intermediate
Duration: 1:10:30

This is the Introductory Module to the Deep Learning Course at CDS, a course that covered the latest techniques in deep learning and representation learning, focusing on supervised and unsupervised deep learning, embedding methods, metric learning, convolutional and recurrent nets, with applications to computer vision, natural language understanding, and speech recognition.

Difficulty level: Intermediate
Duration: 50:17

This module covers the concepts of gradient descent and the backpropagation algorithm and is a part of the Deep Learning Course at NYU's Center for Data Science.

Difficulty level: Intermediate
Duration: 1:51:03
Speaker: : Yann LeCun

This lecture covers concepts associated with neural nets, including rotation and squashing, and is a part of the Deep Learning Course at New York University's Center for Data Science (CDS).

Difficulty level: Intermediate
Duration: 1:01:53
Speaker: : Alfredo Canziani

This lecture covers the concept of neural nets training (tools, classification with neural nets, and PyTorch implementation) and is a part of the Deep Learning Course at NYU's Center for Data Science.

Difficulty level: Intermediate
Duration: 1:05:47
Speaker: : Alfredo Canziani

This lecture covers the concept of parameter sharing: recurrent and convolutional nets and is a part of the Deep Learning Course at NYU's Center for Data Science.

Difficulty level: Intermediate
Duration: 1:59:47

This lecture covers the concept of convolutional nets in practice and is a part of the Deep Learning Course at NYU's Center for Data Science.

Difficulty level: Intermediate
Duration: 51:40
Speaker: : Yann LeCun

This lecture discusses the concept of natural signals properties and the convolutional nets in practice and is a part of the Deep Learning Course at NYU's Center for Data Science.

Difficulty level: Intermediate
Duration: 1:09:12
Speaker: : Alfredo Canziani