This lecture describes how to build research workflows, including a demonstrate using DataJoint Elements to build data pipelines.
This lesson provides an introduction to the Symposium on Science Management at the Canadian Association for Neuroscience 2019 Meeting.
This lesson gives a primer to project management in a scientific context, with a particular neuroinformatic case study.
In this lesson, you will hear about the current challenges regarding data management, as well as policies and resources aimed to address them.
This lesson covers "Knowledge Translation", the activities involved in moving research from the laboratory, the research journal, and the academic conference into the hands of people and organizations who can put it to practical use.
In this lesson, you will hear about the various methods developed and employed in managing performance.
This lesson provides an overview of how to manage relationships in a research context, while highlighting the need for effective communication at various levels.
In this lesson you will hear a panel discussion which hosts experts in the field whom have extensive experience with management in a science setting.
Manipulate the default connectome provided with TVB to see how structural lesions effect brain dynamics. In this hands-on session you will insert lesions into the connectome within the TVB graphical user interface (GUI). Afterwards, the modified connectome will be used for simulations and the resulting activity will be analysed using functional connectivity.
This presentation discusses the impact of data sharing in stroke.
This talks presents an overview of the potential for data federation in stroke research.
This talk focuses on the EAN Scientific Panel of Stroke, in particular on the aims and roles of the panel.
This is the Introductory Module to the Deep Learning Course at CDS, a course that covered the latest techniques in deep learning and representation learning, focusing on supervised and unsupervised deep learning, embedding methods, metric learning, convolutional and recurrent nets, with applications to computer vision, natural language understanding, and speech recognition.
This module covers the concepts of gradient descent and the backpropagation algorithm and is a part of the Deep Learning Course at NYU's Center for Data Science.
This lecture covers concepts associated with neural nets, including rotation and squashing, and is a part of the Deep Learning Course at New York University's Center for Data Science (CDS).
This lesson provides a detailed description of some of the modules and architectures involved in the development of neural networks.
This lecture covers the concept of neural nets training (tools, classification with neural nets, and PyTorch implementation) and is a part of the Deep Learning Course at NYU's Center for Data Science.
This lecture covers the concept of parameter sharing: recurrent and convolutional nets and is a part of the Deep Learning Course at NYU's Center for Data Science.
This lecture covers the concept of convolutional nets in practice and is a part of the Deep Learning Course at NYU's Center for Data Science.
This lecture discusses the concept of natural signals properties and the convolutional nets in practice and is a part of the Deep Learning Course at NYU's Center for Data Science.