Welcome! This is the home page for the Winter 2016 edition of the “Neural Data Analysis” course.
Basic familiarity with MATLAB. Depending on your background and programming experience you might find the following resources helpful:
If you are unsure, take a look at the table of contents of the Primer. If there are things you don't recognize, use the Primer itself, or Chapter 2 of the MATLAB for Neuroscientists book to get up to speed.
Regardless of your MATLAB abilities, some great ways to keep learning are:
If you have no formal training in computer programming (i.e. you have never taken a “Intro to Computer Science” or “Introductory Programming” type course) you will almost certainly find what follows in this course less frustrating if you do the pen-and-paper exercises in this short chapter by Daniel Dennett (“The Secrets of Computer Power Revealed”) before you embark on the MATLAB primer linked to above.
This course is “standalone”, but the following textbooks provide more in-depth treatment of some of the topics.
Overall, the course is designed to provide hands-on experience with management, visualization, and analysis of neural data. Becoming skilled at these things is a rate-limiting step for many graduate projects requiring analysis. Even if your work only requires rudimentary analysis, awareness of what else can be done and how to do it well is valuable, for instance when evaluating the work of others in the literature!
To do so, the focus is on introducing some commonly used tools, such as GitHub and relevant functionality within MATLAB – and then to actually use these on real data sets. Initially, those data sets will be local field potentials and spiking data recorded from various brain areas in freely moving rodents in the van der Meer lab; however, a crucial goal of the course is that after some initial steps you will use your own data. (If you don't have your own, you can do everything with data provided here.)
We will make contact with a few concepts from computer science, signal processing, and statistics. However, the focus is on making initial steps that work and getting pointers to more complete treatment, rather than a thorough theoretical grounding. Nevertheless, to make sure that what you learn is not tied to specific data sets only, a number of principles of data analysis – applicable to any project of sufficient complexity – will be referenced throughout the course. You are invited to think of these and others, not only as you progress through this course, but especially as you organize your own data analyses and read analyses performed by others.
This course will provide a brief introduction to a number of concepts which are themselves the subject of multiple courses and voluminous textbooks. These include signal processing topics such as Fourier analysis and filter design, computer science concepts such as object-oriented programming and binary data formats, and a number of statistical ideas and tools. Be aware that if any of these are particularly important to your research, you should consider taking more in-depth coursework and/or working through relevant textbooks on your own: this short tutorial cannot replace such courses!
Most modules finish with a challenge (or several), in which you are invited to implement some of the ideas in the module yourself. Pick one such challenge from the first half of modules (1-7) and another from the second half (9-16). Submit your code for two chosen challenges to a GitHub repository you created, along with documentation: instructions on what it is supposed to do, how to make it run if applicable, and comments explaining how the code works.
The lab codebase is set up for machines running 64-bit Windows 7 or Mac OS X. If you want to use Linux or some other OS you will probably need to compile some of the low-level loading functions yourself. Some pointers for this are provided in subsequent modules when loading is introduced.
The architecture of the code used in this course was inspired by a similar set of code by my post-doctoral mentor, A. David Redish; several of the data types and functions are re-implementations of Redish lab functions of the same name. Major contributions to the codebase were made by Alyssa Carey (a MSc student and research assistant in the lab) and Youki Tanaka (current PhD student).