Send an email to [email protected] to make sure your submission will be judged!!!
The Santa Cruz Machine Learning Cooperative Cooperative would like to announce the following challenge, open to the public, to help us push forward our understanding of how visual information is processed in mammals and whether machines can replicate it. This is in collaboration with SCIPP (the Santa Cruz Institute for Particle Physics), which develops a device that records from many neurons simultaneously from a live mouse.
-
The competition begins on October 5th, 2017 and will end November 26th, 2017 at 11:59PM!
-
Fork this repo and open the notebook to get started!
Your task is to create a model that reproduces visually evoked spiking activity of the neurons in the superior colliculus. The superior colliculus is a brain structure, behind the cortex, which is directly connected to the retina. It receives visual, auditory, and somatosensory information, combines them together, and sends a motor signal to direct the eyes and head toward an object of an interest. In this competition, we focus on modeling how the visual information is encoded in neuronal activity in the superior colliculus by predicting the timing of spike trains in response to a stimulus.
The input data of your model is a short movie that the mouse saw during the experiment; the output data is time series of the firing rates of all the neurons we recorded. As a training dataset, we provide short clips of movies and corresponding firing rates of the neurons.
The original unedited movie is a YouTube video, which can be seen here. The grayscale version of this movie was presented 6 times during the experiment.
The movie is 290 seconds long, and was divided into 58 five second segments. 48 segments are provided for the training dataset and 10 segments have been randomly removed to later judge the quality of your response.
The provided dataset contains 288 (48 * 6
) 5 second clips (150 frames @ 30Hz
, or 33.3ms/frame
) with X-Y pixels already linearized (128 x 96 -> 12288 pixels
) for training convenience. The movie clips are provided as a 3D numpy
array (288 clips x 150 frames x 12288 pixels
). The notebook demonstrates how the release data can be easily loaded.
The output data, which your model attempts to predict, are spiking activity of 54 neurons in the mouse superior colliculus while the mouse watches the movie. The activity of each neuron has been binned using the same time resolution as the movie for convenience. The output
data used for training your model are provided as a 3D numpy array, similar to the movie (288 clips x 150 frames, x 54 neurons
).
These data can be downloaded from the Github repository of the competition at the data release.
Any algorithm is allowed in this competition! You can use the algorithm that you think is best suited for the goal of the competition. Please describe the method either in the comment in the code or documentation of your entry repository.
The notebook is provided for demonstration or as a basis to begin your modeling effort.
A similar work has been done using the data from ex-vivo retina combined with convnet based network: Deep Learning Models of the Retinal Response to Natural Scenes. This article may give a good introduction for what type of things we would like to achieve.
If you are in need of GPU resources for entering the competition, and have an active UC login, you may make use of the pattern lab server following these instructions.
If you don't have jupyter notebook you can get help installing it here!
Your entry will be judged on a variety of criteria.
To judge an entry on its ability to use resources effectively, the entry must include details of the computation and resources used to create a response.
Simplicity and explainability are desirable characteristics when working with machine learning results. The interpretability of the model and the complexity of the engineering effort are taken into consideration.
Entries will be judged on their ability to work with an unseen dataset. Any trained models used should be supplied with the response as well as providing a function that can test new samples.
To enter a response simply fork this repository to your own github account. Anyone with a github account may submit a response. The tip of the master
branch November 26, 11:59 PM will be considered your entry. Please make it clear within your code how to direct the model to make predictions from new output data, which will be used by judges to test the precision and recall of your entry.
To fork the repository to your own account, start by clicking the fork icon on github.
Then use git clone https://github.com/YOURACCOUNT/fall-17-scml-competition
to copy the code to your local machine. You can then make modifications and later upload your changes using git add .
, git commit -m "my changes"
, git push origin master
.
No responses will be accepted after November 26, 11:59 pm. Only commits made before this time will be judged.
All entrants will be contacted that their response is under review by November 29, 11:59 pm.
A list of awards will be given on December 6th at the SCML meeting!
Please join us for the student meetings to get help on formulating an entry! On the November 22nd meeting of the group entrants will have an opportunity to discuss their competition entry and any challenges in submitting.
You can also post your questions to the github issues or mailing list!
For this challenge these judges have agreed to work with participants to judge the quality of responses. These are subject to change.
Shinya Ito Postdoctoral Scholar, Santa Cruz Institute for Particle Physics
Rob Currie Chief Technical Officer, UCSC Genomics Institute
David Steinberg Bioinformatics Programmer, UCSC Genomics Institute
Judges are expected to use a balance of their expertise and automated methods to judge the quality of a response.
We value work that balances the amount of computational and human effort that goes into a response. All entrants who provide a response by the deadline will receive feedback from our judges and a one-of-a-kind way to remember your effort!
The entry that best balances human effort, resource consumption, and and accuracy takes the Grand Prize.
The brevity prize is the entry that satisfies the challenge requirements in the fewest lines of written code.
The entry that uses the least computational resources, regardless of environment will take the Economy Prize.
The entry that achieves the highest accuracy gets the precision prize.
Please post an issue or send a message to the mailing list if you have a question!