2022
Projects:
Brain Tumor Subtyping
AUTHOR: Jędrzej Kubica
AFFILIATION: University of Warsaw
Brain tumors can be classified into different subtypes to recommend a patient-specific treatment. Characterization of patients into groups based on molecular features can help clinicians to choose proper medications for each individual and to improve the outcome of the treatment. Various classification tools for tumors of the central nervous system have been developed. For instance, MethPed [1] is an open-access classifier which uses DNA methylation as an epigenetic marker for subtype classification. The goal of the project is to develop of a data analysis workflow to analyze publicly-available epigenetic data of brain tumor patients from databases, such as The Cancer Genome Atlas. Future work will include subtype-specific drug recommendations. Further research can also be extended from brain tumors into tumors of the nervous system.
Application of an amygdala parcellation pipeline based on Recurrence Quantification Analysis to resting-state fMRI data acquired in a 7T MRI scanner
AUTHORS: Sylwia Adamus
AFFILIATION: Medical University of Warsaw, University of Warsaw Faculty of Physics
According to animal studies the amygdala consists of several groups of nuclei, which play different roles in emotion-related processes. It has also been shown that this brain structure is important for the development of many psychiatric conditions, such as depression, addictions, and autism spectrum disorder. Up to this day, a number of approaches to the topic of amygdala parcellation have been suggested. One of them, which was recently published, is a pipeline using Recurrence Quantification Analysis (RQA). It enables the division of the human amygdala into two functionally different parts based on brain signal dynamics [1]. The aim of this project is to further develop this pipeline and check whether with its help it is possible to divide the amygdala into more than two subdivisions. To achieve this the pipeline will be applied to resting-state fMRI data acquired in a 7T MRI scanner from a dataset consisting of 184 healthy subjects from the Human Connectome Project. An exploratory approach will be applied using several variations of RQA parameters and clustering algorithms. The pipeline was developed on resting-state fMRI data acquired in a 3T MRI scanner. It has been speculated, that its application to data from a 7T MRI scanner could enable obtaining more detailed parcellations. Therefore the main hypothesis behind this project is that by using this pipeline it will be possible to achieve a parcellation with at least three functionally different subdivisions. It could have the potential to serve as a mask in further studies of human amygdala functional connectivity. All participants will have the opportunity to go through the whole pipeline to fully explore its possibilities. A device, which supports Anaconda Navigator is necessary to take part in the project. No advanced neuroscientific knowledge is required and everyone, who knows some basics of Python programming, is invited to cooperate. References: [1] Bielski K. et al., (2021) NeuroImage 227(117644)
Automatic artifact detection in EEG using Telepathy
AUTHOR: Marcin Koculak
AFFILIATION: Consciousness Lab, Institute of Psychology, Jagiellonian University
This will be a follow-up to a project from Brainhack Warsaw 2022, where we built a library for EEG analysis from scratch. That library since then was named „Telepathy” and is being actively developed by the author of this project. In this edition I would like to focus on implementing methods for automatic detection of artefactual signal in EEG recording, along with methods to deal with them. This includes correcting hardware related issues, rejecting signals coming from non-brain sources, and methods for dimensionality reduction, all of which help to isolate the signal of interest from EEG. Participants will have a chance to better understand sources of noise in EEG data, see what analysis software does under the hood to deal with them, and attempt to implement them in new and dynamically evolving language for scientific computing. If possible, I will try to organise a small demonstration how EEG signal is collected and from where the artefacts might come from.
Pose estimation based long term behavior assesement of animals in semi-naturalistic condtions
AUTHORS: Konrad Danielewski, Marcin Lipiec, Ewelina Knapska, Alexander Mathis
AFFILIATION: Nencki Institute of Experimental Biology
Using Eco-HAB system of four cages equiped with RFID antennas (RFID tagged group of 12 mice) and DeepLabCut, SOTA pose estimation framework we want to create a framework for behavior analysis of those animals. A set of 9 bodyparts is tracked on each animal (pose estimation model is ready and working well). Our test recording is 3 hours long but the goal is for the system to perform non-stop for a week. A synchronization script that will work based on visual information is needed, as there is no TTL signal from the antennas – camera timestamps are available and between frame times are very precise (at the scale of 10s of microseconds – FLIR BFS camera is used). We aim to develop an algorithm that will filter identity based on antenna readout and correct any potential switches. With identity corrected DLC2Kinematics and DLC2Action with some additional functions can be used to assess animal behavior for single animals and pairwise social interations. The goal would be to combine all of this into a user friendly framework with a set of high-level functions and thorough documentation that can be open-sourced and used by anyone interested in highly detailed long-term, homecage monitoring and behavior assesement.
A paradigm shift in experiment design leading to large scale EEG data acquisition for visual attention
AUTHOR: Shaurjya Mandal
AFFILIATION: Carnegie Mellon University
Visual attention often refers to the cognitive processes that allow the selective processing of visual information that we are exposed to in our daily lives. Gaining an understanding of visual attention can be crucial to a number of applications like the study of human-computer interaction and analysis and improvement of advertisements. EEG along with eye-tracking has been popularly used to study visual attention in subjects. In deep learning applications, to model needs to be trained on a significantly large dataset. Although the results obtained with deep learning algorithms tend to be highly accurate, it is always hard to acquire such a dataset. The apparatus required to track the eye movements and comment on visual attention from the gaze of the subject are not portable. Thus, to sync the EEG data with eye tracking/ visual attention data outside the laboratory setup, we would need an additional approach. To study the same, we ask two important questions:
1. To what extent can multi-channel EEG data provide inference regarding eye movement and eye tracking? And are there any specific experiments that help us to observe the behaviour better?
2. Is there a way to find an alternative to proper eye-tracking that can be crowdsourced? To answer the above questions, we would start with analysing 3 distinct publicly available datasets.
EEGEyeNet dataset comprises of EEG and eye-tracking data from 350 participants, 190 female and 160 male, between the ages of 18 and 80 years. This dataset provides enough data to correlate the changes in EEG based on the movement of the eyes. The eye-tracking experiments with synchronized EEG have been divided into 3 main parts: Pro and antisaccade, large grid, and visual symbol search. Knowing the protocols of the experiment while analysing the data would allow gain a better understanding of eye-movement with the EEG data. In the previous literature, eye-tracking for visual attention is linked to mouse-tracking. But to allow this, specific protocols have to be adopted that causes takes care of localised visual exposure with minimal distraction. To validate our performance across eye-tracking and mouse-tracking scenarios, we will make use of the OSIE dataset. OSIE dataset consists of 700 natural images along with publicly available mouse tracking and eye tracking data. To train the deep learning model that will allow us to generate labels for visual attention, we have made use of the SALICON dataset. The dataset consists of 10,000 MS COCO training images. By the end of the project, we hope to:
1. Analyse the pretrained models for visual attention with eye tracking and train our own models with the mouse-tracking data and compare the models.
2. Through effective data analysis, reduce the number of channels while preserving the variance of EEG data during eye-tracking experiments.
3. Develop our own custom software which can be used as a convenient means to collect and sync EEG data to visual attention outside laboratory settings.
Robust Latent Space Exploration of EEG Signals with Distributed Representations
AUTHOR: Adam Sobieszek
AFFILIATION: University of Warsaw
It is said that the advantage of methods, such as Variational Autoencoders (VAEs) and Generative Adversarial Networks (GANs), is their learning of latent space representations of the target domain. E.g. training a GAN that generates EEG signals gives us a latent space representation of the type of EEG signals that were in the training set. These representations could moreover be made „disentangled”, which uncovers independent dimensions that are features that best describe the target domain. Such methods could give us the ability to describe and better understand the type of information contained in EEG signals. However, for such methods to be widely accepted as scientific tools, we need to overcome the issue that each time we train a neural network, the learnt latent space seems on the surface to be completely different. Recently we investigated the ability of GANs to learn disentangled representations of EEG signals. With our network, we’ve obtained multiple latent space representations of EEG signals. The goal of this project is to investigate whether such representations hold the same kind of information, possibly showing the robust nature of neural network representation learning, as well as investigate how such multiple (distributed) representations may be used together. We will implement down-stream tasks, that latent representations are useful for (such as classification, and explainable feature visualisation) and compare how their performance differs between different representations. We will investigate whether combining representations may result in an increase in accuracy. We will test methods for learning probabilistic maps between latent spaces, which could prove useful for a wider array of machine learning applications. Finally, we will attempt to find in these distributed representations the common factors, that are robustly able to describe EEG signals. If time allows we will also attempt to train disentangled variational autoencoders (called beta-VAEs) for generating EEG signals and investigate whether the factors discovered by this method are similar to those found with our GAN-based method.
