List of Project Mentors
Mentor: Quique Toloza
Project:
Biological neurons exhibit multiple layers of input-output processing, resulting in individual units that are much more complex than the simple units used in most artificial neural networks. The benefits of having more complex units in a network are largely unknown and worth exploring as a potential avenue to 1) more flexible and efficient artificial systems and 2) more biologically-realistic models of the brain. This project encourages students to take inspiration from neurophysiology and explore the consequences of simple modifications to artificial neural networks that make them more “brain-like.” Students should be familiar with dynamical systems modeling. Julia is the preferred programming language for this project.
Mentor: Jacob Huh
Project:
Learning separability of continuous representations
Neural networks have an implicit bias towards learning smooth representation over the data distribution.
That is, the output prediction does not deviate when there is a small and natural perturbation in the input data (small euclidean changes).
Yet, our world often requires us to make drastically different predictions based on small changes in the input representation (e.g., take, for example, homonyms, object grasping, medical imaging).
While it is possible to overcome these implicit biases by “learning from data,” this requires industry-scale data and compute, making it an infeasible and inefficient solution to modern-day research and our goals towards “intelligence.”
This project aims to study the neural network’s ability to separate continuous tensors and investigate different types of input transformations that have better empirical and theoretical guarantees for separation.
Lastly, we will investigate how such a method can be used for downstream applications such as medical imaging, reinforcement learning, and many more.
Mentor: Andrei Barbu
Projects:
1. Let’s try to make computational models of social interactions and build a mathematical foundation for social interactions; by this route we will also investigate compositionality and generalization in new kinds of neural network structures. This involves creating zero-shot models of social interactions: models that without any social-interaction-specific training data generate trajectories for different kinds of social interactions or recognize videos of social interactions. We have started going down this road by modifying models from robotics to recursively reason about social interactions. Next steps would be to: formalize new kinds of interactions, learn such models automatically, find good neural network approximations to such models (this relates more deeply to notions of compositionality and generalization in neural networks; how can we make specific networks that are good at understanding the rules of a domain and generalizing to that domain).
2. Large-scale neuroscience of language. Natural language processing is similar to neuroscience these days. We have models and we don’t really know how to work internally. Inspired by recent work in understanding Transformers in computational linguistics, we created a. large-scale dataset of intracranial (high temporal 2kHz and spatial resolution) neural recordings of subjects watching movies. Then, we annotated that dataset extensively by aligning it at the ms level, recording the part of speech of every word, parsing every sentence, recording the identity of every speaker, and annotating every cut and scene type. With the resulting dataset we could investigate how many linguistic concepts are processed by the brain, like, how and where word meaning is represented, are there specific areas or points in time associated with different kinds of compositional operations, how are vision and language integrated, etc.
Mentor: Mengjia Xu
Project: Learning stochastic graph representations for “discrete-time” dynamic graphs with “variable embedding size”
Effective dynamic graph representation learning methods have gained a lot of attention as they can empower diverse downstream inference and prediction tasks (link prediction, recommender system, causal discovery for fraud detection and credit decisions, etc.). However, uncertainty quantification is very important for characterizing the effective embedding dimensions in the latent space; moreover, learning graph embeddings with variable embedding size over different time stamps guided by the uncertainty quantification could generate more scalable and efficient graph embedding approaches and will improve the quality of the embeddings. This is currently an exciting open area of research.
Mentor: Gabe Margolis
Project:
Future in-home assistant robots should operate 24×7 in open-ended environments without supervision; in this setting, they will face the challenges of dynamically changing environments, autonomous battery charging and homing, and safe exploration around humans. The project will apply cutting-edge robot learning techniques to the challenges of long-term robot operation. Candidates should be competent with machine learning in Python. Familiarity with computer vision and robot localization techniques will be useful.
Mentor: Akshay Rangamani
Project:
Deep Networks as Learned Optimizers: Building on the observation that a single step of an optimization algorithm can be seen as a layer of a deep network, we would like to investigate the extent of their generalization capabilities, and how they are able to exhibit such fast convergence. This project will also characterize the efficacy of this approach on convex/non-convex problems
Implicit Neural Representations and their Properties: This is a project inspired by classical signal processing approaches to signal representation. We will learn implicit neural representations and investigate their applicability to transformations like blurring, cropping, super resolution, etc.
Recurrence and Robustness: Iterated associative memories are more robust to perturbation in the inputs than Single-shot associative memories. In this project we will apply this observation to characterize whether recurrent connections make deep networks more or less robust to adversarial noise in the inputs.
Mentor: Ali Hummos
Project:
Predictive autoencoder: Classical autoencoders notoriously learn entangled representations from data. The features in the latent embedding space (where they encode observations into) does not separate the separated causal features in the world that generated the data. For example, the location of an object and its shape could be represented along one dimension in a high dimensional embedding space as a single feature. How about use gradient descent updates on the embedding space to bias the model to use video data to discover the casual features the data. This project involves training an autoencoder on moving MNIST dataset and comparing efficiencies of different architectures.
Inferring the motives of others: in a multi-agent setting, neural network based models do not typically display “theory of mind” design. An agent does not attempt to build theories of what other agents might have in mind. More concretely, let’s build an agent that has an abstract representation of its goal and motivations, and allow it to use the same structure to infer the goals and motivations of others. What would be the benefits of the two abilities sharing the same neural implementation. The project involves building a minigrid RL environment and training RNNs on the task.
Mentor: Quilee Simeon (Yang lab)
Project: Caenorhabditis elegans (C. elegans) has long been a model organism in neuroscience since it is a small (1mm in length), transparent worm that exhibits a fairly rich repertoire of behaviors with only 302 neurons. We have long known the full wiring diagram (connectome) of this model organism. While we can model the impulse response of individual neurons, we are yet to have a holistic computational model that can predict the output of the whole system given a pattern of inputs. The “big picture” goal of this project is to simulate the input-output mapping of the c. elegans nervous system. To that end, we will employ network architectures – both traditional (RNNs) and emerging (GNNs) – constrained by the connectivity of C. elegans to simulate the activity of artificial neurons. We will compare and optimize our models against the activity of real neurons. Eventually, we hope to connect our simulations up to behavioral and neuromuscular models of the worm in order to simulate a virtual C. elegans.
There are several ways to contribute to this project. The following is a non-exhaustive list of examples:
1. Build a pipeline for loading C. elegans connectivity and activity data that has been open-sourced.
2. Learn about and build graph neural networks (GNNs) – an increasingly used neural network framework which generalizes convolution and transformers – that operate on graphs of the C. elegans nervous system which you will construct.
3. Simulate and analyze C. elegans neuron activity using approaches from both dynamical systems theory and data-driven machine learning. Compare the pros and cons of the two approaches.
4. Come up with self-supervised tasks to optimize neural networks that are best suited for prediction of the C. elegans nervous system’s dynamics.
Mentor: Yena Han
Project:
Deep neural networks are suggested as the best models of various areas in the brain, such as the visual cortex. For comparing and validating competing models, functional similarities between artificial neural networks and the brain are one of the key criteria. The idea is as we develop models that have better functional predicity scores, models with more biological relevance emerge. Critically, we question what we can actually learn or infer from these functional tests. For instance, can we differentiate distinct architectures? Or can we make connections to the key motif? We study the roles and limitations of using functional similarity metrics in various domains. We try to identify multiple factors affecting predictivity.