Andrew Lampinen

Interested in cognition and artificial intelligence. Research Scientist at DeepMind. Posts are mine.

2024-09-23

Pleased to share that our paper (sigmoid.social/@lampinen/11249) is now accepted at TMLR! The camera-ready version should be clearer and improved thanks to the helpful reviewer comments (and others). Thanks again to my co-authors @stephaniechan and @khermann
The TMLR version is here: openreview.net/forum?id=aY2nsg
and the arXiv version (arxiv.org/abs/2405.05847) should be updated to match shortly. Check it out if you're interested in interpretability, and its challenges!
#interpretability

2024-08-15

Really excited to share that I'm hiring for a Research Scientist position in our team! If you're interested in the kind of cognitively-oriented work we've been doing on learning & generalization, data properties, representations, LMs, or agents, please check it out!
boards.greenhouse.io/deepmind/
#research #jobs

2024-08-04

Pleased to share my paper "Can language models handle recursively nested grammatical structures? A case study on comparing models and humans" was accepted to Computational Linguistics! Early journal version here: direct.mit.edu/coli/article/do
Check it out if you're generally interested in assessing LMs against human capabilities, or in LM capabilities for processing center embedding specifically.

2024-07-17
2024-07-17

Pleased to share that our paper "Language models, like humans, show content effects on reasoning tasks" is now published in PNAS Nexus!
academic.oup.com/pnasnexus/art

#LanguageModels #lms #AI #cogsci #machinelearning #nlp #nlproc #cognitivescience

Andrew Lampinen boosted:
2024-06-06

@gmusser @thetransmitter @kristin_ozelli @KathaDobs @ev_fedorenko @lampinen @Neurograce @UlrikeHahn 👆I really recommend this. There is a fair bit of comment on this platform on how non-human-like artificial neural networks and #LLMs are, but most of it is not by cognitive scientists or neuroscientists. This article is an informative counterbalance.

Andrew Lampinen boosted:
George Mussergmusser
2024-06-06

At a detailed level, artificial neural networks look very different from natural brains. At a higher level, they are uncannily similar. My story for @thetransmitter, edited by @kristin_ozelli, features @KathaDobs @ev_fedorenko @lampinen @Neurograce and others. thetransmitter.org/neural-netw

2024-05-24

@cigitalgem thanks! Curious to hear any reactions you have

2024-05-23

This paper is really just us *finally* following up on a weird finding about RSA (figure on the here) from a paper Katherine Hermann & I had at NeurIPS back in the dark ages (2020): x.com/khermann_/status/1323353
Thanks to my coauthors @scychan_brains & Katherine! 9/9

2024-05-23

We also just find these results inherently interesting for what they suggest about the inductive biases of deep learning models + gradient descent. See the paper for lots of discussion of related work on (behavioral) simplicity biases and much more!

We’ve also released a colab which provides a very minimal demo of the basic easy-hard representation bias effect if you want to explore it for yourself: gist.github.com/lampinen-dm/b6
8/9

2024-05-23

Similarly, computational neuroscience compares models to brains using techniques like regression or RSA. But RSA between our models shows strong biases (the stripes); e.g. a model doing two complex tasks appears less similar to another model doing *exactly the same* tasks than it does to a model doing only simple tasks! 7/9

RSA plot showing strong biases (banding patterns), and in some cases stronger similarity between models doing *very different* tasks than models doing the same task.
2024-05-23

Why care about biases? They can hamper our ability to interpret what a system is “doing” from its representations; e.g. PCA, SAEs or other model simplifications will be biased towards particular aspects of the model’s computations. 6/9

PCA plot showing clear clusters by an easy feature, but the hard feature (represented as shapes) is not captured at all in the top PCs.
2024-05-23

We find similar biases in Transformers (on language tasks) and CNNs (on vision tasks), but also variability; e.g. biases can depend on architecture component (e.g. encoder vs. decoder) and many other properties. Feature prevalence and output position also bias representations! 5/9

Figure showing feature biases for token-level and structural features from transformers trained on Dyck language tasks; biases are qualitatively different in the encoder (biased by more structural features/those that cover more tokens) and decoder (biased towards simpler features).
2024-05-23

This effect seems to be driven by a combination of 1) the fact that the easy feature is learned first, and 2) the fact that there are more ways to represent a non-linear feature. The learning dynamics are quite interesting! 4/9

Learning dynamics for different training regimes: pretraining one of the features first, then the other, or training both simultaneously, and including variance explained if accounting for all ways the harder feature can be represented. Even if the hard feature is pretrained, it still carries less variance than the easier feature, although the gap is somewhat closed. However, pretraining + accounting for all ways the hard feature can be represented closes the gap entirely.
2024-05-23

For example, if we train a model to compute a simple, linear feature and a hard, highly non-linear one, the easy feature is naturally learned first, but both are generalized perfectly by the end of training. However, the easy feature dominates the representations! 3/9

Left: Learning curves showing test accuracy for both easy and hard features converges to 100%, but the hard feature is learned much more slowly.  Right: variance explained by each feature in the model’s penultimate layer representations — the easy feature has ~ 55% variance explained by the end of training, while the hard feature carries only ~5%. Intriguingly, the variance explained by the easy feature even *inflates* when the hard feature is learned, even though the easy feature performance has long been saturated.
2024-05-23

Our goal was to study the relationship between representation and computation in a setting where we know what a model is computing. To do so, we train MLPs, Transformers, and CNNs to compute multiple features from their inputs.

We manipulate data & feature properties, while preserving the features’ computational roles in the task. We then analyze the models’ internal representations, and find representations are strongly biased by these seemingly-irrelevant properties. 2/9

Overview figure: training models on data with varying properties, than analyzing how those properties are reflected in the model representations, and considering downstream implications for interpretabiltiy (e.g. that it may be biased towards simpler features a model computes). Bottom panel: the experiment domains: MLPs on binary features inputs and outputs, transformers on language-like inputs and outputs, and CNNs on vision inputs with classification outputs.
2024-05-23

How well can we understand an LLM by interpreting its representations? What can we learn by comparing brain and model representations? Our new paper (arxiv.org/abs/2405.05847) highlights intriguing biases in learned feature representations that make interpreting them more challenging! 1/9
#intrepretability #deeplearning #representation #transformers

2024-03-13

And thanks to all of the wonderful SIMA team who made this project take off! 5/5

2024-03-13

However, our agent is still often unreliable, and doesn’t match human instruction-following performance yet (though some of our tasks are hard, and even human game players are far from perfect). SIMA is very much a work in progress; we’re looking forward to sharing more soon. But I’m excited about the progress we’ve already made towards scaling language grounding in video games (videos in this thread are from the agent). Check out our blog post and tech report for more! 4/5

2024-03-13

We built our Scalable Instructable Multiworld Agent (SIMA) around the representations of internet-pretrained models, including a video model, and trained them primarily via instruction-conditioned BC, augmented with classifier-free guidance. We found that our agent learns many diverse skills, and shows positive transfer among environments (outperforming single-game-specialized agents), and even some zero-shot transfer of basic skills like navigation to held-out environments! 3/

Performance across environments: SIMA achieves ~160% the performance of environment expert agents overall; and even achieves around 90% performance zero-shot on a domain it has never trained on. An ablation without pretrained components do not perform as well as SIMA, and an agent without language performs very poorly.

Client Info

Server: https://mastodon.social
Version: 2025.04
Repository: https://github.com/cyevgeniy/lmst