Andrew Lampinen Profile Banner
Andrew Lampinen Profile
Andrew Lampinen

@AndrewLampinen

6,594
Followers
1,397
Following
116
Media
1,993
Statuses

Interested in cognition and artificial intelligence. Research Scientist @DeepMind . Previously cognitive science @StanfordPsych . Tweets are mine.

Joined November 2019
Don't wanna be here? Send us removal request.
Pinned Tweet
@AndrewLampinen
Andrew Lampinen
2 months
Excited to share one of the main things I’ve been working on: scaling towards grounded language agents that can follow natural language instructions across many video game environments—using the same pixels to keyboard/mouse controls as humans do. 1/8
12
61
325
@AndrewLampinen
Andrew Lampinen
1 year
Ted Chiang is a great writer, but this is not a great take and I'm disappointed to see it getting heavily praised. It's not in keeping with our scientific understanding of LMs or deep learning more generally. Thread: 1/n
@jburnmurdoch
John Burn-Murdoch
1 year
Ted Chiang’s piece on ChatGPT and large language models is as good as everyone says. The fact that the outputs are rephrasings rather than direct quotes makes them seem game-changingly smart — even sentient — but they’re just very straightforwardly not.
Tweet media one
92
532
3K
27
202
1K
@AndrewLampinen
Andrew Lampinen
4 years
Officially done with my PhD! My dissertation is now online (), if you fancy reading way too many pages about flexibility and transfer in humans and deep learning models.
21
56
567
@AndrewLampinen
Andrew Lampinen
1 year
What is emergence, and why is it of recent interest in AI, and long-standing interest in cognitive science? And why is this an exciting time for considering emergence across these fields? A thread: 1/
14
93
471
@AndrewLampinen
Andrew Lampinen
11 months
Computational physicists doing most their operations (e.g. derivatives) via an FFT because it's more efficient
Tweet media one
@robertskmiles
Rob Miles (✈️ Tokyo)
11 months
Their results are bizarre and inhuman. @NeelNanda5 trained a tiny transformer to do addition, then spent weeks figuring out what it was doing - one of the only times in history someone has understood how a transformer works. This is the algorithm it created. To *add two numbers*!
Tweet media one
71
262
2K
6
38
464
@AndrewLampinen
Andrew Lampinen
1 year
What can be learned about causality and experimentation from passive data? What could language models learn from simply passively imitating text? We explore these questions in our new paper: “Passive learning of active causal strategies in agents and language models” Thread: 1/
Tweet media one
11
75
403
@AndrewLampinen
Andrew Lampinen
2 years
I'm not a scaling maximalist, but it's surprising to me how many people are 1) interested in differences between human and artificial intelligence and 2) think scaling to improve performance means deep learning is doing something fundamentally wrong. 1/n
15
64
394
@AndrewLampinen
Andrew Lampinen
5 months
Research in mechanistic interpretability and neuroscience often relies on interpreting internal representations to understand systems, or manipulating representations to improve models. I gave a talk at @unireps at NeurIPS on a few challenges for this area, summary thread: 1/
Tweet media one
3
55
392
@AndrewLampinen
Andrew Lampinen
2 years
Abstract reasoning is ideally independent of content. Language models do not achieve this standard, but neither do humans. In a new paper ( co-led by Ishita Dasgupta) we show that LMs in fact mirror classic human patterns of content effects on reasoning. 1/
7
56
381
@AndrewLampinen
Andrew Lampinen
4 years
I'm excited to announce that next month I'll be joining @DeepMind as a research scientist! Looking forward to continuing to work with @FelixHill84 and others!
11
7
356
@AndrewLampinen
Andrew Lampinen
3 years
How can RL agents recall the past in detail, in order to behave appropriately in the present? In our new preprint "Towards mental time travel: A hierarchical memory for RL agents" () we propose a memory architecture that steps in this direction.
3
69
340
@AndrewLampinen
Andrew Lampinen
3 years
What are symbols? Where do symbols come from? What behaviors demonstrate the ability to engage with symbols? How do the answers to these questions impact AI research? We argue for a new perspective on these issues our preprint: Summary in thread: 1/6
7
70
333
@AndrewLampinen
Andrew Lampinen
4 years
How can deep learning models flexibly reuse their knowledge? How can they adapt to new tasks zero-shot, as humans can? In our new preprint (), we propose a new approach based on learning to transform task representations: meta-mapping. Preview in thread:
3
65
296
@AndrewLampinen
Andrew Lampinen
6 months
Very excited to share a substantial updated version of our preprint “Language models show human-like content effects on reasoning tasks!” TL;DR: LMs and humans show strikingly similar patterns in how the content of a logic problem affects their answers. Thread: 1/
Tweet media one
2
42
253
@AndrewLampinen
Andrew Lampinen
6 months
Exactly. Nobody (serious) thinks transformers can magically infer things totally unsupported by training. However, they can generalize, not just memorize, things that are supported. The key question is what, is/isn't supported by training on Internet-scale distributions? 1/
@QuanquanGu
Quanquan Gu
6 months
I believe the message in the paper is straightforward and uncontroversial. However, it seems there might be a misunderstanding in @abacaj 's interpretation. Pre-trained transformers can effectively acquire in-context knowledge for tasks related to their pre-training data and…
6
34
219
7
16
232
@AndrewLampinen
Andrew Lampinen
2 years
How should we compare the capabilities of language models and humans? Is the answer different for LMs than cognitive models? In I offer some thoughts, focusing on a case study of LM processing of recursively nested grammatical structures. Thread: 1/11
6
56
227
@AndrewLampinen
Andrew Lampinen
2 years
Explanations play a critical role in human learning, particularly in challenging areas—abstractions, relations and causality. We show they can also help RL agents in "Tell me why!—Explanations support learning of relational and causal structure" (). Thread:
3
38
214
@AndrewLampinen
Andrew Lampinen
9 months
The debate over AI capabilities often hinges on testing abilities that humans are presumed to have — reasoning, logic, systematic, compositional generalization, grammar, etc. But how reliably good are humans actually at these skills? 1/
11
45
206
@AndrewLampinen
Andrew Lampinen
2 years
Interested in how RL agents could recall the past in detail, in order to overcome the challenges of the present? Come chat with us about "Towards mental time travel: A hierarchical memory for RL agents" at #NeurIPS2021 poster session 1 (4:30 GMT/8:30 PT, spot E1)!
Tweet media one
4
31
195
@AndrewLampinen
Andrew Lampinen
6 months
What is representational alignment? How can we use it to study or improve intelligent systems? What challenges might we face? In a new paper, we describe a framework that attempts to unify ideas from cognitive science, neuroscience and AI to address these questions. 1/3
Tweet media one
3
28
187
@AndrewLampinen
Andrew Lampinen
11 months
Why does asking language models to respond as an expert or think step-by-step improve answers? What does it have to do with conditional BC and role play? Thread on the power of conditional sequence modeling, and why evaluating LM capabilities or safely deploying them is hard: 1/
9
36
180
@AndrewLampinen
Andrew Lampinen
1 year
The recent discussions of what language models can and can't accomplish highlight some important issues in how cognitive science, linguistics, etc. think about human capabilities or competencies, and thus how to test them in models. Thread: 1/
1
45
178
@AndrewLampinen
Andrew Lampinen
2 years
It's often claimed that learning language alone can't lead to understanding because understanding requires relating language to external meaning. I'm all for grounding and social learning of language, but I think that argument is wrong, or at least uninteresting. 1/n
14
30
174
@AndrewLampinen
Andrew Lampinen
1 year
Chomsky favoring thought experiments based on his mental model of a system over actual empirical results? What a shocker
@jayelmnop
Jesse Mu
1 year
PSA to anyone who wants to write an op-ed criticizing LLMs (yes, including Noam Chomsky): if you're going to come up with hypothetical failure cases for LLMs, at a minimum, please actually check that your case fails with a modern LLM
Tweet media one
Tweet media two
Tweet media three
31
87
870
6
20
173
@AndrewLampinen
Andrew Lampinen
1 year
When we and others study the performance of LM-like models under controlled situations, where we know exactly what is trained, we find that these models can learn generalizable strategies that perform well on truly novel test examples rather than just memorizing + rephrasing. 3/
8
15
150
@AndrewLampinen
Andrew Lampinen
1 year
I think this paper does a good job of highlighting the point that apparently discontinuous transitions in performance may correspond to more continuous change in a smoother metric (also made in ). However, I have a terminological quibble. 1/5
@arankomatsuzaki
Aran Komatsuzaki
1 year
Are Emergent Abilities of Large Language Models a Mirage? Presents an alternative explanation for emergent abilities: one can choose a metric which leads to the inference of an emergent ability or another metric which does not.
Tweet media one
24
187
978
3
26
148
@AndrewLampinen
Andrew Lampinen
7 months
Nice, though not too shocking, to see that human inductive biases on these sequential tasks can be meta-learned. I particularly appreciate the emphasis on the ability of the transformers to learn subtler biases in the human behavior. 1/4
@LakeBrenden
Brenden Lake
7 months
Today in Nature, we show how a standard neural net, optimized for compositional skills, can mimic human systematic generalization (SG) in a head-to-head comparison. This is the capstone of a 5 year effort with Marco Baroni to make progress on SG. (1/8)
Tweet media one
24
395
2K
3
13
146
@AndrewLampinen
Andrew Lampinen
3 years
Excited that our work on "Transforming task representations to perform novel tasks" is now published online @PNASNews ! We provide a new perspective on flexible adaptation, which we hope will be inspiring for both AI and cognitive science. Summary thread:
2
24
128
@AndrewLampinen
Andrew Lampinen
3 years
Honored to share that I have been awarded one of the CogSci 2021 Glushko Prizes for my dissertation on learning to transform task representations to adapt to novel tasks! And I'm especially honored to share it with these incredible other scholars.
@cogsci_soc
CogSci Society
3 years
The #CogSci2021 Glushko Dissertation Prize Winners have been announced! Congratulations to all - very well deserved. Visit the website for further info on our winners:
Tweet media one
2
8
76
16
3
117
@AndrewLampinen
Andrew Lampinen
2 years
I admire the authors, and I find their work interesting. However, I think this piece makes subtle equivocations between strong and weak notions of compositionality that may be problematic (and are common in this space). Thread:
@RTomMcCoy
Tom McCoy
2 years
🤖🧠NEW PAPER🧠🤖 What explains the dramatic recent progress in AI? The standard answer is scale (more data & compute). But this misses a crucial factor: a new type of computation. Shorter opinion piece: Longer tutorial: 1/5
Tweet media one
8
112
601
5
15
113
@AndrewLampinen
Andrew Lampinen
1 year
This is not to imply that LMs never do (lossy) memorization, of course; that definitely happens in many instances, is important to study, and may underlie their success in some cases. But saying that is all they do seems out of keeping with the state of the science. 7/7
4
2
112
@AndrewLampinen
Andrew Lampinen
1 year
In case you missed our paper "Can language models learn from explanations in context?" and want a 7.5 minute video overview of the experiments and main findings, I've now put one on YouTube!
4
19
110
@AndrewLampinen
Andrew Lampinen
1 year
But it's also a general finding that neural networks generalize better from more diverse learning experiences. E.g. the open-ended learning work () or our work on how richer environments improve generalization (). 5/
2
5
93
@AndrewLampinen
Andrew Lampinen
2 years
Why? Because evolution made some pretty hefty tradeoffs in energy cost, soft/fractured skulls at birth, etc. in order to scale the human brain. As a consequence, we have perhaps ~1 quadrillion synapses (). 2/
1
2
90
@AndrewLampinen
Andrew Lampinen
7 months
Definitely out here attacking anyone who believes in probabilities less than zero, ngl
@GaryMarcus
Gary Marcus
7 months
It’s almost impossible to be in the middle on p(doom). If you think it is greater than zero, you get attacked, if you think it is less than zero, you get it attacked. If you think we don’t know, you get attacked. For the record, I think we don’t know.
50
6
140
2
1
93
@AndrewLampinen
Andrew Lampinen
1 year
So if anything, we should expect LMs with internet training to generalize better than models trained on toy tasks. And indeed, LMs perform well at BIG-Bench () — tasks researchers chose specifically to be hard for a model that memorized the internet. 6/
2
4
93
@AndrewLampinen
Andrew Lampinen
3 months
This is a valuable paper for helping to understand how general the inductive biases that give rise to common language structures may be. But furthermore, I think the prior claim that these languages are "impossible" for humans is weak at best. Thread:
@JulieKallini
Julie Kallini ✨
4 months
Do LLMs learn impossible languages (that humans wouldn’t be able to acquire) just as well as they learn possible human languages? We find evidence that they don’t! Check out our new paper… 💥 Mission: Impossible Language Models 💥 ArXiv: 🧵
Tweet media one
12
114
478
2
9
90
@AndrewLampinen
Andrew Lampinen
2 years
If you're curious about my thoughts on language models and reasoning (and more), I'll be speaking at the London Machine Learning Meetup on November 16th! (It's virtual, so you don't need to be in London to attend.)
3
12
87
@AndrewLampinen
Andrew Lampinen
8 months
Disappointed, but not all that surprised to see these kinds of opinions going around. A few years ago, we wrote a paper called "Publishing fast and slow: a path towards generalizability in psychology and AI" that seems relevant again. Short thread:
@shortstein
Thomas Steinke
8 months
I'm shocked that some people hold such backward views and saddened that this is coming from a person with significant power and influence in the research community. I ❤️ @arXiv
Tweet media one
64
71
980
4
5
90
@AndrewLampinen
Andrew Lampinen
1 year
One important approach to the scientific study of complex phenomena like human intelligence or the behavior of language models is to create a simplified model which captures the key elements, while maintaining full control over the system, and study its behavior. 2/
3
3
88
@AndrewLampinen
Andrew Lampinen
2 years
We're not going to be able to run brain scale models anytime soon, and I think innovations in architecture, algorithms and training will contribute to advances in AI. But I think dismissing scale may lead to fundamentally misunderstanding the basis of intelligence. 9/9
5
5
85
@AndrewLampinen
Andrew Lampinen
3 years
I'm pleased to share that we've open-sourced two environments and the hierarchical attention mechanism for our "Towards mental time travel: A hierarchical memory for RL" paper: Paper summary thread/links:
@AndrewLampinen
Andrew Lampinen
3 years
How can RL agents recall the past in detail, in order to behave appropriately in the present? In our new preprint "Towards mental time travel: A hierarchical memory for RL agents" () we propose a memory architecture that steps in this direction.
3
69
340
1
9
86
@AndrewLampinen
Andrew Lampinen
5 months
Really cool results! It makes sense in light of the fact that LMs have to model lots of bad reasoning and inaccurate info that stripping out some of what they've learned can help. Pretty surprising that just targeted rank reduction works though! I'm curious though, 1/2
@pratyusha_PS
Pratyusha Sharma @ ICLR
5 months
What if I told you that you can simultaneously enhance an LLM's task performance and reduce its size with no additional training? We find selective low-rank reduction of matrices in a transformer can improve its performance on language understanding tasks, at times by 30% pts!🧵
Tweet media one
37
281
2K
1
13
84
@AndrewLampinen
Andrew Lampinen
3 years
Super excited to announce the lineup of panelists and discussion moderators for our Neural Network Models of Cognition Affinity Group at CogSci this year! An awesome collection of legends in the field and rising stars.
1
13
85
@AndrewLampinen
Andrew Lampinen
2 years
So our largest models are *nowhere near* the scale/complexity of the human brain, which the evolutionary evidence suggests may be important to our intelligence. Sometimes more is different (), and you get emergent phenomena (cf. ) 4/
2
2
82
@AndrewLampinen
Andrew Lampinen
4 years
Interested in the relationship between cognition and AI or neural networks? I have a new @cogsci_soc blog post describing some factors that allow neural networks to be more flexible, and how those factors relate to human cognition and development:
1
20
83
@AndrewLampinen
Andrew Lampinen
1 year
If you're interested in recent discussions about what language models can and can't do, you may be interested in this paper on the challenges of comparing LM and human capabilities! I've recently updated it with some minor edits and some new experiments.
@AndrewLampinen
Andrew Lampinen
2 years
How should we compare the capabilities of language models and humans? Is the answer different for LMs than cognitive models? In I offer some thoughts, focusing on a case study of LM processing of recursively nested grammatical structures. Thread: 1/11
6
56
227
1
15
81
@AndrewLampinen
Andrew Lampinen
1 year
Will be at NeurIPS! Let me know if you want to chat about explanations, reasoning, language models, or language-augmented RL!
8
3
76
@AndrewLampinen
Andrew Lampinen
2 months
I like the empirical results, but have some questions about the way they're being discussed. In particular, what is reasoning? By this bar, humans also don't learn to reason, because they tend to be less accurate in out of distribution situations as well. 1/7
@sirbayes
Kevin Patrick Murphy
2 months
@fchollet claimed that "learning to reason" (robustly across problem instances) is hard because of the nature of the MLE objective. This paper proves that claim. The reason is "statistical features inherently exist in data distributions, but can hinder model generalization…
9
72
391
1
10
75
@AndrewLampinen
Andrew Lampinen
5 months
Very excited to head to NeurIPS! Feel free to reach out if you want to chat about any of our recent work on LMs, agents, interpretability, representational alignment, etc. You can find me:
2
0
72
@AndrewLampinen
Andrew Lampinen
2 years
That's many orders of magnitude more than the largest models I'm aware of. And synaptic interactions are likely far more complex than a single parameter can capture (e.g. ). And this all neglects the increasing evidence for glial contributions etc. 3/
2
1
69
@AndrewLampinen
Andrew Lampinen
1 year
See also this thoughtful thread from @raphaelmilliere !
@raphaelmilliere
Raphaël Millière
1 year
I don't think lossy compression is a very helpful analogy to convey what (linguistic or multimodal) generative models do – at least if "blurry JPEGs" is the leading metaphor. It might work in a loose sense, but it doesn't tell the whole story. 1/
22
71
410
3
1
68
@AndrewLampinen
Andrew Lampinen
7 months
Some great analysis and thoughtful comments here! However, are these really embers of *autoregression*? It seems to me that they are quite general phenomena, that are also observed in humans (as the paper notes), CNNs, RL agents, ... any system that learns a prior from data 1/
@RTomMcCoy
Tom McCoy
8 months
🤖🧠NEW PAPER🧠🤖 Language models are so broadly useful that it's easy to forget what they are: next-word prediction systems Remembering this fact reveals surprising behavioral patterns: 🔥Embers of Autoregression🔥 (counterpart to "Sparks of AGI") 1/8
Tweet media one
36
299
1K
6
1
69
@AndrewLampinen
Andrew Lampinen
2 years
And as the scaling laws paper showed, larger models generalize *better* from the same amount of data (). Larger models *can* use more data, but they can also make better use of less! 5/9
1
3
67
@AndrewLampinen
Andrew Lampinen
1 year
And though some particularly challenging BIG-Bench tasks are harder, models can perform better on many of them if given time to produce some reasoning steps before their answer (). Memorization + rephrasing is not a good explanation of this. 7/
2
2
67
@AndrewLampinen
Andrew Lampinen
1 year
Always fun chatting with @MLStreetTalk ! Check this out for my current thoughts on language, grounding, NeurIPS, and more!
@MLStreetTalk
Machine Learning Street Talk
1 year
First interview from #NeurIPS2022 uploaded! This is a conversation with @AndrewLampinen about NLP, symbols, grounding and even Chomsky! We recorded a LOT of footage this week, keep an eye out!
Tweet media one
4
21
108
1
5
67
@AndrewLampinen
Andrew Lampinen
3 months
Since there continues to be some confusion about the term "passive," let me unpack this in more detail. Part of the motivation for the use of the term here was to draw attention to two dimensions that get conflated in discussions of langauge models, but that should not be. 1/
@AndrewLampinen
Andrew Lampinen
1 year
What can be learned about causality and experimentation from passive data? What could language models learn from simply passively imitating text? We explore these questions in our new paper: “Passive learning of active causal strategies in agents and language models” Thread: 1/
Tweet media one
11
75
403
1
7
67
@AndrewLampinen
Andrew Lampinen
1 year
As someone who is both interested in the successes of language models, and also in the contributions of rich learning environments to generalization, I've been hoping to see some results like this for a while! The evaluations are relatively simple, but it's very suggestive.
@szxiangjn
Jiannan Xiang
1 year
🤔Humans can learn from embodied experiences in the physical world. Can Language Models also do that? 🔥Check out our new paper about enhancing Language Models with World Models! 👇 1/n
Tweet media one
5
67
356
1
7
66
@AndrewLampinen
Andrew Lampinen
2 years
Pleased to share that our paper "Can language models learn from explanations in context?" () was accepted to Findings of EMNLP 2022! Camera ready version coming soon, with some better control experiments for the tuning conditions.
@AndrewLampinen
Andrew Lampinen
2 years
Can language models learn from explanations of answers in a few-shot prompt? In our new preprint (), we explore this question! Thread:
3
45
191
2
7
65
@AndrewLampinen
Andrew Lampinen
2 years
@NicoleCRust Something about NNs brings nasty responses out! A few excerpts from an email I got from a full professor (at a reasonably prominent institution) who didn't like my answer to their question in a talk I gave about my dissertation:
Tweet media one
Tweet media two
Tweet media three
Tweet media four
19
5
65
@AndrewLampinen
Andrew Lampinen
21 days
Our SIMA tech report is now on arXiv!
@AndrewLampinen
Andrew Lampinen
2 months
Excited to share one of the main things I’ve been working on: scaling towards grounded language agents that can follow natural language instructions across many video game environments—using the same pixels to keyboard/mouse controls as humans do. 1/8
12
61
325
1
9
64
@AndrewLampinen
Andrew Lampinen
3 years
I've often thought that some of what people consider a separate "system 1 vs system 2" might be well captured with a single architecture that just decides how long to think. Glad to see some progress in that direction, and excited for further work in this area!
@GoogleDeepMind
Google DeepMind
3 years
Introducing PonderNet, a new algorithm that allows artificial neural networks to learn to “think for a while” before answering, like humans do. This improves the ability of neural networks to generalize outside of the training distribution:
Tweet media one
24
297
1K
2
4
62
@AndrewLampinen
Andrew Lampinen
2 years
In our paper "Symbolic Behaviour in AI" () we discuss a perspective on symbols, that we suggest 1) better describes human capacities 2) provides a more promising direction for AI. Briefly, we focus on symbols as subjective, following C.S. Peirce.
@GaryMarcus
Gary Marcus
2 years
Deep Learning Is Hitting a Wall. What would it take for artificial intelligence to make real progress? #longread in ⁦ @NautilusMag ⁩ on one of the key technical questions in AI.
62
248
843
2
10
62
@AndrewLampinen
Andrew Lampinen
9 months
The most thorough and compelling paper yet in the line of work suggesting that there is signal in even minor adversarial perturbations, that humans are somewhat sensitive to; but with time they can override that signal (but still detect it if asked).
@gamaleldinfe
Gamaleldin Elsayed
9 months
Nature Comms paper: Subtle adversarial image manipulations influence both human and machine perception! We show that adversarial attacks against computer vision models also transfer (weakly) to humans, even when the attack magnitude is small.
12
87
389
2
7
62
@AndrewLampinen
Andrew Lampinen
3 years
Our RL agent memory was accepted at #NeurIPS2021 ! Thanks to the reviewers and AC for thoughtful reviews. We'll have a revised version out soon!
@AndrewLampinen
Andrew Lampinen
3 years
How can RL agents recall the past in detail, in order to behave appropriately in the present? In our new preprint "Towards mental time travel: A hierarchical memory for RL agents" () we propose a memory architecture that steps in this direction.
3
69
340
3
1
61
@AndrewLampinen
Andrew Lampinen
1 year
Emergence is the idea that a large system composed of many small parts can have fundamentally different properties than those parts do — or that“more is different” as Anderson described it (). 2/
2
4
60
@AndrewLampinen
Andrew Lampinen
8 months
Got pretty lucky this year with 3/3 papers I was involved in accepted to NeurIPS! Our work on passive learning of causal strategies:
@AndrewLampinen
Andrew Lampinen
1 year
What can be learned about causality and experimentation from passive data? What could language models learn from simply passively imitating text? We explore these questions in our new paper: “Passive learning of active causal strategies in agents and language models” Thread: 1/
Tweet media one
11
75
403
2
0
59
@AndrewLampinen
Andrew Lampinen
2 years
But again, I don't think scale is everything; architecture, algorithms and learning experiences obviously matter. E.g. transformers scale much better than LSTMs and may also have better inductive biases for certain interesting emergent behaviors () 6/9
2
1
57
@AndrewLampinen
Andrew Lampinen
2 years
I think this is a really exciting time for cognitive science, as @davisblalock highlights from our paper—for the first time we have various models that can exhibit many of the complex behaviors humans do, on naturalistic stimuli like those that are used to test humans. 1/3
@davisblalock
Davis Blalock
2 years
"Language models show human-like content effects on reasoning" So the specific results are interesting and actionable, but my main takeaway from this paper is that cognitive science is becoming adjacent to AI in a way it hasn't been since the 50s or 60s. [1/6]
Tweet media one
6
65
430
1
5
57
@AndrewLampinen
Andrew Lampinen
8 months
This is an interesting case of generalization from LMs, where they clearly cannot just "parrot" training data, but rather use explicit statements in tuning to non-trivially alter their behavior in a related test context. However, I find the overall framing misleading. 1/
@OwainEvans_UK
Owain Evans
8 months
Our experiment: 1. Finetune an LLM on descriptions of fictional chatbots but with no example transcripts (i.e. only declarative facts). 2. At test time, see if the LLM can behave like the chatbots zero-shot. Can the LLM go from declarative → procedural info?
Tweet media one
1
3
51
3
10
56
@AndrewLampinen
Andrew Lampinen
1 year
Cool work! Some more nice evidence that you shouldn't just *assume* what humans would do when evaluating models, you should actually compare. Really like the conclusion "future research should not idealize human behaviors as a monolith"
@Brown_NLP
Brown NLP
1 year
Last year, we criticized LMs for performing “too well” with pathological prompts, and many papers have now shown similar results with corrupted ICL or CoT. In our new work, we find that *humans* also perform surprisingly well with irrelevant prompts! (But not misleading ones.) ⅕
Tweet media one
Tweet media two
2
24
137
3
7
56
@AndrewLampinen
Andrew Lampinen
2 years
If you're interested in my thoughts on cognition and AI, or how my research interests have evolved (or if you just need something to listen to while hiding from omicron), this may be the podcast for you!
@thesisreview
The Thesis Review Podcast
2 years
Episode 38 of The Thesis Review: Andrew Lampinen ( @AndrewLampinen ), "A Computational Framework for Learning and Transforming Task Representations" We discuss learning to rapidly adapt to new tasks, cognitive flexibility in minds & machines, and more
Tweet media one
2
5
35
1
4
55
@AndrewLampinen
Andrew Lampinen
1 year
There's a trend recently of arguing that we shouldn't describe language models (and other AI systems) with words like "understand" or "see" that anthropomorphize them. A thread on benefits + costs of metaphor in language and thought, and why I'm ambivalent about these arguments:
6
11
55
@AndrewLampinen
Andrew Lampinen
2 years
What does it take to make symbols useful? We think it's embracing their status as subjective, constructed tools. If you want to find out more you can check out this fun conversation with @MLStreetTalk about our paper on Symbolic Behaviour! Thanks for having me on!
@MLStreetTalk
Machine Learning Street Talk
2 years
We discuss @AndrewLampinen from @DeepMind paper on symbolic behaviour in artificial intelligence with @ecsquendor @DoctorDuggar
Tweet media one
3
5
23
2
4
52
@AndrewLampinen
Andrew Lampinen
9 months
This claim seems, ironically, to be over-hyped. At a glance, it appears to rely on some very strong assumptions, including a crucial one that is unstated as far as I can see. 1/
@o_guest
Olivia Guest · Ολίβια Γκεστ
9 months
My fav part of this paper — we're not able to stop climate change directly, but we can rain a bit on this silly AI summer: "as we formally prove herein, creating systems with human(-like or -level) cognition is intrinsically computationally intractable."
1
23
61
4
5
53
@AndrewLampinen
Andrew Lampinen
3 years
New commentary on "The Generalizability Crisis" by @talyarkoni : "Publishing fast and slow: A path toward generalizability in psychology and AI." We argue that these fields share similar generalizability challenges, and could learn from each other.
2
11
51
@AndrewLampinen
Andrew Lampinen
1 year
This will indeed probably be a point of confusion. But also the Chinchilla paper only considers training compute in their optimality calculations afaiu; considering inference cost/speed would lead to different trade-offs, and presumably more "overtrained" models.
@BlancheMinerva
Stella Biderman
1 year
@MetaAI @GuillaumeLample Another thing I anticipate being a massive source of confusion: Their smaller models are massively overtrained. The fact that their 13B model meets or exceeds GPT-3 in performance is NOT contradictory to anything in the Chinchilla paper, because it's not compute-optimally trained
5
9
94
1
2
49
@AndrewLampinen
Andrew Lampinen
4 years
We've updated our preprint on zero-shot adaptation by transforming task representations! Now including more carefully controlled experiments, and analyses of how task space transforms under various meta-mappings. (See attached pic!)
Tweet media one
1
5
52
@AndrewLampinen
Andrew Lampinen
1 year
Some reflections on our Symbolic Behaviour in AI paper, after two more years of rapid progress in the field:
2
7
52
@AndrewLampinen
Andrew Lampinen
7 months
@yudapearl @EliSennesh @ylecun @tdietterich @GaryMarcus @geoffreyhinton Addressed in our recent paper, now accepted at NeurIPS (updated version coming soon):
@AndrewLampinen
Andrew Lampinen
1 year
What can be learned about causality and experimentation from passive data? What could language models learn from simply passively imitating text? We explore these questions in our new paper: “Passive learning of active causal strategies in agents and language models” Thread: 1/
Tweet media one
11
75
403
2
4
47
@AndrewLampinen
Andrew Lampinen
1 year
Dataset contamination is an increasing issue. Was also really disappointed to see BIG-Bench eval omitted due to contamination — there are lots of interesting tasks there.
@cHHillee
Horace He
1 year
I suspect GPT-4's performance is influenced by data contamination, at least on Codeforces. Of the easiest problems on Codeforces, it solved 10/10 pre-2021 problems and 0/10 recent problems. This strongly points to contamination. 1/4
Tweet media one
Tweet media two
81
696
4K
2
1
48
@AndrewLampinen
Andrew Lampinen
4 years
Interested in automated curricula for RL? Check out our ICLR poster in sessions 3 & 4 tomorrow. Our method works for goal-conditioned agents in complex environments. Paper (open access): Poster: #ICLR2020 #ReinforcementLearning
1
10
45
@AndrewLampinen
Andrew Lampinen
11 months
Interesting to contrast with our recent work studying causal strategies () — a key difference is that rather than inference from correlations, we focus on the fact that LMs can *intervene* at test time, e.g. via a tool/API, which can unlock generalization
@HochreiterSepp
Sepp Hochreiter
11 months
ArXiv : Benchmark dataset to test causal inference of large language models (LLMs). A key shortcoming of LLMs is causal inference. LLMs achieve close to random performance on the causal tasks. Causal inference works only for already known tasks.
2
59
260
2
5
44
@AndrewLampinen
Andrew Lampinen
1 month
Really excited for this interview to come out! Check it out if you're interested in LMs, causality, and agency! Also recorded just before I left London, so a bit of a throwback for me to see :)
@AleksanderMolak
Aleksander Molak (CausalPython.io)
1 month
Causal Inference, Agents and LLMs Join Andrew Lampinen and me for a conversation on... 1/n #CausalBanditsPodcast #causalinference #causaltwitter @machinelearning
1
3
12
1
2
44
@AndrewLampinen
Andrew Lampinen
1 year
Great demonstration of an important point that can't be repeated enough — it's hard to interpret LM evaluations if you don't know whether examples appeared in the training data.
@osainz59
Oscar Sainz @ICLR 2024
1 year
As an example, for popular datasets like CoNLL03, ChatGPT is capable of generating the training, validation, and even test splits. It turns out that ChatGPT has been evaluated as a zero-shot or few-shot NER system on this dataset by multiple papers. 🧵2/5
Tweet media one
3
14
56
3
3
44
@AndrewLampinen
Andrew Lampinen
2 years
And as my colleagues and I have explored in lots of prior work, experiences like explanations () or richer, more embodied environments () can fundamentally change learning & generalization. 7/9
1
1
43
@AndrewLampinen
Andrew Lampinen
11 months
If you're interested in what LMs or agents can learn about causal strategies from passive data, a recording of a talk I gave at @ic_arl a few weeks ago is out now!
1
7
43
@AndrewLampinen
Andrew Lampinen
1 year
It’s not surprising that recent developments in AI are eliciting a range of skeptical and fearful reactions. This almost always happens when technology develops. Thread: 1/
4
3
42
@AndrewLampinen
Andrew Lampinen
11 months
Very cool paper! A bit reminiscient of our work () where we likewise showed that learning to output explanations (~=thoughts) could improve learning and OOD generalization; but letting the agent observe its thoughts is a key idea we didn't get to!
@jeffclune
Jeff Clune
11 months
Introducing Thought Cloning: AI agents learn to *think* & act like humans by imitating the thoughts & actions of humans thinking out loud while acting, enhancing performance, efficiency, generalization, AI Safety & Interpretability. Led by @shengranhu 1/5
14
145
604
1
6
42
@AndrewLampinen
Andrew Lampinen
1 year
Form-meaning arguments are trending again, guess it's time to reboost this (), and also remind everyone that distributional semantics existed long before current LLMs (e.g. "Distributional semantics in linguistic and cognitive research", Lenci, 2008).
@AndrewLampinen
Andrew Lampinen
2 years
It's often claimed that learning language alone can't lead to understanding because understanding requires relating language to external meaning. I'm all for grounding and social learning of language, but I think that argument is wrong, or at least uninteresting. 1/n
14
30
174
4
8
42
@AndrewLampinen
Andrew Lampinen
1 year
Will shortly be presenting our paper on how data distributions give rise to few-shot learning in transformers! Come by poster 524 this afternoon, and/or check out @scychan_brains oral presentation next week!
@scychan_brains
Stephanie Chan
2 years
Intriguingly, transformers can achieve few-shot learning (FSL) without being explicitly trained for it. Very excited to share our new work, showing that FSL emerges in transformers only when the training data is distributed in particular ways! 🧵👇
Tweet media one
14
190
1K
0
6
42
@AndrewLampinen
Andrew Lampinen
4 years
A paper supporting our observation () that egocentric perspective improves generalization in RL!
@hardmaru
hardmaru
4 years
Rotation, Translation, and Cropping for Zero-Shot Generalization Makes a lot of sense. Try playing Doom not from an agent-centric perspective! I think agent-centric view is a better prior for encoding useful information using fewer bits for the policy.
Tweet media one
2
20
73
0
6
41
@AndrewLampinen
Andrew Lampinen
10 months
Lots of interesting results! Most importantly, data dominates the match to neural data much more than architecture (*as long as you use a decent architecture)!
@talia_konkle
talia konkle
10 months
5/ Surprisingly, across both measures, we found that even major differences in model architecture (e.g. CNNs vs. Transformers) did not significantly lead to better or worse brain alignment…
Tweet media one
2
10
59
4
6
41
@AndrewLampinen
Andrew Lampinen
5 months
How far can we trust representation analyses or mechanistic interpretations? @danfriedman0 's new work shows that analyses based on simplifying the model or its representations can be misleading about how the model will behave out of distribution! Check out his thread:
@danfriedman0
Dan Friedman
5 months
We often interpret neural nets by studying simplified representations (e.g. low-dim visualization). But how faithful are these simplifications to the original model? In our new preprint, we found some surprising "interpretability illusions"... 1/6
Tweet media one
3
49
298
2
4
39
@AndrewLampinen
Andrew Lampinen
1 month
Very cool! Reminds me a bit of the hypernetwork meta-learning architecture we found was beneficial in — awesome to see some theoretical justification for why it might be useful!
@ssmonsays
Simon Schug
1 month
Many tasks are compositional. Their constituent parts can be recombined in exponentially many ways. How could neural networks possibly learn all those tasks? That is what our ICLR 2024 paper shows: How to fight exponentials with exponentials 🧵️
3
22
118
0
4
41
@AndrewLampinen
Andrew Lampinen
1 year
@kareem_carr Are you familiar with the PDP books ()? They focused on NNs as cognitive models more constrained by some aspects of neural processing, and also introduce concepts like representation learning by backpropagation which are rather fundamental to modern DL.
2
0
39