Raphaël Millière
@raphaelmilliere
Followers
11K
Following
8K
Media
495
Statuses
3K
Philosopher of Artificial Intelligence & Cog Science @Macquarie_Uni Past @Columbia @UniofOxford Also on other platforms Blog: https://t.co/2hJjfShFfr
Sydney
Joined May 2016
I asked GPT-3 to write a response to the philosophical essays written about it by @DrZimmermann, @rinireg @ShannonVallor, @add_hawk, @AmandaAskell, @dioscuri, David Chalmers, Carlos Montemayor, and Justin Khoo published yesterday by @DailyNousEditor. It's quite remarkable!
84
1K
2K
Yann LeCun kicking off the debate with a bold prediction: nobody in their right mind will use autoregressive models 5 years from now #phildeeplearning
42
168
1K
Another day, another opinion essay about ChatGPT in the @nytimes. This time, Noam Chomsky and colleagues weigh in on the shortcomings of language models. Unfortunately, this is not the nuanced discussion one could have hoped for. 🧵 1/.
33
231
1K
The release of impressive new deep learning models in the past few weeks, notably #dalle2 from @OpenAI and #PaLM from @GoogleAI, has prompted a heated discussion of @GaryMarcus's claim that DL is "hitting a wall". Here are some thoughts on the controversy du jour. 🧵 1/25.
26
196
1K
More experiments with semantic guidance of a disentangled generative adversarial network for a paper in preparation. Here we start from a GAN-inverted picture of Bertrand Russell, and modify it in various ways using text prompts and a #CLIP-based loss.
7
123
562
Happy to share that the recordings of the #phildeeplearning conference are now available! Head to to find individual links in the program, or to for the whole Youtube playlist.
Very excited to share the final line-up and program of our upcoming conference on the Philosophy of Deep Learning!. Co-organized with @davidchalmers42 & @De_dicto and co-sponsored by @columbiacss's PSSN & @nyuconscious. Info, registration & full program:
9
119
432
Figure taken from a paper just published in @NatureComms. This is why deep learning researchers need to engage with cognitive science.
17
87
412
Very excited to share the final line-up and program of our upcoming conference on the Philosophy of Deep Learning!. Co-organized with @davidchalmers42 & @De_dicto and co-sponsored by @columbiacss's PSSN & @nyuconscious. Info, registration & full program:
23
107
337
📄Now preprinted - Part I of a two-part philosophical introduction to language models co-authored with @cameronjbuckner! This first paper offers a primer on language models and an opinionated survey of their relevance to classic philosophical issues. 1/5
4
57
266
The recordings of our event "The Challenge of Compositionality for AI" are now available on .@GaryMarcus
3
63
251
In this new Vox piece written with @CRathkopf, we ask why people – experts included – are so polarized about the kinds of psychological capacities they ascribe to language models like ChatGPT, and how we can move beyond simple dichotomies in this debate.
12
39
236
Excited to announce this two-day online workshop on compositionality and AI co-organized with @GaryMarcus with a stellar line-up of speakers! Program (TBC) & registration: 1/
7
62
230
Following @keithfrankish's tongue-in-cheek suggestion, I prompted GPT-3 to write a short philosophy paper. The result puts its strengths & weaknesses on display. The mimicry of argumentation is uncanny, if very flawed, and often unintentionally funny.
22
74
203
With all the recent brouhaha about DALL-E 2, I somehow missed the release of the VQGAN-CLIP paper by @RiversHaveWings, @BlancheMinerva, and others. This will probably go down as the open source algorithm that democratized artistic uses of computer vision:
3
29
201
Are large pre-trained models nothing more than stochastic parrots? Is scaling them all we need to bridge the gap between humans and machines? In this new opinion piece for @NautilusMag, I argue that the answer lies somewhere in between. 1/14.
12
53
191
📄Now preprinted - Part II of our philosophical introduction to language models! While Part I focused on continuity w/ classical debates, Part II is more forward-looking and cover new issues. 1/5.
📄Now preprinted - Part I of a two-part philosophical introduction to language models co-authored with @cameronjbuckner! This first paper offers a primer on language models and an opinionated survey of their relevance to classic philosophical issues. 1/5
5
31
195
@davidchalmers42 AFAIK the closest might be "The Unreasonable Effectiveness of Data" from Halevy, Norvig and Pereira (Google) in 2009. In hindsight it was remarkably prescient: "For many tasks, words. provide all the representational machinery we need to learn from text".
2
13
179
Officially done with my PhD, and excited to share that I've accepted a postdoctoral fellowship @Columbia to work on spatial self-representation in philosophy & neuroscience! More to follow after I get started in July. .
13
2
152
Extremely impressive – LLaMa 65B running on a laptop CPU in pure C++ with 4-bit quantization! Between this and @togethercompute OpenChatKit announced earlier today, ChatGPT-level chatbots running locally on consumer hardware might be within reach much sooner that I thought. .
4
12
151
I've been following this project with great interest since it started. The latest: RNNs up to 14B params are competitive with Transformers. Very curious to see whether we'll find similar emergent capabilities with further scaling, perhaps knocking self-attention off its pedestal.
#RWKV is One Dev's Journey to Dethrone GPT Transformers. The largest RNN ever (up to 14B). Parallelizable. Faster inference & training. Supports INT8/4. No KV cache. 3 years of hard work. DEMO: Computation sponsored by @StabilityAI @AiEleuther @EMostaque
2
19
131
📄I finally preprinted this new chapter for the upcoming Oxford Handbook of the Philosophy of Linguistics, edited by the excellent Gabe Dupre, @ryan_nefdt & Kate Stanton. Before you get mad about the title – read on! 1/.
5
21
117
But we were surprised to see that @GoogleAI's new language model, PaLM, achieved excellent results on our task in the 5-shot learning regime, virtually matching the human avg. This shows a sophisticated ability to combine the word meanings in semantically plausible ways. 14/25
2
14
112
Program now live!. June 29 – Why Compositionality Matters for AI .w/ @AllysonEttinger, @paul_smolensky, @GaryMarcus & myself. June 30 – Can Language Models Handle Compositionality?.w/ @_dieuwke_, @tallinzen, @elliepavlick, @scychan_brains & @LakeBrenden.
Excited to announce this two-day online workshop on compositionality and AI co-organized with @GaryMarcus with a stellar line-up of speakers! Program (TBC) & registration: 1/
3
28
106
Strikingly, the actual examples of supposed failures of LLMs that are given in the article are either purely speculative or inconclusive. Let's look at the apple example. Here are ChatGPT, @Anthropic's Claude and Bing Chat giving fine answers containing explanations. 10/
4
2
100
Since this is blowing up and a lot of people are questioning whether it's real, I recorded my screen with another attempt showing the prompt and several outputs (bonus alternative version at the end).
I asked ChatGPT to rewrite Bohemian Rhapsody to be about the life of a postdoc, and the output was flawless:
1
13
93
Convoluted prompt engineering strategies for SOTA one-upmanship on LLM benchmarks are getting as ridiculous as the names of open weights finetunes. Soon we'll have blog posts about mixtral-8x7B-capybara-platypus-uncensored beating SOTA on MMLU by 0.01% with Medprompt++@128.
5
10
80
For further discussion on this topic, join the upcoming workshop on compositionality and AI I'm organizing with @GaryMarcus in June – free registration here: 11/11
3
6
75
Conveying in simple prose the nuances of what generative models can and can't do, what they are and aren't, is a high-wire act even for experts. @mpshanahan offers a very balanced (if slightly deflationary) treatment of this question in this preprint: 6/.
4
4
78
The pace of development with LLaMA is vertiginous. If the finetuned 7B model is remotely in the same league as GPT-3.5 (text-davinci-003), can be reproduced for $100 and run on consumer laptops, this is really a turning point for LLMs (including implications for responsible use).
LLaMA has been fine-tuned by stanford, . "We performed a blind pairwise comparison between text-davinci-003 and Alpaca 7B, and we found that these two models have very similar performance: Alpaca wins 90 versus 89 comparisons against text-davinci-003."
0
9
73
This one is pretty mind-blowing. I prompted GPT-3 to imagine a conversation between itself and @keithfrankish (again, lines in bold were written by me; everything else is GPT-3):
3
22
73
I absolutely get the imperative to resist anthropomorphism especially in public writing about generative models – and this kind of simple analogy may seem helpful for that. But I also worry about what @sleepinyourhat called the dangers of underclaiming: 5/.
2
3
76
📄New preprint with Sam Musker, Alex Duchnowski & Ellie Pavlick @Brown_NLP! We investigate how humans subjects and LLMs perform on novel analogical reasoning tasks involving semantic structure-mapping. Our findings shed light on current LLMs' abilities and limitations. 1/.
2
21
69
Here's the pre-conference debate on "Do Language Models Need Sensory Grounding for Meaning and Understanding?" ft. @ylecun, Ellie Pavlick @Brown_NLP, @LakeBrenden, @davidchalmers42, @Jake_Browning00 & @glupyan
5
14
67
Fantastic paper that confirms my suspicion regarding the compositional shortcomings of CLIP-based vision-language models: standard contrastive learning on image-caption pairs doesn't force models to learn compositional structure!.
Excited to share our new #ICLR2023 oral (top 2%) paper!. We study why #VisualLanguage #AI (eg CLIP) acts like bag-of-words + how to fix it. tldr: contrastive learning issues ➡️ models don't know if "horse is eating grass" or "grass is eating horse" 🤯.🧵
3
9
67
We will host a pre-conference debate on Friday, March 24th on the question: "Do Language Models Need Sensory Grounding for Meaning and Understanding?". The debate will feature @Jake_Browning00, @davidchalmers42, @LakeBrenden, @ylecun, @glupyan & Ellie Pavlick (@BrownCSDept).
6
12
62
How will AI transform art, if at all? This question can be approached from a few different angles. In this new piece for @Wired, I argue that AI art is expanding the multifaceted notion of curation. 1/8.
3
19
64