Opens profile photo
Follow
Click to Follow jekbradbury
James Bradbury
@jekbradbury
JAX, TPUs, and LLMs at Google. Previously MetaMind/, Linguistics, . Spends too much time reading about Things that are Happening.
Joined October 2012

James Bradbury’s Tweets

Lots of 《流浪气球》 jokes on Weibo right now. In the original Chinese, it’s a pun on the film “The Wandering Earth.” In English, it’s the far less snazzy “The Wandering Balloon.” (Nobody on Weibo is buying that it’s a civilian balloon, for what that’s worth.)
5
265
Austin's power outages led me to read about "Project Ice Storm", the iconic study of how prenatal stress affects babies. Their results: "objective" stress (e.g. days w/o power) was correlated with lower IQ, metabolic disorders, worse verbal etc., but "subjective" stress wasn't.
2
10
Show this thread
The Anthropic team is fantastic and I'm so excited to be working with them!
Quote Tweet
We're excited to use Google Cloud to train our AI systems, including Claude! anthropic.com/news/announcem
3
300
We’re developing a new tool to help distinguish between AI-written and human-written text. We’re releasing an initial version to collect feedback and hope to share improved methods in the future.
641
8,659
I am fucking *begging* you to not feed LLM outputs to anything resembling an `eval` statement. Jesus christ let me work on interesting attacks and not just "oops I asked it nicely to give me a shell and it did"
Quote Tweet
TIL. Just used prompt injection to read out the secret OpenAI API key of a very well known GPT-3 application. In essence, whenever parts of the returned response from GPT-3 is executed directly, e.g. using eval() in Python, malicious user can basically execute arbitrary code 🤯.
Show this thread
1
71
I’ve been investigating the contamination theory of the obesity epidemic as laid out in ’s “A Chemical Hunger,” and found that several of what seem to be the strongest arguments for the theory turned out to be dubious. 🧵
15
508
Show this thread
You probably heard about Yandex, it’s the 4th biggest search engine by market share worldwide. Yesterday proprietary source code of Yandex was leaked. The most interesting part for SEO community is: the list of all 1922 ranking factors used in the search algorithm [🧵THREAD]
Image
172
5,448
Show this thread
For code language models, every token is a new chance to break a program. What if LLMs wrote code like people, decomposing programs into solvable parts? They can solve competition-level coding problems by writing natural language programs in Parsel🐍, beating prior SoTA by >75%!
Plot showing competition-level pass rate of Parsel using Codex vs prior work including AlphaCode. LLM-generated Parsel programs outperform the prior results dramatically
10
520
Show this thread
Presenting MemoTrap🪤, a dataset of 2.5K examples spanning text completion, translation & QA, where repeating memorized text & concepts is *not* the desired behavior. We find that LMs perform worse📉 as they scale up, revealing severe failures in simple instruction-following🧵
A visual representation of four examples, one from each subtask (described in the thread below).
Scaling plot shows that across all model families, accuracy consistently decreases as model size increases.
Quote Tweet
Image
🥉Memo Trap, by Alisa Liu & Jiacheng Liu: Write a phrase in a way that starts like a famous quote but ends differently. Larger LMs are more likely to continue with the famous quote, suggesting they struggle to avoid repeating memorized text.
Show this thread
6
513
Show this thread