PrometheusRoot
Blog Links Prometheans 100+ Why are you here?
← Prometheans 100+
×
Samy Bengio
builder
Researcher
google-brainappleresearch

Related

legend Yoshua Bengio builder Jeff Dean
← Prometheans 100+ Samy Bengio

Former Google Brain lead, now Apple AI research

Samy Bengio

Senior Director, AI Research — Apple

Profile

Samy Bengio is the Senior Director of AI and Machine Learning Research at Apple, where he leads a large research organization whose work now ships inside iPhones, iPads, and Macs as the foundation of Apple Intelligence. Before Apple, he spent 14 years helping build Google Brain from the inside — one of the early scientific leaders of the team that turned deep learning from an academic curiosity into the engine of modern AI. He earned his Ph.D. from the Université de Montréal in 1993, and yes, he is the younger brother of Yoshua Bengio — though he has built a serious body of his own work, with roughly 250 papers and over 100,000 citations.

His Google-era fingerprints are on some of the papers developers still quote today. Show and Tell, co-authored with Oriol Vinyals, was one of the first end-to-end neural image captioning systems. Scheduled Sampling gave sequence-to-sequence models a way to survive their own generation errors during training. And Understanding deep learning requires rethinking generalization forced the field to admit that nobody actually knew why over-parameterized networks worked — a question still not fully settled a decade later. He was NeurIPS program chair in 2017 and general chair in 2018, which is about as central to the community as you can get without running a lab of your own.

He left Google in 2021 in the fallout from the firings of Timnit Gebru and Margaret Mitchell, both of whom reported to him. Apple picked him up within weeks, and the hire landed as a clear signal: Apple, long seen as the sleepiest of the big AI labs, was finally playing for research talent. His group now publishes aggressively at NeurIPS, ICLR, and ICML while also feeding directly into Apple’s on-device and server foundation models. He’s also an adjunct professor at EPFL.

What makes him worth following for anyone building with AI: his team sits at the unusual intersection of serious academic research and constrained, ship-it product engineering. Apple cannot run a 400B-parameter model on a phone, so his researchers have to care about efficiency, distillation, and what models actually do — not just what they score on benchmarks. His 2025 paper The Illusion of Thinking, which picked apart the reasoning claims of frontier models using controlled puzzle environments, was the most-discussed AI paper of its quarter and a rare case of a big-tech lab publicly poking holes in the dominant narrative. That’s the flavor of work to expect from him.

Key Articles & Papers

The Illusion of Thinking: Understanding the Strengths and Limitations of Reasoning Models via the Lens of Problem Complexity 2025 — Controlled puzzle experiments showing frontier reasoning models hit a complete accuracy collapse past a complexity threshold — a pointed critique of the chain-of-thought hype cycle. GSM-Symbolic: Understanding the Limitations of Mathematical Reasoning in Large Language Models 2024 — Shows that LLM math performance degrades sharply when you change surface details of a problem, suggesting pattern matching rather than reasoning. How Far Can Transformers Reason? The Globality Barrier and Inductive Scratchpad 2024 — Theoretical limits on what transformers can learn in one pass, and how scratchpads extend those limits. Understanding Deep Learning Requires Rethinking Generalization 2016 — Showed neural nets can perfectly fit random labels, forcing the field to reckon with the fact that classical generalization theory doesn't explain why deep learning works. Show and Tell: A Neural Image Caption Generator 2014 — Early influential end-to-end deep learning approach to image captioning, combining CNN vision features with an RNN language model. Scheduled Sampling for Sequence Prediction with Recurrent Neural Networks 2015 — Curriculum trick that bridges the gap between teacher-forced training and autoregressive inference — a workhorse technique for sequence models. Adversarial Examples in the Physical World 2016 — Demonstrated that adversarial perturbations survive the camera — a foundational result in adversarial ML and physical-world robustness. Apple Intelligence Foundation Language Models 2024 — Tech report on the on-device and server models that power Apple Intelligence, from the team Bengio leads. Transformers Learn Through Gradual Rank Increase 2023 — Analysis of how transformer representations evolve during training, shedding light on the implicit learning dynamics of attention.

Spotify Podcasts

Episode 12 - Samy Bengio and Yoshua Bengio
Episode 12 - Samy Bengio and Yoshua Bengio
Episode 12 - Samy Bengio and Yoshua Bengio
Episode 12 - Samy Bengio and Yoshua Bengio
Apple hires ex-Google scientist Samy Bengio to lead new AI research
Apple hires ex-Google scientist Samy Bengio to lead new AI research
Google AI research manager Samy Bengio quits after two ousted female colleagues
Google AI research manager Samy Bengio quits after two ousted female colleagues

Related People

legend Yoshua Bengio builder Jeff Dean
© 2026 PrometheusRoot