Gary Marcus
Professor Emeritus of Psychology & Neural Science, New York University
Cognitive scientist and AI's most persistent skeptic — argues that deep learning alone will never achieve real intelligence and that the AI hype cycle is dangerously overblown.
Credentials
PhD in Brain and Cognitive Sciences (MIT, advised by Steven Pinker), Professor Emeritus at NYU, founded Geometric Intelligence (acquired by Uber in 2016), CEO of Robust.AI, author of five books on AI and cognition, regular contributor to The New Yorker and Scientific American
Why They Matter
Marcus is the contrarian voice that keeps AI discourse honest. While everyone else hypes LLMs, he systematically documents their failures and argues they cannot achieve true understanding. For business leaders, his critique is a reality check: he helps separate what AI can actually do reliably from what the marketing says. If your AI strategy depends on LLMs being more capable than they are, Marcus is the person pointing that out.
Positions
AI Timeline View
True AGI is much further away than the hype suggests. Current approaches (deep learning, LLMs) hit fundamental ceilings that require entirely new paradigms to overcome.
Safety Stance
Key Beliefs
Deep learning is necessary but not sufficient for AGI. We need hybrid architectures that combine neural networks with symbolic reasoning.
Rebooting AI: Building Artificial Intelligence We Can Trust (with Ernest Davis)
Large language models do not understand language — they are sophisticated pattern matchers that lack genuine comprehension, reasoning, and reliability.
Various Substack posts and media appearances
The AI industry suffers from a massive hype problem that leads to misallocated investment and public misunderstanding.
Substack: "The Road to AI We Can Trust"
AI safety is a real concern but the near-term risks (misinformation, bias, reliability failures) matter more than speculative existential scenarios.
Congressional testimony on AI oversight
Controversial Take
Marcus argues that scaling up current LLM architectures will not lead to AGI — a direct challenge to the "scaling hypothesis" that drives billions of dollars in investment at OpenAI, Anthropic, and Google. He predicts that LLMs will hit a capability wall and that the industry will eventually recognize the need for fundamentally different approaches combining neural and symbolic methods.
Track Record
How well have Gary Marcus's predictions held up?
Deep learning would hit a wall in tasks requiring systematic generalization, compositional reasoning, and reliability
Made: 2018
LLMs still struggle with reliability and hallucination, but they have exceeded expectations on many reasoning benchmarks. The debate continues.
Self-driving cars were much further away than the industry claimed (criticized Musk's "next year" timelines)
Made: 2018
Full self-driving remains unachieved as of 2026 despite repeated predictions of imminent arrival.
GPT-3 and its successors would not solve the fundamental problems of AI understanding and reliability
Made: 2020
GPT-4 and Claude dramatically improved capabilities beyond what Marcus predicted, but hallucination and reliability remain unsolved.
Key Quotes
“Deep learning is not going to give us artificial general intelligence. It's an important tool, but it's not the whole story.”
“We are nowhere near artificial general intelligence and the hype around it is both scientifically inaccurate and potentially dangerous.”
“Large language models are like autocomplete on steroids. They can be impressive and useful, but they don't understand what they're saying.”
“The biggest risk of AI right now is not that it's too smart. It's that people think it's smarter than it is and trust it with things they shouldn't.”
Publications
Rebooting AI: Building Artificial Intelligence We Can Trust
2019
Guitar Zero: The New Musician and the Science of Learning
2012
The Algebraic Mind: Integrating Connectionism and Cognitive Science
2001
Deep Learning: A Critical Appraisal
2018
Connections
Agrees With
Last updated: 2026-04-12
←Back to AI Minds Directory