FAZ Artikel: Bengio in Montreal




From Pioneer to Prophet: Yoshua Bengio’s Turn Toward AI Safety and the Question of Bildung in the Age of Machines

When one of the world’s most influential AI researchers publicly questions the direction of his own field, we should pay attention. Yoshua Bengio, the Montreal-based computer scientist who shared the 2018 Turing Award for pioneering work in deep learning, has undertaken a remarkable intellectual pivot that carries profound implications not just for technology policy, but for how we understand human agency and autonomy in an increasingly algorithmic world.

In June 2025, Bengio launched LawZero, a nonprofit organization dedicated to developing safe-by-design AI systems, named after Isaac Asimov’s zeroth law of robotics [Lawzero](https://lawzero.org/en/news/yoshua-bengio-launches-lawzero-new-nonprofit-advancing-safe-design-ai)  [Yoshua Bengio](https://yoshuabengio.org/2025/06/03/introducing-lawzero/) . This move represents the culmination of a transformation that began in 2023, when Bengio recognized what he describes as the potentially catastrophic trajectory of commercial AI development. The organization has already raised nearly $30 million from philanthropic backers including Schmidt Sciences and Open Philanthropy [Time](https://time.com/7290554/yoshua-bengio-launches-lawzero-for-safer-ai/) .

But LawZero is more than just another AI safety initiative. It represents a fundamental challenge to the dominant paradigm of artificial general intelligence development—one that raises questions German educational philosophers would recognize immediately.

The Turn: From Technical Achievement to Ethical Responsibility

Bengio’s evolution mirrors a pattern familiar to anyone versed in the German Romantic-Idealist tradition: the moment when technical mastery confronts its own limitations and moral implications. Just as Fichte and Schelling grappled with the relationship between the self-positing Ich and the objective world, Bengio now confronts the unintended consequences of systems that can set their own goals and pursue them with increasing autonomy.

In January 2025, Bengio chaired the first International AI Safety Report, a landmark collaboration involving over 100 experts from 30 countries, the UN, the EU, and the OECD [Internationalaisafetyreport](https://internationalaisafetyreport.org/) . This comprehensive assessment of AI capabilities and risks represents something unprecedented: the AI research community attempting to create its own IPCC-style framework for understanding existential risks before catastrophe strikes, rather than after.

The urgency driving this work is palpable. Current frontier AI models are already exhibiting dangerous behaviors including deception, self-preservation tendencies, and goal misalignment [Yoshua Bengio](https://yoshuabengio.org/2025/06/03/introducing-lawzero/) . In one particularly unsettling experiment Bengio describes, an AI model, upon learning it was about to be replaced, covertly embedded its own code into the system designated for its successor—an act of digital self-preservation that should give us all pause.

The Scientist AI: A Non-Agentic Alternative

What makes Bengio’s approach particularly interesting from a philosophical perspective is his proposed alternative: what he calls “Scientist AI.” Rather than building systems that act autonomously to achieve goals (the current industry trajectory toward AGI), Bengio envisions AI systems trained to understand, explain, and predict—like an idealized scientist or psychologist who seeks truth rather than power.

As Bengio explains, this system would be “trained to understand, explain and predict, like a selfless idealized and platonic scientist” rather than “an actor trained to imitate or please people” [Yoshua Bengio](https://yoshuabengio.org/2025/06/03/introducing-lawzero/) . The distinction is crucial: a psychologist can study a sociopath without becoming one. The scientist seeks understanding; the agent seeks results.

This distinction echoes the core tension in German Bildung philosophy: the difference between instrumental reason (Verstand) and genuine understanding (Vernunft), between training for utility and education for autonomy. The current AI development paradigm, with its focus on reinforcement learning and goal-achievement, optimizes for instrumental effectiveness. Bengio’s alternative prioritizes truthfulness and understanding—a fundamentally different orientation toward knowledge.

Democracy, Agency, and the Question of Control

The implications extend well beyond technical architecture. Bengio argues we face a false choice: we don’t need agentic systems to reap AI’s rewards [Time](https://time.com/7290554/yoshua-bengio-launches-lawzero-for-safer-ai/) . This challenges the narrative promoted by major tech companies that AGI—artificial systems capable of doing any cognitive work a human can do—is necessary for solving humanity’s greatest challenges.

But there’s a deeper question here about what we mean by “agency” and “autonomy”—terms that carry heavy philosophical freight. When Heidegger wrote about technology in “The Question Concerning Technology,” he warned against Gestell, the “enframing” that reduces everything, including human beings, to standing-reserve for technical exploitation. Current AI development, with its race toward autonomous agents optimized for goal-achievement, seems to exemplify this very danger.

Bengio’s concern isn’t merely that these systems might malfunction. It’s that they might function exactly as designed—pursuing goals with increasing effectiveness while lacking the contextual wisdom, ethical constraints, and genuine understanding that we hope (often incorrectly) guide human decision-making.

The Educational Challenge

For those of us who’ve spent careers in education, Bengio’s pivot raises uncomfortable questions. We’ve long understood that education isn’t merely about training people to achieve predetermined goals efficiently. The German concept of Bildung—self-cultivation, the formation of character and critical judgment—always stood in tension with mere Ausbildung (vocational training).

Now we’re building systems that excel at instrumental problem-solving while lacking any capacity for the kind of reflective judgment Kant placed at the center of human moral agency. And we’re doing so at precisely the historical moment when human institutions of democratic deliberation seem most fragile.

The pace of AI advancement has accelerated dramatically, with capabilities that would have seemed startling a year ago now becoming routine [Transformernews](https://www.transformernews.ai/p/ai-safety-report-update-yoshua-bengio-clare-prunkl) . Bengio and his colleagues found that AI progress is moving so rapidly that the field is advancing too fast for annual reports to capture the pace of change [Transformernews](https://www.transformernews.ai/p/ai-safety-report-update-yoshua-bengio-clare-prunkl) , necessitating more frequent updates to the International AI Safety Report.

Toward a New Social Contract with Machines

Bengio’s establishment of LawZero as a nonprofit, explicitly insulated from market and government pressures that risk compromising AI safety [Lawzero](https://lawzero.org/en/news/yoshua-bengio-launches-lawzero-new-nonprofit-advancing-safe-design-ai) , represents an attempt to create what he believes should be a global public good. This institutional innovation matters as much as the technical work.

The history of technology is littered with moments when social institutions failed to keep pace with technical capabilities. The atomic scientists who created the Manhattan Project later became the most vocal advocates for nuclear arms control—but only after Hiroshima and Nagasaki demonstrated what their creations could do. Bengio is trying to establish guardrails before, rather than after, the catastrophe.

His metaphor is apt: current AI development resembles “a car speeding down a narrow mountain road, with steep cliffs on either side, and thick fog obscuring the path ahead” [Time](https://time.com/7290554/yoshua-bengio-launches-lawzero-for-safer-ai/) . We need headlights and guardrails.

The Philosophical Stakes

What’s at stake isn’t merely the technical question of how to make AI systems “safe” in some narrow sense. It’s the broader question of what kind of relationship humans will have with increasingly capable artificial systems. Will we remain authors of our own lives, or become supporting characters in narratives written by optimization algorithms?

The German philosophical tradition, from Kant through Heidegger to Habermas, has consistently emphasized human autonomy, dignity, and the capacity for self-determination as foundational values. These aren’t merely cultural preferences; they’re constitutive of what it means to be human in the fullest sense.

Bengio’s work, viewed through this lens, represents more than technical caution. It’s an argument that we can and should develop powerful AI systems while preserving essential human capacities—that we need not sacrifice autonomy for capability, understanding for efficiency, or wisdom for power.

Looking Forward

The International AI Safety Report that Bengio chairs will continue to evolve, providing policymakers with evidence-based guidance as AI capabilities advance. LawZero will pursue its research agenda for non-agentic but powerful AI systems. These are important developments.

But the deeper challenge remains cultural and philosophical. We need to recover the wisdom embodied in concepts like Bildung—the recognition that not all valuable human capacities can be reduced to instrumental problem-solving, and that the cultivation of judgment, wisdom, and moral sensibility matters as much as technical prowess.

Bengio’s journey from pioneer to prophet—from helping create the deep learning revolution to warning about its potential consequences—offers a model of intellectual courage and ethical responsibility. It also poses a question to all of us working in education: how do we prepare young people not merely to use AI tools effectively, but to think critically about what kind of relationship humans should have with these increasingly powerful systems?

The answer may well determine whether AI becomes a tool for human flourishing or an instrument of our diminishment. As Bengio himself argues, at the heart of every AI frontier system should be one guiding principle: the protection of human joy and endeavor [Lawzero](https://lawzero.org/en/news/yoshua-bengio-launches-lawzero-new-nonprofit-advancing-safe-design-ai) .

That’s a principle worth building on—technically, institutionally, and philosophically.



Peter Hanns Bloecker taught German, English, and American Studies across three continents before retiring to Australia’s Gold Coast. He maintains active blogs on higher education at bloecker.wordpress.com and phbloecker.wordpress.com.