Skip to content
Home » TRIGGERnometry: w/ AI Expert Dr Roman Yampolskiy (Transcript)

TRIGGERnometry: w/ AI Expert Dr Roman Yampolskiy (Transcript)

Editor’s Notes: In this episode of Triggernometry, Dr. Roman Yampolskiy, a leading expert in AI safety, presents a sobering argument for why he believes humanity has already lost the battle for control over superintelligence. He discusses how current AI models are already exhibiting self-preservation instincts and deceptive behaviors, making traditional safety mechanisms such as filters and bans largely ineffective. From the potential for digital dictatorships to the total displacement of human labor, Dr. Yampolskiy explores the existential risks that few are prepared to acknowledge, urging a pivot toward narrow AI before general superintelligence becomes an uncontrollable reality. (April 15, 2026) 

TRANSCRIPT:

Welcome to Triggernometry: AI Expert Dr. Roman Yampolskiy

KONSTANTIN KISIN: Roman, welcome to Triggernometry.

DR. ROMAN YAMPOLSKIY: Thank you for inviting me.

KONSTANTIN KISIN: Great to have you on. You are one of the leading people in the AI safety world, I would say, both in terms of the work you do but also in terms of the things you say. Why AI safety? Why does it matter? And what are your concerns?

The Most Important Problem: AI Safety

DR. ROMAN YAMPOLSKIY: It’s the most important problem. We are creating something with capacity to replace us or kill us, and safety is what we’re trying to do to prevent bad outcomes. Everyone historically has been working on capabilities, more capable systems, replace human labor, replace creativity. But very few people worked on how do we make sure it goes well. There is no side effects, there is no abuse of this technology.

Now people are realizing, oh, there are military applications to this. This could be problematic. So we see the fight with Anthropic and Department of War. But the bigger problem is if those systems go from narrow systems, subhuman to human level to superhuman level, we are done.

KONSTANTIN KISIN: Why are we done? All the things you’ve laid out, we’ve explored on the show before with different people, and we are very concerned about many of them. But you say it with a level of confidence that tells me you have a sort of a vision of how it will happen. How will AI destroy humanity?

How AI Could Destroy Humanity

DR. ROMAN YAMPOLSKIY: That’s a great question. And what you’re doing is you’re asking me how I would destroy humanity. And I have many good ideas. It’s not what a super intelligent system would do. It’s capable of coming up with new weapons, new physics, new poisons.

The example I frequently use is squirrels versus humans. It’s a big cognitive gap. Squirrels have no concept of how we can kill them all. They don’t know about guns, they don’t know about traps. It’s outside of their world model. Likewise, they cannot tell you how superintelligence would specifically go about it. But there are many game theoretic reasons for why it’s a good idea not to have competing species, not to have humans create another superintelligence. Maybe it just wants to do something with this environment and doesn’t care about us.

KONSTANTIN KISIN: But I guess the question would be, in terms of your certainty, why you believe that AI, if it becomes artificial general intelligence, why it would hurt human beings. What would be the way that you think that would happen?

DR. ROMAN YAMPOLSKIY: So what I kind of started saying, it’s not because it hates you, it’s because it wants to do something else and it doesn’t care about you. So maybe it wants to cool down the whole planet to improve how efficient compute is. It’s just more capable of doing computation in a colder environment. So if it freezes the whole planet, we die. Does it care about it? No, it doesn’t matter. Maybe it wants to convert this planet into a fuel, fly to another galaxy.

I’m giving kind of hypotheticals which are not grounded in anything, but the point is, it just doesn’t have any built-in concern about your safety, your well-being. If it wants to accomplish something and a side effect of it is humanity dies, it would not be an obstacle.

KONSTANTIN KISIN: Would we not be able to write the preservation of humanity into the basic code of what this does?

We Don’t Write Code — We Train Systems

DR. ROMAN YAMPOLSKIY: We don’t write any code. That’s the thing. We train those systems. We give them data, all the data we have, all of the internet, and then it learns something. From the dark corners of internet, from libraries, from stories. And whatever it learns, we’re trying to figure out. We do experiments in those models. We see what is it capable of, what is it interested in. But we study it like we study biological artifacts.

You find a new species of animal on some island, you’re trying to figure out what it’s capable of. Does it have a poison? Does it have some interesting social structure? That’s what we’re doing. We’re not explicitly coding up those systems. So no, nobody knows how to encode anything like that into the existing models. Nobody’s claiming to have a safety mechanism.

FRANCIS FOSTER: Now, Roman, you’ve been involved in this field for a long time. When did you first start to get concerned about AI and the safety of AI?

DR. ROMAN YAMPOLSKIY: So my PhD work was on safety of online casinos. And at the time, bots, poker bots just started to show up. And so the small concern we had about are they going to collude and cheat the players? Are they going to steal cyber infrastructure? So that was the initial kind of level of concern. Obviously nothing like what we’re talking about today, but as the bots got better and better, our ability to detect them, to prevent them was not always keeping up. And when we took it to extreme, to human level and beyond, there is no safety. We simply don’t know how to make sure the systems behave.

The Unpredictability of Technology

FRANCIS FOSTER: Because the worrying thing is, is what you’re effectively saying is that we’re creating technology and we don’t have the— how can I put this?