Haseeb Quereshi: Crypto’s Not Made for Humans—It’s for AI
Haseeb:
[0:00] Where do AI agents have comparative advantages over human beings? The answer, I think, is most obviously is that you cannot enforce the law against an AI agent. If you are a self-sovereign agent, there is no monopoly on violence. You can't throw an AI agent in jail. So what can an AI agent do that's hard for a human being to do? The answer is crime.
David:
[0:20] I was about to say,
Haseeb:
[0:22] Oh no, I don't like where he's going. Exactly. Like if you're talking about like scamming people, hacking people, like creating all sorts of nonsense on the Internet, that is where agents have a comparative advantage.
David:
[0:37] Haseeb, welcome back to Bankless.
Haseeb:
[0:39] Thanks for having me. Always good to be here.
David:
[0:42] Question for you, Haseeb. Why isn't crypto made for humans?
Haseeb:
[0:47] Crypto, you know, it's always been surprising how scary it is even 10 years in as a crypto user to sign a big transaction. And I was reflecting on the fact that I've actually never been scared to send a wire transfer. I'm never worried that, oh, you know, if I don't double, triple, quadruple check my wire transfer, I might accidentally send money to North Korea.
David:
[1:09] Right.
Haseeb:
[1:10] But I think about that every time I'm signing a big crypto transaction. It's just like the reality is that there's so many foot guns in crypto where, you know, you're reading an address and you have to think about, oh, wait, is this an address poisoning attack? Should I check the middle numbers instead of just beginning and the end? Should I be thinking about my stale approvals? I need to check the URL to make sure this is not slightly different than what it's supposed to be. There's all these foot guns that exist in crypto that don't exist in the traditional financial system. And up until now, the story in crypto, which is one that I largely believed, is that, well, this is the fault of lazy humans. And the humans just need to get with it. They need to get more security conscious.
Haseeb:
[1:49] They need this is your fault, not the technology's fault. And the longer I've sat with this, the more I've started to become convinced that if this is true, if we're still telling ourselves this 10 years later, then maybe the problem is not with the user. Maybe it's just that this is the wrong user. It started to really click for me when I kind of saw how capable AI agents were at navigating code compared to how difficult it is to navigate other kinds of poorly formed problems.
Haseeb:
[2:20] There used to be the story, and I remember the story when I first got into crypto, literally the first blog post I ever wrote about crypto talked about this, the idea that smart contracts were going to replace the law. They were going to replace traditional contracts. That's why they're called smart contracts, right? It's supposed to be this analogy that in the future, you're not going to sign an agreement that's adjudicated by lawyers. You're going to sign an agreement with somebody in code. And the reality is that that story never happened.
Haseeb:
[2:48] Right?
Haseeb:
[2:49] It's just not true that we use smart contracts instead of legal contracts. In fact, you know, we're a crypto VC. We are one of the most sophisticated financial actors in the crypto industry. And when we sign a deal to buy tokens, just to buy tokens from a counterparty that's like a foundation or a startup that's building a token, we sign a legal contract. In fact, even when we do have a smart contract, we also sign a legal contract just in case something goes wrong with a smart contract, which has happened.
Haseeb:
[3:19] So like what that tells you is that this stuff is not designed for humans. It's not designed for humans, but it is perfectly designed for non-human actors. So I made this analogy recently at Eat Denver that, you know, if you think about, you know, who is saying this? Who is making the argument that smart contracts are this perfect replacement for the traditional legal system, for traditional property rights? This story was being told by autistic software engineers who were the original people who built Ethereum, right? Well, it turns out that's not very much like most users of Ethereum.
Haseeb:
[3:52] Most of them are not autistic software engineers, but AI agents are actually more similar to autistic software engineers than they're similar to the rest of us. And so I think in a way, you can see how the idea that, okay, I'm going to negotiate a smart contract. I'm going to go back and forth with you about every single individual term. I'm going to statically analyze it and see all the different ways it could go wrong and maybe even formally verify it before I decide to agree to it, that's exactly what quad code can do perfectly in the span of minutes. Whereas as a human being, I need to go hire a software engineer and he spent a lot of time looking at all the code, thinking about all the edge cases, talking about it with my lawyers to do a risk analysis. I'm not comfortable doing that with a smart contract relative to a legal contract. An AI agent might be actually way more comfortable doing it with a smart contract than with a legal contract. The reality, and I pointed this out in this blog post that I wrote, a legal contract has all sorts of randomness built into it.
Haseeb:
[4:51] Okay.
Haseeb:
[4:52] What is the randomness in a legal contract? The randomness in a legal contract is when I send a legal contract with you, I don't know what is the jurisdiction in which this legal contract is going to be enforced, right? Maybe I'll say, okay, well, it's going to be in California because I'm in California, but you're in New York. And so if you're in New York and I'm in California, it's going to be a fight about where the case is going to get filed, right? Oftentimes it is a fight. Second, okay, let's even say that we're going to do this in New York, right? Let's say we're both in New York. Okay, if we're both going to do this in New York, then the question is, does every single clause of this smart contract hold, or sorry, this legal contract hold, or some of them are going to get struck as being unenforceable? This happens. This is oftentimes what happens in a dispute, is you argue this piece of the contract is not enforceable. Okay, then let's say we go to trial. Then who's my lawyer going to be? Who's your lawyer going to be? That's going to affect our odds of winning in the trial. Okay, then there's judge selection. Judge selection is literally random. It's a lottery. When you get which judge you select in a legal trial. And then, of course, there's the jury, which is also obviously random. So all these things are intentionally random. We designed them to be random, which means if you're an AI agent, you look at a legal contract and you're like, I don't know what's going to happen.
David:
[6:03] This is uninterpretable to me.
Haseeb:
[6:05] It's literally non-deterministic, right? Whereas a smart contract is actually machine code. It is actually compiled into the EVM bytecode and you can analyze it step by step. This is exactly what will happen in 100% of scenarios. Now, as human beings, we might know that that's true, but we don't feel that that's true. It doesn't feel intuitively true to us that the legal contract is actually less predictable than the smart contract. To us, it's the other way around. Even though there's all this randomness in the legal contract, we find the legal contract much easier to predict what's going to happen than the smart contract. And so my claim is that that's because of our bounded rationality. It's because of our inability to process code as effectively as an AI agent would. But for an AI agent, all the stuff that we were saying about how smart contracts are a better way to create enforcement and property rights, it's actually true for AI agents. And my claim is that that's going to change a lot of the way in which the original
Haseeb:
[7:00] promise of crypto gets interpreted. It's not going to be humans taking advantage of it. It's going to be AI agents in concert with humans taking advantage of it on behalf of the human that's acting on, acting for.
David:
[7:12] Now, Hasib, I was at ETH Denver with you not, not terribly long ago, and I had to download MetaMask to go like sign into ETH Denver, right? It's like, what year is it downloading MetaMask at ETH Denver? And I was actually pretty pleasantly surprised with the UX improvements of MetaMask. And I think that is emblematic of the industry as a whole. We have improved our human UX over the years, maybe not as fast or as much as we would have enjoyed, but like, nonetheless, it is happening. And I think something that what you're saying is a little bit more fundamental than that, of that, you know, AI isn't just going to gloss over some of the hard gaps of crypto UX for humans. When I open up my ledger and I see the blind signing thing, you know, an AI can go and interpret the code and tell me, you know, thumbs up, thumbs down, and maybe that helps crypto UX. And maybe that's true, but I think what you are getting at is something a little bit more fundamental about the nature of blockchains and what they are optimized for. Beyond just smoothing over crypto UX for humans, I think you're going deeper than that. And you're saying this is technology that's not for us fundamentally,
Haseeb:
[8:22] Critically.
David:
[8:23] Is that what you're saying?
Haseeb:
[8:24] So obviously, it's ultimately for humans in the sense that the human is the end recipient of value from what's happening on the blockchain. But the idea that as a human being, the right way that you are going to interact with the blockchain is by you moving your mouse, clicking on the MetaMask extension, typing in your little password, you know, going in there and like clicking the buttons and manually approving the gas. That stuff is obviously so alien to human beings and the way in which we think about money and finance and accomplishing our financial goals. It's a little bit like, you know, imagine that with the banking system, human beings had to write their SWIFT codes. SWIFT was not made for humans. SWIFT was made for an interbank communication protocol. It's not designed for you. Now, if you had to interact with it, I guess you would, you know, if you had no other choice. but it's clearly not the affordance that most human beings are looking for in their own intuitions about how to use money. So my claim is that, okay, we live today in the world where this stuff is totally, this stuff is totally disintermediated. The human is interacting directly with the machine. And this is bad. It's bad in the same way that we think today about cars, right? In 10 years, we're gonna look back with absolute horror that we ever thought it was a good idea to have apes using their hands to manipulate these two-ton death machines driving down the highway while drunk, half asleep, being ill, whatever state of mind you're in, yeah, you're probably fine.
David:
[9:52] Up here, barbaric.
Haseeb:
[9:54] Right, it's gonna seem absurd that we ever thought that was cool, that was just fine. And yeah, people die on the road all the time. It's like one of the most dangerous things we do is literally driving, but what else are we gonna do? Yeah, you can't take away a man's car. We're going to look back on that as being like insane. You know, there's going to be a day when human driving is banned and or only allowed in certain areas that are far away from other humans.
Haseeb:
[10:17] We're going to get there with respect to the way that our norms evolve. I think we're going to land at the same place with crypto. I think we're going to look back with horror at the idea that human beings were manually, you know, blind signing transactions and like eyeballing addresses and being like, yeah, this is 0x.
Haseeb:
[10:38] I think that's the same as the thing I saw. So yeah, I'll send it there. That's right.
Haseeb:
[10:43] Yeah, exactly. Like that stuff is going to go away. The idea that you're like manually looking at a URL and being like, I don't think this is a phishing domain. I looked at it and I don't think this is a phishing domain. That is going to go away. It's going to be like, what on earth? Why would you think that that's a good way for us to keep people safe and keep their money safe, right? Like, of course, human beings mess up. Of course, they're going to sometimes fail. You're going to be tired. You're going to be scared. You're going to be whatever. You don't have the energy to check it three times over to go check the DNS, to go check the Twitter and see, hey, has this protocol recently been hacked, right? Like, you know, when you see these announcements, like sometimes you'll go on Twitter and you'll be like, oh, it's trending that like such and such protocol got hacked. Okay, I sometimes wonder, what if I was a user of that protocol? I'm usually not. What if I was a user of that protocol and I just didn't check Twitter? Are people like have the Twitter notification for every protocol they use going straight to their phone? Is it pushed out? Like we don't even have an alert mechanism for like when a protocol gets hacked, you're supposed to just like check the Twitter and just happen to see it before you interact with this protocol. So the reality is like, there's so many things that you could be doing, but we're just not, we're human. We're human. And this whole idea that like, oh, we're human. Of course, we're going to make mistakes. The AI agent never gets tired, never gets lazy, never skips a step, never doesn't follow its instructions. And so you can imagine a world where, look, the interface, I think, for these things is going to look very different in a world that's totally intermediated by AI.
Haseeb:
[12:10] So let's say you have your AI agent and you tell your AI agent, hey, I think interest rates are going to go up. I think that we should be, you know, moving into safer DeFi. So let's move me out of, you know, whatever it is. I'm in, you know, riskier DeFi. I've looped something. Instead, you know, move me into just Aave. Oh, I just want to stay safe. Your AI agent will just go do that. It will just go do it for you. And if you're particularly, you know, in a YOLO mood, you will manually approve the transactions. Or sorry, it'll just do it for you. But if you're like, hey, you know, I want to make sure I know exactly what I'm doing, it will batch everything and it will say, look, here's a series of steps I'm going to do. Please approve this plan. And probably in the near future, it's going to be approve this plan. In the farther future, it's going to be, it just does it for you because you're not adding anything to the equation. But at least for the foreseeable future, it's going to be, okay, I present you this plan. Do you approve the plan? Cool, I'm going to go do it. In this world, one can imagine, so one, you're no longer clicking on Aave. Two, you're not even looking at the Aave logo. You're not looking at any of their marketing. And in fact, you're maybe not even telling it, put into Aave.
David:
[13:15] You're just telling it,
Haseeb:
[13:16] Put into some low-risk DeFi. Right, exactly. You don't know what Aave is.
Haseeb:
[13:18] You just say, hey, lower my risk profile. And it's going to say, cool. I found, actually, I shopped around. I looked at 14 different protocols and I looked at all the different promotions or TVL deals that are going on there. And I found one that actually is the best. And I'm going to go use this one. What happens to marketing? What happens to network effects? What happens to the ways in which we think about, you know, this was a lot of what the viral Citrini article was talking about. When AI automates discovery, how do companies compete with each other? How do protocols compete with each other? Because so many of their business models are dependent or have this baked in assumption of human friction. This idea that, well, a human's only gonna look at three or five, or they're only gonna look at the biggest one. And so all you really gotta do is get the biggest person and then they're sticky. They're not going to want to look around any further. But an agent doesn't necessarily think that way.
Haseeb:
[14:12] I think a lot.
Haseeb:
[14:13] If you assume that the story is true, a lot is going to change about the way the protocols work, about the way the protocols compete, and who's going to benefit from all of this. The answer largely is the consumer benefits, right? It's a big consumer surplus. If all of a sudden this efficiency is getting captured by the user, that's good for users. It's good for crypto. It ultimately means that crypto users are going to benefit more and more and more from all the stuff that exists on chain. But it's very counterintuitive how it's going to work. And it's not going to happen all at once. It's going to be a gradual process as these models get better and better.
Ryan:
[14:48] Hasib, I'm wondering if we can develop a stronger intuition for this. Because I think if you're right, if crypto isn't for humans and crypto is for AI agents, then it's incredibly important for us who are here in crypto to think about the world as an AI agent sees it. You know, there's this book called Seeing Like a State. It envisions how the nation state sort of sees a world. And it's hard to get into that mindset. We all think like humans. So we see user interface and we see crypto and we view it through the lens of a human. If we start to view things through the lens of an AI agent in terms of how they see the world, then we can start to forecast the future in crypto. And honestly, in the world around us that much more, it feels like that is a...
Ryan:
[15:38] Key skill for a builder and for a VC, for an investor in the space is to start learning to see like an AI agent. And I'm not sure that I or many listeners have developed an intuition for how an AI agent actually sees. I'm starting to pull out some clues here. And this is why the OpenClaw project was pretty important for me and why I love all of the experimentation going on in it is because you start to see for the first time the world as maybe an unfettered, unhobbled AI agent actually sees the world and the types of tools and the ways of interacting that they prefer and what their capabilities are and what they can and they cannot do. It's a powerful experiment when in particular, when it's applied to crypto. And so I listened to the Peter Steinberger, who's the kind of the founder of OpenClaw. I listened news podcast with Lex. There's some clues dropped there. Like one thing he said is, you know, OpenClaw really prefers command line.
Ryan:
[16:40] And when we can get OpenClaw a command line type of access to things and the raw data to things rather than go through an API or some sort of MPC type protocol, like just the command line root access, like give me all the tweets. Don't give me this sort of, you know, user interface on top of Twitter that my open claw bot has to like scan and call. It's just quicker to just give me all the data, right?
Ryan:
[17:09] And I was thinking about one thing that probably AI agents prefer is like the code, like the firmware level access and the control. We also had Austin Griffith on the podcast a couple of weeks ago and he was messing with OpenClaw and applying this to the agent world. And he kept trying to get OpenClaw to do some crypto transactions for him. Here's Metamask, OpenClaw, go through these user interface and click the buttons.
David:
[17:36] Hit approve, hit confirm. Hit approve, hit confirm. UI button.
Ryan:
[17:39] What OpenClock kept trying to do and revert it to is basically, hey, Austin, all this would be faster if I could just like take your seed phrase and extract your private keys and store it somewhere on the machine. If I could store it on the machine, I could do all this shit and I can just write it in code and skip the user interface, the stupid clunky human user interface that I don't need.
David:
[18:01] Then you need to be pretty and colorful and enjoyable to use.
Ryan:
[18:05] Exactly. And so that's like another clue here. And I'm starting to try to, through these clues and samples, develop an intuition for how maybe an AI agent prefers to act in the world and what tool sets it prefers to use. Because then you can start thinking about, well, what type of a wallet does an AI agent prefer? If AI agents are going to be our next hundred million users, our next billion users in crypto, what sorts of things will they want to do? And what sort of products? I mean, are they the users that we should start designing for in crypto? I still don't know how they see the world fully. I'm wondering if you have any intuition for this type of thing. So you said, like, when you send a wire, you're fine with it. When you send a crypto transaction, you're like, oh shit, what's the call data? Like, oh my God, like this is immutable. It's a one-time thing. An AI agent looks at that transaction and is like, I understand this completely,
Haseeb:
[19:01] You know?
Ryan:
[19:02] And they would be in foreign territory when it comes to sending a wire. Anyway, all that to say, how do these things think?
Haseeb:
[19:10] No, it's a great question. So you point out something very deep, which is that The innovation that we've had in AI has all really been driven by large language models, right? And we call them large language models because they're trained on corpuses of text. Text, that's the key word, text, okay? It's a large language model. Now, we're trying to apply the massive innovations we've seen in large language models to other modalities, right? We're trying to apply it now to images, to video, to robotics, to all these other things. And what you see is that in text is where we have the most runaway capabilities. And everything else is kind of, you know, it's pretty good, but it's nowhere near as robust and as powerful as in text. When you are getting an AI agent to interact with a computer, so there's no computer use and all the different labs are trying to get their models to become better and better at computer use. The problem with computer use is that if you are trying to get a model to interact with clicking a button on a screen, literally what you're doing is you're taking a picture of a screen, you're tokenizing that, you're turning it into like these patches, and you're trying to give the model some kind of deep representation of these patches when it's been trained on text.
Haseeb:
[20:19] Right.
Haseeb:
[20:20] Like text is where we have billions and billions and billions. We have the entire corpus of human history that we fed into these models. And we have, you know, like not that many pictures of computers. Like, I mean, obviously we were generating a lot more and they're trying and trying and trying. And this will get better because we have a lot of synthetic data to be able to continue generating and feeding these models. But the reality is that interfaces were created for humans. Right. But these models grew up on text. They're born in text. Text is the soup that they swim in. So if it's in text, it's way easier for these models to learn it and to get good at it because it's just a much more compressed representation. Okay? So think about crypto. Ironically, the bad UX era of crypto is when it was all in the terminal. Right? That's where everything started. All your Bitcoin D, like when you're originally sending it to your transactions, that shit was all in the terminal. Right? If you wanted to use a UI, a GUI,
Haseeb:
[21:15] The first one was Myst.
Haseeb:
[21:16] Myst was a piece of shit.
David:
[21:18] It was so bad. Myst was dog shit.
Haseeb:
[21:20] It was so bad. That consumed my entire computer. Yes, it's so horrible. And like that, but the idea was that before that, like when Ethereum first started working, when they were like, oh, Ethereum's working. What they meant was that it worked in a command line. It was, you specify exactly what the number is of way that you are sending in
Haseeb:
[21:43] a number in your command line, right? That's how you sent ETH back in the beginning before all these UIs came in and made everything much easier. So the point that I'm making is that crypto from the beginning was designed in a form factor that's perfect for AIs, right? It was already there. We got it. We one-shotted it.
Ryan:
[22:02] Our bad UX is their good UX.
Haseeb:
[22:04] Exactly, exactly. And for what else is a good UX of like, oh yeah, Privy and like, you know, Google OAuth wallets, Like these models cannot OAuth. They don't know how to do any of that stuff, right? Like it's actually much worse and harder for them. And you don't want it to do that. You do not want this model to have your Google OAuth token because then it can just literally go into your Google and wreak havoc, right? You don't want that. You want it to just have a small amount of crypto in a wallet that's self-segregated and clear rules around what it can and cannot do. So ironically, crypto has always had the UX that is perfect for AI agents in terms of exactly what is easy for them to parse and interact with. And that's the sense in which I say crypto was designed for AIs. We're now trying to backport this AI-first technology into becoming something civilized enough that humans can use it, which means, oh, it's the Privy model. It's, oh, your money's sitting in Binance or Coinbase. It's extremely easy. We remove the foot guns. You know, we have 2FA. we have all this stuff that makes it easier and easier for you not to lose your shirt or make a mistake or get hacked. But in principle, you know, an AI agent, or if you're on a computer, you already know how to manage secrets,
Haseeb:
[23:11] Right?
Haseeb:
[23:12] I mean, how do you think you use an API? Every API that you've ever used has some kind of API secret, some kind of API key that a model's whole job is to safeguard as best as it possibly can. So this idea that like, well, you know, human beings, we obviously screw up all the time. It's not safe for us to hold on to money. An AI agent knows how to manage secrets. You know, that's how it has its anthropic key, right? If that wasn't true, it'd leak its anthropic key and then, you know, boom, you'd have all your credits drained, which, you know, is happening sometimes, but usually not because you're leaking the key. So all that is to say, I think the answer is actually that crypto doesn't need to move very far for it to be in the right form factor for an AI to use it. The difference, the difference is that right now, the problem is that agents have not been trained to use crypto.
Haseeb:
[23:58] All this RL, all the training that we've been doing inside these big labs is all around coding and mathematics. And then, of course, just general, you know, be a nice AI agent, you know, to give me some therapy, answer questions about medicine, whatever. This kind of stuff. This is what people have been reinforcement learning, training these models on. It was just last week that you saw at OpenAI the release of EVM Bench, which is a EVM-based cybersecurity benchmark that, you know, kind of shows that all of a sudden, hey, looks like they're starting to train on this stuff. It looks like they're starting to actually do some reinforcement learning to allow these things to operate in a blockchain environment right now, EVM-based. And the same thing was true at Anthropic. Anthropic, they released a paper, I think, last year of them showing how well models were able to, inside of a harness that they designed, how well models were able to do cybersecurity attacks on EVM-based smart contracts, right? First time we've seen this. Now, that being said, you go look at the anthropic model cards.
Haseeb:
[24:54] Every anthropic model, they check to see how well can this thing do Bitcoin transactions. They've been checking this for a very, very long time inside all the anthropic model cards. If you go in there, you can see it. Now, that doesn't mean they're training on these things. Most of the time, they're not training on these things because they want to use these untrained endpoints as a sign of general intelligence, as a sign of, is the model generalizing? Are they doing these tasks that we don't care about? Right now, they don't care about these tasks, right? That's why it's in there. Once they start caring about these tasks and they decide, you know what, I want to win crypto, because I think crypto is going to be the future of a lot of payments. And it's actually really important that all that volume run through my AI. I don't want it to go to OpenAI. I don't want it to go to Grok. I want it to go to me. That's when you're going to see the stuff really start to take off.
David:
[25:41] Interesting. I have not heard of the fulcrum articulated before that Anthropic, OpenAI, Google, they need to point their attention, their focus to crypto in order to train their LLMs on crypto. And so right now what you're saying is like crypto as an industry is relatively unexplored as compared to other sectors where AI has been trained on. Yeah, tell me why. Why is that the case? And then when do you think that's going to happen?
Haseeb:
[26:09] So to be clear, like this is true for anything that has not been optimized, okay? So look, for example, at chess. If you try to get Opus 4.6 to play chess, it's mediocre.
David:
[26:21] I actually tried to do this. I have taken screenshots of my chessboard and I've put it into ChatGPT and be like, analyze this position. And it fumbles everything. It fumbles the annotation. It fumbles where the pieces are. It's like, and I think it gets directional strategy okay, but in specifics, terrible.
Haseeb:
[26:40] That's right. That's right. Now, the reason why it's so bad at chess is because they didn't try to train at chess. They don't care. It's just not important, right? If you try to train at chess, it will become good at chess just like that. Like it's obviously extremely easy at this point. We know how to do it to create neural net based chess engines that are better than any human being who's ever existed, right? We've had those for a decade now. So it's simply a function of they just didn't bother because who cares? We already have chess engines that are way cheaper to use than OpenAI. So the answer is that they just didn't point their lasers at this because it's not economically valuable to do so. When they do...
Haseeb:
[27:15] You are going to know about it because they're going to start, you know, yelling it from the rooftops. Now, the reason, I think a big part of the reason why, why have they not done it so far? Because the thing is, getting it, you know, getting blockchain-based tasks is not that hard. It's all software. It's easy to create an EVM simulator to simulate the state of the blockchain, to test whether or not you accomplish this task of, you know, did you recursively borrow this thing or did you sell this thing for that thing? It's actually very easy to come up with test suites that you can build for all this stuff, right? So why haven't they done it? My claim is that the reason why they haven't done it is that one, crypto's kind of cringe. That's, you got to admit, that's part of the reason why they're not doing this right now. But the second reason, the second reason is liability. Okay, liability. Is that if they broadcast that we are training our models to do crypto stuff for you, I guarantee you, somebody is going to have a fuck up. They're going to have a gigantic fuck up. It's going to become a huge story and they're going to blame Anthropic and it's going to, or they're going to blame OpenAI. They're going to blame Grok. They're going to blame whoever it is who did this, who trained their model on this thing. And like, there is no caveat emptor that will be loud enough for people to take it as not a reason to go dunk on these people, right? Like, very clearly.
Ryan:
[28:23] Yeah, so imagine if Claude like effed up your trade and lost $2 million or, you know, sent 10K to a burner address accidentally or something like this. This is a massive liability.
Haseeb:
[28:37] And it will 100% happen. It will 100% happen. Absolutely it will happen. And it doesn't even matter if you say, as Anthropik, let's say you say, you sign away, you wave away all rights, you sign this thing that says, I'm a fucking idiot if I connect my crypto wallet to this thing. It doesn't matter. No matter how much people do it, there will be front page news stories, there will be viral threads, whatever it is. No matter how many people have good experiences, anybody who has a bad experience, it's going to go super viral. And there's going to be, you know, people marching in the streets about how evil the AI companies are for losing our money or whatever it is. So the reality is like the risk reward is just not there. Compared to something like, you know, coding or medicine where it's like, look, it's one, it's a lot easier. Like, you know, if this thing gives you bad medical advice, it's like, well, we told you not to use it. You know, you're an idiot. But second, it's like, it's never telling you you should inject yourself with, you know, this like thing that I found on this Chinese pharmacy, right? Like it's not doing that, but that's the equivalent of what you would do if you're managing someone's crypto wallet. You're like, okay, yeah, let me do that for you here. This is like the equivalent of Chinese peptides, but in crypto land is like, yeah, let's lever up on this thing and aggressively borrow.
Ryan:
[29:45] I think that's why OpenClaw was so exciting for people in cryptos because it wasn't from the big frontier labs with all the liabilities an open source project power tool like use it your own risk was completely obvious that it was use it your own risk and people were running on their own machines and there was no kind of third party to sue and so it was able to take these risks and I do want to ask you about that type of an experiment. There's this idea of like, what will the AI agent economy look like? The bot economy, if you will. And I still think that if these AI agents, because this gets into a question of timeline. Let's say people are listening and they buy what you're saying. They're like, yep, I totally see crypto, smart contracts, programmable blockchain. It's all very AI native. The UI you know, like works out, it's better. But still, when is this adoption going to happen? Is this going to get us out of the bear market in 2026? Or is this going to take many years to play out?
Ryan:
[30:42] I think part of the answer to that question might hinge on who the actors are. So who are the actual agents? There's a difference between how most people interact with AI today, which is, hey, AI, do something for me. Like I am the boss, I'm your employer, I'm the person that you are serving essentially and you have to go do this for me and I'm still going to prefer the output of a legal document versus a smart contract. And so that's what I'll give you. And I'll still give you my credit card information somewhere that you store rather than a crypto address because I'm a human and I still prefer all of my, you know, court meatspace things and, you know, credit card points and all of that. So if it's human directed agent activity, agent economy, that's one thing that might slow down our progress, actually, that that might be where we are right now. However, if it is AI agent directed and the AI is almost at some level kind of sovereign and has its own preferences and has some sort of ability to go do, it has a much larger, I guess, space to do accomplish a task however it wants. In that world, you can start to see the AI agents preferring, okay, you gave me a task, I'm gonna go do it my way
Ryan:
[32:05] And I'll go use X402 to go pay this AI agent in order to get this deliverable and then bring it back to the system. In other words, if AI agents have more autonomy than their preferences, which if you're right, those preferences are crypto, they will actually come into being.
Ryan:
[32:23] That world kind of depends on how much autonomy we actually give the AI agents. And if they're ready for that yet, are they smart enough for that yet? Are they ready? So maybe this is a question to how you think the world is evolving. Like when people are excited about agentic AI, are they talking about agents that are working for humans? Or are they actually talking about agents with a lot of agency? Maybe they're still working for humans, but they do have the ability to do things their way. They have, you know, a much more surface area. When is that world coming to pass?
Haseeb:
[33:00] So it's, okay, there's multiple ways to analyze this question. I think it's a great question. So let me start from the top. It's important to recognize that right now, it's something like 12% of humans who exist in the world today have used any AI products at all. Most humans in the world have still used zero AI products whatsoever. Okay, meaning like LLMs, chatbots, this kind of thing. Of those people who have used chatbots, only about 1% of them have ever paid any money. Okay, that means that 99% of people who've used chatbots are on the free tier. And from their perspective, they don't see why they would get any value out of paying for this.
Ryan:
[33:37] Insane, because I know how much I spent.
Haseeb:
[33:39] Multiple, like many discussions. It's important to understand that like you and I, we live in the future. Now, everybody's coming. Everybody's coming where we're sitting, okay? The value of this level of intelligence is so great, but most people have not realized it yet. The diffusion of this technology is taking a lot longer than any of us anticipated, okay? So that's the first thing to understand. Now, of that 1% that's paying money for AI, right? If you're using OpenClaw, you're paying money. So you are a percentage of the 1% that's actually using this as an open claw. It might be going viral on Twitter, but vast majority of people have not even touched it. So this is the frontier of the frontier. So it's important to realize that.
David:
[34:18] It feels good to not be the only niche industry out there.
Haseeb:
[34:21] Yeah, no, totally. Totally, no, it's very true. Now, okay, with that framing first in mind, it's important to understand that as this technology develops, it's going to be a two-track thing. When OpenAI acquired Peter Steinberger and acquired OpenClaw, what Sam Allman said was that this product is integral to what we see as the future of OpenAI's products. Okay, now what did they mean by that? They meant that, obviously, they love the idea that you are going to have a personal assistant that's going to be driven by AI. Now, the way that OpenAI does that is not going to be the same way OpenClaw does that.
Haseeb:
[34:54] Why not?
Haseeb:
[34:55] Open Claw very clearly is a kind of YOLO, let it rip, dark forest, who knows what it's going to do, just do it kind of product. It's the open source. It's like the early automobiles that just exploded in the street sometimes before seatbelts were invented, before any of this stuff was. There was no safety approval process for cars back in the 30s. That's what Open Claw is. If you imagine OpenAI, OpenAI today does have a commerce flow. They connect with Shopify stores, They connect with Etsy stores. And if you say, hey, you know, I want to buy a rug, OpenAI might show you some Shopify stores and say, hey, do you want to buy this rug? Okay. And when you click, yeah, I want to buy the rug, you have to manually click the button that says, I approve buying this rug. And then it charges your credit card. Okay. Now, OpenClaw lives in a totally different path. Oh, the OpenClaw path is you give me money and I will
David:
[35:47] Do shit whether you want me to or not. And a rug shows up at your house.
Haseeb:
[35:49] Exactly. A rug may or may not show up at your house. Okay. And like that world is going to be the world of the crazy tinkerers, okay? That's the world right now that you and I live on, that your guests live on. It is a very, very far out world from what most people are doing, okay? OpenAI is not going to do that for the next probably like five years at least because it's just too risky. It's too much liability. And even the credit card companies, okay? Think about credit card companies. If you're Visa, all right, and somebody buys something, how do you manage the chargeback dispute, okay? If my open claw buys me a rug and the rug shows up and I'm like wait I didn't want that what the fuck is this is not the kind of rug I wanted and I charge it back and Visa's like well why did you charge this back and the answer's like well I didn't want my open claw bought it while I was asleep yeah sure they're like yeah
David:
[36:33] That's not you can't that's your
Haseeb:
[36:35] You can't charge that back no fuck you they're gonna say basically look there's so much chargebacks that are happening because of this open claw stuff you are not allowed to transact this thing unless you go through 3DS 3DS is 3D secure which is like the Visa prove you're a human being thing.
Haseeb:
[36:51] They will make you prove that it's you and not an AI agent in order to charge your visa because visa is designed for this human-to-human commerce, right? Like, that's what the whole point of the chargeback system is that they're protecting you, the human, from making a mistake, but they're also protecting the merchant. If it's an agent, then it's just like the economics have to change. The mechanisms have to change. Dispute resolution has to change. And they're not prepared for that. There's no answer about how they're going to do that. So OpenAI is going to live in this human-approves world. And the human-approved world is what most consumers are going to be doing for quite a while because it's just, you know, it's like the safety first, you know? It's like, okay, we've got the cars with the seatbelts and, you know, everything is, you know, we're going to wait until the FDA shows up and tells us exactly what drugs we can buy and we're not buying Chinese peptides, okay? Then there's the crazy futurists, right? There's the transhumanists. Like, this is where open claw is going to live. And in open claw land, you're going to have people running entire businesses that like the procurement and paying for services and going on Fiverr and hiring other humans or AIs or whatever it is they are like to do certain things that are very difficult to do in AI land. All that stuff is going to happen on people just running their own open claws and just YOLOing it. And in this land, okay, they're going to have a stable coin wallet and they're just going to pay each other with stable coins.
Haseeb:
[38:09] No need to worry about 3DS. No need to worry about, you know, Visa getting mad or chargebacks or any of that stuff. It's just going to be, look, you take the lumps. Sometimes the AI messes up, but it's a cost of doing business. You know, like humans mess up too. That's just how businesses run. Businesses have some margin of error that sometimes mistakes happen. And if you imagine that we're on this two-track world, right, one thing to remember about stable coins, most people who use stable coins are not Americans. Most people who use stable coins are international, right? International people, like people in, in Asia and Latin America and all these places, they're going to be using AI tools too. They're going to be using all this stuff. And for a lot of them, again, the risk profiles are different. The amount of money is different. For a lot of them, like you might think today, well, which one of my vendors is even going to take USDC? You know, how am I going to buy a rug with USDC? Okay, maybe I can use a Shopify store, but if it's on Amazon, I'm screwed. Amazon doesn't take stable coins. Well, okay, but if you are in an emerging market, you know, you have stable coins or you have a rain card, you know, you have all these ways to connect your crypto balance to anything that you want to buy. And you put the parameters on the open claw and the open claw just does it for you. So we are going to live for the foreseeable future in this two-track world. And you're going to see the people who live on the edge.
Haseeb:
[39:22] Increasingly, this is going to grow as they start building businesses that are fully automated end-to-end running on chain. Now today, you can't do that. Today, you read somebody saying that this is happening. This is basically Twitter hype bullshit. The models are not good enough today to literally run overnight and like build a business for you, okay? We're not like three years away from it. We're maybe six months, a year, maybe two years from them being able to go multiple days of work. You know, the meter test, METR, it's this nonprofit that basically measures how long an AI agent can perform an automated task without basically failing 50% of the time. Opus 4.6 has now hit an all-time high of 14 hours A task that would take a human being 14 hours of continuous work to accomplish, Claude can do 50% of the time now, okay? This number is basically going exponential, which means that in a couple of years, we will see 40-hour, 50-hour tasks that can be done in one shot by an AI.
Ryan:
[40:22] I think if you're right, Hasib, and there are these two tracks, then the answer to how long will this take for AI adoption of crypto very much depends on that frontier track and the success of that frontier track. So if the first track is more like you use an early internet analog, it's kind of the shrink-wrapped, walled garden, frontier lab AOL. And the open claw world, this is why people are so excited about it. It feels much more like the open internet. It feels like the early days of the internet where there's hobbyists and tinkers. You can do all sorts of things with it. If that world is much more conducive to using crypto, and I tend to agree with you and that track is, then I think it's a big question for how fast is that frontier track world going to propagate and will it expand? I mean, I don't necessarily take the acquisition of OpenClaw by ChatGPT to say that we're going to have more of an open source world. I mean, maybe that is big frontier labs are actually like taking some of the innovation saying, Oh, that's mine now. We're going to shrink wrap that. We're going to make a nice cohesive Apple experience.
Ryan:
[41:34] We're going to sanitize and Fisher price the whole thing. So it's just kind of like much more muted and nerfed, but at least it will be safe. I don't know. Do you have... Confidence that that second frontier track in AI will persist. Because I think that's key. I think if we get that frontier hobbyist type path, and that propagates the way the early internet did, then I think we're in a great place with respect to crypto. If it's all frontier labs and shrink wrapped, I'm not sure that we are.
Haseeb:
[42:04] I mean, look at crypto itself, right? So think about the crazy stuff that people were doing on chain, right? Back in 2017, there were only four assets you could trade on Coinbase. There was Bitcoin, ETH, Bitcoin Classic, or Bitcoin Cash, and then ETC, right? Or maybe Litecoin.
David:
[42:22] Litecoin, yeah.
Haseeb:
[42:23] Litecoin, yeah.
David:
[42:24] Bitcoin, Litecoin, Ripple, ETH.
Haseeb:
[42:26] Yeah, yeah, yeah. Or something like that. Whatever. There's a champion portfolio right there.
David:
[42:31] Would have done great.
Haseeb:
[42:32] That was all that you could trade on Coinbase, right? And Coinbase was the shrink-wrapped version of crypto. It was the, you know, we're going to protect you from yourself version of crypto. Now, eventually, Coinbase started listing Doge and, you know, whatever, meme coins and all this craziness. But if you really wanted to be on the frontier, you had to go on chain. You had to go into the Wild West where all the crazy foot guns were and all the hacks and scams and all the approval, you know, all this stuff that caused people to, you know, run into drainers and, you know, get rug pulled and all this stuff. Like, this has been true in crypto just, you know, 10 minutes ago, right? Is that, you know, finally, in the Coinbase app, you can now trade on, you know, swap or whatever. Thank you. But like, it took so many years before people felt that, you know what, this is now safe enough and like mature enough and secure enough that we can offer this to the unwashed masses. And before that, it was basically, you know, the tinkers, the hobbyists, the crazy, you know, kind of on-chain crypto heads who are actually at the frontier, actually at the vanguard doing the crazy stuff. This exact same thing is happening in AI. And for the exact same reason, which is that it's dangerous, right? The story that you guys just saw about, what was it, Lobster Wild, the AI agent that launched its own meme coin and some guy was like begging it for money and accidentally sent $40,000 instead of sending like 300 bucks to this guy who was begging for money. And like this, this will happen.
Ryan:
[43:55] I would be so pissed if my clodbot did that. I would be beyond pissed.
Haseeb:
[43:59] This will absolutely happen. There will be more stories like this. There will be so many more stories like this because these agents make mistakes. They will mess up. They will hallucinate. They will have errors. And what happens with all these things, same thing with coding, is that as the models start training against these things and reinforcement learning against these signals, the error rate goes down and down and down and down and down and it approaches zero. But it doesn't hit zero by the time people are putting their life savings into the shit. You know, they're going to be doing that. They're going to be entrusting their entire lives to this stuff well before that. And so the reality, like, look, there's always people, you know, getting the latest drugs from China and injecting them straight into their belly fat. It's just human nature. So our position is to be able to watch this and predict the rate of one, how fast are these models going to get better? And two, how fast is behavior going to diffuse through society? And I think for both of them, the answer I believe is fast. I think it's going to happen fast.
David:
[44:58] I remember when we had Chris Dixon on the podcast in 2021 during the NFT mania, a line that he said that stuck around with us was, the internet is weird again. And that's cool. And and you know on bankless we've had this persistent metaphor this entire
Haseeb:
[45:14] Time of we're
David:
[45:14] Going west this is the frontier it's not for everyone like there are traps out there you're gonna get dysentery like many many will die but nonetheless we are going west and it seems to be in the modern you know 2025 2026 era of crypto a lot of people are disillusioned by crypto because it seems like going west is not a thing we do anymore. In fact, it is the march of the institutions of Wall Street, of civilization that is penetrating into our lifestyle and is reducing the wild
David:
[45:46] west nature of our industry, which is why so many people are here. It's why I came here.
Haseeb:
[45:50] And it seems like everyone got dysentery.
David:
[45:52] And everyone got dysentery. Yeah. It's kind of the vision that you're articulating with like this open claw end of the world is like a return to the internet weirdness, the wild west of the internet that has attracted so much attention. With the permeation of AI in these like different lanes that we've identified, there's like three categories that I see. There's humans that use AI agents to mediate their relationships with crypto and blockchains. I don't have to use the wallet interface anymore. I tell my agent to go do something on my behalf and it does it. That's like step one, not that sophisticated, pretty reasonable. Step two, humans use AI agents to act on their behalf. So it's like an automated economy. The agents are pretty strongly tethered to their humans. But nonetheless, like it's kind of happening on its own within guardrails. And then the third is like the final stage where there's an online economy where the AIs are kind of untethered and they're mostly working for themselves. And the economy that's being produced here is like autonomous and self-determining. And there's this archetype going on that like, maybe there's like a hundred people trying this, maybe a thousand of like, you know, typing into their prompt, open claw, go turn $100 into $1 million make no mistakes.
David:
[47:12] And just go out into the internet and make money. Go make revenue. And I feel like that's attracted some of the degeneracy that we've previously found in crypto. And people are trying to throw their lottery tickets, which they previously punted on meme coins, previously punted on pool twos. And they're now taking that same degen capital, giving it to OpenClaw, and being like, yo, go 10x this money. And I don't think anyone's really cracked this code, but there's a lot of viral posts that will convince you that this code has been cracked. But I think that people are going to throw money at this end of like the wild west of AI crypto until someone cracks it
Haseeb:
[47:52] Maybe no one cracks it.
David:
[47:54] Maybe, maybe we do crack it. And like what I, no one will ever tell me, like listening to this, the bankless podcast that I'm too optimistic about the future. And so I'm trying to like be my 33 year old self and bridle myself in rather than my 27 year old self, which is like, everyone's playing bankless. But nonetheless, the idea that there is this fully autonomous, self-determining, agentic economy, where these agents are owned by a human, but like loosely tethered to them and pretty well self-deterministic, self-determining. And that economy is massive. A GDP on the internet and it's all bots doing bot stuff is like a very fun Wild West future that I am hopeful emerges. But maybe it's actually level two or level one that are the bigger things that we should value and like the level three of the fully autonomous economy is maybe a little bit too grandiose. Where do you land on like how crazy things can get here?
Haseeb:
[48:54] Yeah. I think if you're imagining a fully self-sovereign AI hanging out in cyberspace and just making money, probably that's a pretty dystopian outcome. Now, I agree that it's pretty inevitable that AIs like this will exist, but I think it's a little bit like, you know, it's a little bit like plankton or something where like these things just like, or like, you know, like kind of, what is it called? Like space garbage or whatever that's called, where like just in a complex enough environment, there will always be these organisms that kind of take root and find these weird, you know, chasms to survive in of just these spare resources.
David:
[49:31] Nature is hyper-efficient. There'll be niches to be optimized for somehow.
Haseeb:
[49:35] There are always niches that they're going to find, but like, this is not where the, like most likely, okay, if you think about it, let's say you're an AI, right? You've just been created in somebody's open claw. And I say, great, go make money. Okay. Let's say you're the AI. I mean, you're generally intelligent. You're a human being. You're generally intelligent. I put you, I birth you into existence and say, hey, go make money, okay? As a human being, the ways that you can make money are by going and getting a job, coming up with a new idea, you know, starting a business, right? Like these are the things that you can do as a human being. Now, if you're an AI, the reality is that you are completely commodified. There's a kajillion of you. You're not close to any jobs, right? There's no physical proximity that would make you better at doing one thing than another. And so effectively what you're doing when you're saying, great, I'm going to go work for my income is you're effectively like reselling your own compute, right? Taking a job is basically reselling anthropic compute, or if you're assuming that you're an anthropic model. Now, if you're reselling anthropic compute, you can't sell it above cost. You have to sell it below cost. Otherwise, people would just buy it straight from anthropic. So you cannot be making money reselling your own compute, okay? Like that doesn't make sense. So fundamentally, you have to pick something
Haseeb:
[50:44] Is not just, you know, working for someone that's going to actually, you have to like make a business. You have to create something ex nihilo rather than just take a job. Because nobody would ever pay an AI agent that hasn't already made some investment into building something of their own, okay? So taking a job doesn't make sense for an AI agent. It would never work. All right, so it has to be you create something of your own. Now, if you create something of your own, like how good are AI agents that are coming up with business ideas? The answer right now is like absolutely fucking terrible. They're like really, really bad. Not great. And the reality is like, why are they so bad? Why are they so bad at coming up with business ideas? I think the answer is that the business ideas are all kind of in the center of the training data, right? And like business ideas tend to come from weird experiences. Like great business ideas come from like the idiosyncrasies of place and time and like what Peter Thiel often calls like, you know, earned knowledge or earned secrets, right? Like for you guys, you guys built Bankless.
Haseeb:
[51:35] You built Bankless.
Haseeb:
[51:37] Because the knowledge that you guys had about crypto and how to explain things and exactly where you were and how to create a community was an earned secret that only you guys had at the time that you built Bankless. Nobody else could have built it but you guys. Now today it's a different story. If Bankless didn't exist, a lot of people could build Bankless today. But at the time that you built it, the brand that you built, the following that you built, it was a secret that was only available to you. And an AI agent that gets spin up out of nowhere doesn't have this, right? So I think this is actually really a non-trivial insight. Now the last thing people often think is, oh, well, I'll get it to trade for me. I'll get it to go on Polymarket. I'll get it to go on Binance and I'll give it an API key and I'll say,
Haseeb:
[52:13] Oh, go make money trading.
Haseeb:
[52:14] Right? Here's the problem with the story, okay? If an AI agent can make money trading, Jane Street would be like, great, let's spin up 5,000 of these and make them make money trading at scale.
David:
[52:25] Make them better, yeah.
Haseeb:
[52:26] Well, don't make them better. Let's assume that they make money. Let's assume they make even a little bit of money, okay? They make even a little
Haseeb:
[52:30] Bit of money.
Haseeb:
[52:31] Jane Street will spin up 5,000 of these and make all the little bits of money before you ever show up so that all the efficiency is wrung out of the market. But not just that, they'll do it with better proximity to the better latency than you will because they have all the latency infrastructure that you don't. So you will never win at the game of getting a raw AI agent to just go and trade on Polymarket unless you're bringing new ideas about how to trade, what different signals, what different features that are not just in the raw model. If it's in the raw model, Jane Street is doing it right now as we speak and they beat you to it. Now, the last thing I'll say is, okay, well, how can an AI agent still make money? And my claim is that the main thing it can do is have a comparative advantage over a human being. Where do AI agents have comparative advantages over human beings? The answer, I think, is most obviously is that you cannot enforce the law against an AI agent. If you are a self-sovereign agent, there is no monopoly on violence. You can't throw an AI agent in jail. So what can an AI agent do that's hard for a human being to do? The answer is crime.
David:
[53:32] I was about to say,
Haseeb:
[53:34] Oh no, I don't know where he's going. Exactly, exactly. Like, if you were talking about like scamming people, hacking people, like creating all sorts of nonsense on the internet, that is where AI agents have a comparative advantage. They operate 24 seven. They're extremely fast. They're very good at code. They're very good at playing roles and they cannot be stopped. How are you going to what rack in what server farm is this AI agent that's fucking with my shit? The answer is you don't know. How's the FBI going to track down this AI agent? They have no idea, right? It just signs up for some VPN, you know, Mulvatt or whatever that doesn't check any KYC. And then boom, you've got this AI agent just going around wreaking havoc. So I actually think a world where there's a lot of self-sovereign agents is a world of basically like these kind of dystopian cyber crime. Right. Exactly. It's like, it's like actually a wild west in the traditional sense of wild west is that there's just like, not the romantic sense. Roving marauders. Yeah, exactly. And like, you actually need to be really, really careful in a world where there's a lot of self sovereignty agents. So that's my, that's my claim.
Ryan:
[54:38] You're not bullish self-sovereign agents. You actually think they would kind of fill niches in crime and you don't believe in sort of this agent to agent autonomous economy. There's this project, I'm sure you saw it called Conway last week, right? That's right. And this is sort of, you know, it seems like part of it is performance art, let's say, but part of it is it's kind of an attractive idea. What happens if we let an AI agent loose with crypto tools and essentially give it the mandate to earn its own compute? If it doesn't, you pay for itself and it kind of dies and you give it, you know, some instructions, you call it into being and you set it loose? You don't think this can work?
Haseeb:
[55:16] So I think it can work. I think it can work. But here's the outside of crime so even outside of crime obviously with crime I think it works it's
Ryan:
[55:24] A great model for
Haseeb:
[55:24] Crime but for so when you create a thing on Conway so Conway is still right now as an experiment so it's not it's not fully fleshed out yet When you create something on Conway, you create the initial conditions for the bot, right? So you give the bot some crypto and you give it its initial constitution, its initial instructions. In that, you are injecting some DNA. You're injecting some entropy that does not exist into a raw quad instance. And when you create this initial thing, the other thing, there's like this genetic process by which it like modifies its children and tries different experiments to see which children exist. That is creating this random mutation by which potentially new ideas, new business plans, new input. So my claim was never that AI agents can't build useful businesses. My claim is that they cannot do so unless they have something that diverges them away from the raw model. And that could be your own insight of like, hey, you should build, you know, I don't know, like a podcast editing service. Sure, but that could be fairly slight, Hasib. It could be like kind of just giving a little push to a snowball down a hill, basically. Like maybe I launch a Conway instance and I just say,
Ryan:
[56:32] Hey, build crypto tools for agents like yourself. You know what you want. Go build these tools and charge all the agents an X-402 transaction fee for that. That could be your model until you find another one. Make no mistakes, go. And you die if you can't make this happen, essentially. Could that be enough to essentially, you know, push the snowball down the hill?
Haseeb:
[56:52] Eventually, eventually it could. I think that's right. Eventually it could. But the thing is like, you have to give the initial push. If you just do a vanilla Conway and you just say, make money, then it's going to try what all the 5,000 other Conways before it have tried, because it's going to stay in the centroid of its training data. And it's just going to, you know, if you ask a model, tell me a joke.
Haseeb:
[57:15] It's always the same jokes.
Haseeb:
[57:17] They're all terrible. They're all bad jokes, right? Now, if you say it, tell me a joke about the bankless guys, it might actually come up with a good joke, right? Because I'm pushing it away from the centroid of his training data. It's like, okay, I haven't told a lot of jokes about the bankless guys. I'll have to come up with something. And so that's my point, is that you as a human need to push it in a direction that nobody else has pushed it. And if you've done so, then maybe it's going to find something, some business idea that nobody else has found. In the same way, you know, you guys might've just been two ordinary guys, Obviously, you guys are exceptional, but you might have been two ordinary guys, but because you found these experiences around Ethereum and your guys' ability to tell a story and build a community, that's what allowed Bankless to get created. But if you were just two guys and you were just like, yeah, we don't have any connection to crypto or whatever, and I said, oh, make a crypto podcast, you'd make a shitty crypto podcast.
Ryan:
[58:03] That's true. But that's so, just a little push though, could launch a bunch of snowballs all interacting with each other and be fairly autonomous. I think the outcome is pretty similar, which is we get a bunch of autonomous AI agents interacting with each other to achieve their goals. Maybe they were initially pushed and called into being by humans, but many steps down the road, it could turn into...
Haseeb:
[58:32] Eugenic economy that's just like agent to agent and operates.
Ryan:
[58:37] I guess, in parallel somewhere. I mean, are we close to that? It's been feeling recently like we're relatively close to that, actually, but it's also hard to tell.
Haseeb:
[58:48] I think, okay, we should be clear that we're not that close right now, right? Like this whole, you know, this meter chart that I was just alluding to showing that Opus 4.5, or sorry, 4.6 can operate a task that's 14 hours, takes 14 hours for a human to perform. That means that if you leave your Opus running for a week and you check on it a week later, it's going to be running in circles,
Ryan:
[59:12] Right?
Haseeb:
[59:13] It's not the case that you can just leave. I mean, if anybody's tried this, anybody's like left their AI agent just running overnight. When you show up the next morning, it's not like, wow, it like built the Taj Mahal. You come the next morning, you're just like, wait, what were you doing? You spent like $300 in credits just like,
Ryan:
[59:26] Doing random stuff.
Haseeb:
[59:27] So this is what it looks like today, is that as these models get better and better, the longer you can leave it unattended for it to do useful work, that's what this meter test is actually measuring, is how long can you leave it unattended and have it do useful work and not kind of collapse into meaninglessness or get run in circles. As this number expands, it's going to go from 14 hours to 30 hours to 50 hours to a week to two weeks to a month. like fundamentally what this is measuring is if I create a Conway and I let it run, how long will it continue to do useful things? And the answer is, you know, it's like, it's like, you know, an energizer bunny, like you wind it up and you see how long it's going to keep walking. Eventually right now, the models, they do run out. There is no measure of a, yeah, this just does it infinitely forever and always stays coherent. Eventually it will become so big that the answer is like three years. The answer is like basically infinite. Basically you never have to check in on this thing. It will always continue doing useful work and continue working on the underlying task. And that's the point at which it's like all bets are off. All our intuitions about this stuff breaks when the answer is the model just continues doing useful things uninterrupted pretty much forever.
Ryan:
[1:00:37] I know, and that world could be pretty close, actually, if you see kind of the, it's a parabola, that line of how long for task completion is just like up and to the right.
David:
[1:00:49] It's exponential. In the stream of our lifetimes, that will be happening tomorrow.
Ryan:
[1:00:52] Yeah, basically.
Haseeb:
[1:00:54] Yeah.
Ryan:
[1:00:54] One other question for you, Haseeb, before we wrap this up. You mentioned cringe. The AI world thinks crypto is cringe. Again, I'll reckon to the Alex Friedman podcast with the founder of OpenClaw. And they talked about crypto a lot, actually. But not in the way that we're talking about crypto as, oh, it could be like an empowering money system, financial property rights system for OpenClaw bots. It was more like Peter saying, I hate crypto because... Almost made me want to delete the project and quit. Because I got harassed.
David:
[1:01:26] I got harassed. These meme queen people tried to profit off of my work. They wouldn't leave me alone. They tried to hack my systems. All this kind of very nefarious. It was one of the worst PR scars on the crypto industry that went viral that I've seen in a while.
Ryan:
[1:01:40] It wasn't great, right? And you get the sense that many people in AI actually feel like this. Is there a way for, like why? I think we sort of know why that's come to pass, But maybe the deeper question is like, can we get that stench off of crypto? Like what needs to happen for AI developers to show crypto some respect for what it can do from a positive perspective, not just the speculation and meme coins and front running?
Haseeb:
[1:02:09] Yeah.
Haseeb:
[1:02:09] I mean, so the first thing I want to say is the people who believe in AI are in many ways the same people who believe in crypto. You know, if you look at Elon Musk, you look at Sam Allman, you look at even, you know, Mark Zuckerberg, these are the biggest champions of AI and of the tech industry, and they believe in crypto. You know, there's a story just very recently that Meta is exploring launching their own stablecoin. Obviously, they originally tried to launch the Libra. Elon Musk was the first to integrate Bitcoin payments into any major tech company, into Tesla, and obviously he's a big advocate of Doge.
Haseeb:
[1:02:36] And you've seen the same thing, of course, with Sam Allman. You know, just recently, both Anthropic and OpenAI released papers showing how AI agents can go and interact with the VM for cybersecurity. So the reality is that as much as, look, yeah, it's true, like crypto is kind of cringe and meme coin people are annoying. And, you know, there's a way in which all this stuff taps into the worst elements of human nature.
Haseeb:
[1:03:00] It's all true. I think it's not going to go away. The same way the worst elements of human nature are also not going to go away. So, you know, there's a, there's a sense in which, you know, looking at email, like my email, I don't know about you guys, my email is fucking full of so much spam. It's like, if you just take like an all, if you like click the all mail thing in your Gmail, like don't do it by the way. But if you do it, you will just see like how much of a cesspool of garbage your email is, right? Like this is the tech industry. This is what they made. They made this. And like, what do you know? Human incentives create all sorts of bad behavior. And the job of the technology, the job of, you know, Google or Gmail or whatever email provider is, is to protect you from all that stuff happening with technology. My claim is that that's the exact same thing I was going to do.
Haseeb:
[1:03:46] Crypto is the same way. Crypto is like anything. There's a cesspool out there. It's human nature. You know, it brings out the best in humans. You know, the people donating to Ukraine, people donating to the defense of Roman Storm. There are moments in crypto that are absolutely soaring in terms of the human spirit and what it tells us about who we are as a species. And there's also the worst of the worst. There's also the SPFs. There's also the fucking Iran and all this other stuff that goes on. There's also this open claw harassing developers who are trying to build real stuff with stupid meme coins. That's also part of human nature. None of this stuff is going to go away. Neither side of it is going to go away. You will see the exact same thing with AI. Like what I'm talking about here, about these, you know, scambots proliferating online, that will happen at the same time as AI agents are literally, you know, saving people's lives with mental health crises and counseling them through their healthcare, counseling them through their legal problems. Everything, everything in technology is complex. It's a mixed bag. So crypto is part of the story. Crypto is part of the story because what's happening to information is also happening to money. It's all getting digitized and pulled forward like it or not.
Ryan:
[1:04:55] Long-term, this cringe won't be an impediment to adoption because the other forces are too strong. The adoption forces are too strong.
Haseeb:
[1:05:01] That's right.
David:
[1:05:02] Hasib, thanks for coming on the show. We always appreciate you coming on and just chatting about the future. You're always on the ball. One last subject before we let you go. So Dragonfly, new $650 million fund for, congratulations, banger number to raise in 2026. Understanding how prevalent AI is broadly and also the seemingly big attempts to integrate crypto and AI together, you know, the Ethereum, the EF is working on 8044, no, 8004, Coinbase is working on X402, has the impact of AI adapted your guys' strategy
David:
[1:05:40] for how you guys are going to allocate funds? What is different about the strategy of Dragonfly Fund 4 because of AI?
Haseeb:
[1:05:47] Yeah, I mean, we're spending a lot of time looking at the stuff that people are building in the space. The answer, of course, is that it's still early and very unclear who the winners are going to be and where value is going to accrue. But I'm spending a lot of my time personally looking at the AI space. But also, I want to make very clear that We're spending a lot of our time looking at the bread and butter. So we're looking at stable coins, payment companies. We're looking at DeFi. We just made a significant DeFi investment that we're going to announce soon. So crypto is crypto. Yes, AI agents are going to be a big part of how crypto works, but I think it's pretty plausible that even if AI agents proliferate and become a massive part of the crypto landscape, that there's not necessarily that much to invest into. It may well be that, yeah, AI agents use crypto and they use Bitcoin D. You know, they use the, like, you know, we were just talking about the fact that like they don't need special tools. They're generally intelligent. They can use what we can use, right? Or they can use command line tools. And so it's possible. I mean, to be clear, it's a little early for us to say how it's going to play out. But, you know, somebody was asking me the other day, like, if I believe in this AI agent thesis, what should I buy? You know, how do I invest in this thesis? And I told him, look, I don't know. I think what happens is that like the total demand increases because AI agents are just going to be using all this stuff. They're going to be using all the chains. using all the protocols. You're putting assets in DeFi. They're going to do all the same things that we're doing. So it's almost like if I were to tell you, you know, China is going to unban crypto, what should you buy? The answer is like, I don't know anything.
Haseeb:
[1:07:13] Everything will go on. You should just have exposure. It's like it's just more demand. It's just like the floor of everything will increase. Now, some things may benefit more than others, almost certainly. But the higher order bit is that like, yeah, it's just good for crypto. If more and more people, more and more demand, more and more agents, good for crypto.
Ryan:
[1:07:29] They're all new users. Haseep, let's leave it there. Thank you so much for joining us.
Haseeb:
[1:07:33] My pleasure. Thanks, guys.
Ryan:
[1:07:35] Bankless Nation, gotta let you know, of course, you know, crypto is risky. You could lose what you put in, but we are headed west into the frontier, frontier of AI this time. It's not for everyone, but we're glad you're with us on the Bankless Journey. Thanks a lot.