# 352: Google Next: Rebrandapalooza Duration: 105 minutes Speakers: A, Justin, Matt, Jonathan Date: 2026-05-05 ## Transcript [00:07] A: Welcome to The Cloud Pod, where the forecast is always cloudy. We talk weekly about all things AWS, GCP, and Azure. [00:14] Justin: We are your hosts, Justin, Jonathan, Ryan, and Matt. [00:18] A: Before we get into this week's news, we want to take a minute to tell you about We Are Developers World Congress, which is finally making its way to North America this September. If you've spent any time in the European tech scene, you probably know the team behind it. They've been running World Congress in Berlin for over a decade, and it's a big deal over there, pulling in more than 15,000 developers every year. Our friend Kote from Software Defined Talk is actually speaking at the Berlin event this July, and from what we've seen, these are the people who know how to put on a good developer conference. This September 23rd through 25th, they're bringing it stateside to San Jose. Organizers are expecting more than 10,000 developers with over 500 speakers across 18 different content tracks. Covering the entire stack, including cloud, DevOps, AI, security, software architecture, data engineering, frontend, and developer experience. If you've got a team, everyone's going to find a full schedule. It's not just sit and listen sessions. There are keynotes, workshops, masterclasses, and hands-on labs, the kind of stuff you can take back home and work on on Monday. There's an impressive list of speakers, including names from Datadog, Honeycomb, Sentry, Google, LinkedIn, Stack Overflow, Netflix, Microsoft, and Stripe, plus Kelsey Hightower, Oliver Pommel, Christine Yen, Scott Hanselman, and Angie Jones. Head over to wearedevelopers.us to grab your ticket and use code DEVPOD26 for 15% off. That stacks with their group rates if you're bringing 4 or more people, and honestly, at that price, you should probably bring the whole team. [01:50] Matt: Episode 352, recorded for April 28th, 2026. Google Next, Rebrandapalooza. Good evening, Ryan and Matt. How you guys doing? Hello. [02:01] Justin: Doing good. [02:02] Jonathan: Good. Welcome back from Vegas. [02:04] Matt: I mean, Ryan and I survived it. That was the most important part of Vegas. Barely. Barely. [02:10] Jonathan: We survived and we did not succeed in recording. There was a medical emergency, so I'll take that as the excuse. [02:16] Matt: That's a valid, completely valid excuse. So we, we We understand, we appreciate your attempt. I mean, the most important part about Vegas for me was that it's over. Cause I gotta go home. But Ryan, Ryan got to do his first, first Sphere experience. So I'm putting him on the spot. [02:31] Justin: Yeah. [02:31] Matt: He got to go to the Sphere. It was not the Eagles as much as I've tried to get him to go to the Eagles at the Sphere. [02:35] Justin: It is not. [02:36] Matt: And I, I do, I, but I, I made the statement earlier this week to you that I said we should go see the Eagles at the Sphere cuz they may, the Sphere might actually redeem the Eagles for you. [02:45] Justin: Zero chance. [02:46] Matt: The visuals could be awesome. I don't know. I am just saying there's a chance because I've been to a couple of concerts now at the Sphere and it's amazing. And so this is your first experience. [02:53] Justin: This is my first going to the Sphere at all. [02:54] Matt: You should share. You should share how you feel about it. Yeah. [02:57] Justin: So I went and I saw Phish with a buddy of mine at the Sphere. And I got to say, like, the— I'm not a huge Phish fan, like passing, but that is the greatest, like, meld of two things, like getting a jam band with those visuals. And the sound in the sound for that venue is amazing. [03:17] Matt: Oh yeah. Like when you read the, of the details behind that building and like how many speakers and how many directional speakers at your seats there are. And like, it's, it's a crazy impressive venue. [03:27] Justin: Like, I mean, it's really, really, really impressive. And none of it is like, it's not like a typical concert where there's this huge array of things. Like it's all behind basically this scrim that they use for all kinds of really cool visuals. And it. Really changes the way that you do concert lighting. I thought like, it's really neat cuz they, well, they had sort of like faux concert lighting, like, uh, almost like cartoon lighting at certain parts of the show, but it really changed the lighting in the venue. Like in terms of like looking around on the crowd and stuff, it's a lot brighter. There's a very different concert-going experience than I've ever experienced. And so that was, uh, I think it's really wild. Really cool. I like, I look forward to looking at, you know, going to other shows there. [04:07] Matt: Yeah, I've seen, um, the, you know, of course there's a movie, it's like a This Planet Earth type movie where they show you nature, which is cool in there. Although really big animals staring you down. And then I've seen of course Wizard of Oz there. Uh, I saw the preview at Google Next last year. Then I, I saw the actual movie there as well, uh, which is kind of like a 4D experience where they have like, you know, they add wind and all kinds of stuff to it. [04:29] Justin: Oh really? [04:30] Matt: Pretty neat. And then I've seen now a couple of concerts there. I saw Backstreet Boys, and I'm not a Backstreet Boys fan, so, but my wife is. And so I, we went and it was great. I had an amazing concert experience. It was awesome. And so I'm definitely looking forward to Metallica who's supposedly coming. Yep. Oh, it's, I think it's official now. [04:48] Justin: Yeah, it's official. October through May or something like that. It's a huge residence. [04:52] Matt: Oh wow. [04:52] Jonathan: It's a long time. [04:53] Justin: Yeah. [04:53] Matt: I mean, the Backstreet Boys that were there forever. And then, um, you know, even the Fish is, this is not the first time Fish has been there. [04:59] Justin: No, it's not. I don't think they have a residency. I think they just hit it up. [05:03] Matt: But yeah, then the Eagles, you know, they, they keep coming back regularly too. So yeah, there's always that option for you. [05:08] Jonathan: Mm-hmm. [05:08] Matt: But yeah, no, I definitely, uh, it's, it's a cool place and it's worth going to Vegas for that concert experience. It's expensive. That's my only complaint. [05:15] Justin: It is very expensive. [05:16] Matt: It only holds, I think, what, 8,000 people? I mean, it's not a huge venue either, so it, it's expensive for the right reasons. But, uh, you know, it's, it's pretty cool. Definitely worth checking out if you're there. So I'm glad you gotta do it. I'm glad you had a good time. Yeah. It's, it's amazing. [05:30] Justin: It really is. I didn't realize how small the venue is, cuz I didn't realize it was only 8,000. [05:36] Matt: I'm surprised by that. I never— you're low. Sorry. It might be the floor capacity. Uh, beer seating capacity. Let's do real-time follow-up. [05:43] Jonathan: You said it with such definitivity. [05:45] Matt: I did. I did. Standing floor capacity is 1,400 and then total capacity is 20,000. Seat capacity is 17,000. [05:50] Justin: That makes more sense. [05:51] Jonathan: Yeah. [05:52] Matt: Yeah. That's about right. Yeah. [05:52] Jonathan: That feels better. [05:53] Matt: Yeah. But it's still, it's, yeah. And They keep trying to build more of them. I think the only place that's agreed to build another one is Dubai so far. 'Cause of course Dubai. [06:00] Justin: No, there's another one I heard, uh, that's being built out in Virginia. [06:04] Matt: Like, oh really? I heard that one. I know they wanted to build one in London and London was like, yeah, no, the light pollution. But I'm like, the outside of the Sphere is cool. Like it's definitely awesome on its own, right? Especially in this, in the whole thing. But you know, the venue inside is actually way cooler than even the outside of it. So I don't, I don't know why they are insistent on having the outside of the Sphere. Other than the marketing potential of it. But, uh, you know, apparently got killed it in London, which, uh, makes sense cuz no one in London wants a giant eyeball looking at everybody. [06:30] Justin: Yeah, it is kind of like, yeah, like where else could, other than Dubai, would it, would you build one where it makes sense? [06:37] Matt: Singapore makes sense to me. [06:39] Justin: Oh, good call. [06:41] Matt: I get maybe Sydney as well, but that one I'm not sure about. So yeah, Tokyo would definitely probably fit in right in there. Uh, there's definitely some places I think it would work, but they're expensive to build. They're like, I think it was how many billions of dollars? It was many, many billion, $5 billion, something like that. [06:57] Jonathan: No, I think they did it once, so it should be cheaper the second time. [07:00] Matt: That's what they say. [07:01] Jonathan: That's what they do say. [07:02] Matt: But, but, uh, you have seen that inflation and, uh, tariffs have raised prices of everything. So maybe not. You'll never know. [07:09] Jonathan: It's kind of fitting with the giant eye and being in DC, you know, with all the government stuff. So everyone's spying on each other. Kind of makes sense there. [07:18] Matt: Maybe they should make a ballroom out of it and then they could just put it right next to the White House. So I'm just saying. All right, let's get into real news here. Google Next happened. We'll get to that in a minute, but there's a lot of news before Google Next that we have to get through here quickly. First up, Amazon and Google have invested— Amazon $25 billion and Google $40 billion into Anthropic. Both of these partnerships are a commitment to buy more capacity and measured by gigawatts On the Amazon side, they're going to be using the Trainium 2 and Trainium 3 capacity. And on the Google side, they're going to be using the TPU architectures. So Anthropic is definitely taking advantage of all the GPUs they can get as their growth continues to be astronomical. And so, yeah, they invest the money and then they pay it back to them, which is sort of a weird money on paper problem. I don't really understand how it works, but that seems to be what happens all the time right now in the AI space. [08:09] Justin: It does. [08:10] Matt: And so Anthropic continues to get big amounts of money and their valuation continues to be absolutely crazy. [08:16] Justin: Yeah, there's, uh, yeah, if I understood how the, you know, the, the money thing worked, I saw a funny thing where it was someone was plugging an extension cord into itself as a description of how, how this was working. [08:32] Jonathan: Yeah, kind of what it feels like though. Yeah, I give you money, you give me it back to me cuz you're gonna spend it on my platform. And somewhere in there, somebody gets a salary. Like, it's kind of like you can't infinitely make power and there's always some loss. There's some lost money somewhere in here. [08:49] Justin: And you probably can't tax this in the same way, right? So it's all just sort of how they're getting away with, like, I don't know, we're investing. It's not buying, investing. [08:58] Matt: But like, how, I mean, like how much shares eventually that means Anthropic has to either return that investment if it's a loan. And they get paid back with interest, or they own some portion of the company and then the company has to go public to basically redeem that money to them. So I, it's a lot of money. I don't know. Yeah. The IPO on this company's gonna be massive when it happens. [09:19] Justin: It is. It's gonna be amazing. And, but I wonder like the financials then. I think it'll be one of those that balloons up right away and then craters to ashes because I think the, I, I, AI still is very unsustainable in terms of cost. So it's, yeah, kind of nuts. [09:36] Matt: I mean, the biggest thing is gonna be when they file their, was it the 10-K that comes out before they go public or the prospectus, whatever. You know, that's what, that's what killed, um, what was the name of the coworking space? Uh, WeWork. [09:49] Jonathan: WeWork. [09:50] Matt: He's like, when you actually looked at the financial numbers, you were like, oh my God, this is a house of cards. And it basically imploded their entire IPO. So, The big question's gonna be when one of these companies, OpenAI or Anthropic, finally goes public and they start publishing these things, what people's actual reaction is to their financials. [10:06] Justin: And yeah, and is it, can they sustain like the level of growth that the Wall— that the street wants? [10:11] Matt: So, right. [10:13] Jonathan: Well, they can't sustain it and they're essentially subsidizing it. Something I saw for, I think it was GitHub, where they were saying it's, they The $30 you pay per person, a low user, and don't hold me to these numbers, was like $20 it was costing them. The medium was around $30, but the highs are like $90 per agent. It's like they're losing money on every single, or like over half the seats that they're spending right now just to get people on it. [10:40] Matt: Well, they, they were until today. [10:42] Justin: Until today. [10:43] Matt: We'll talk about that. We'll talk about that next week. But yeah, 'cause it came after our cutoff for the week. But yeah, no, they, They've realized they're about to lose a bunch of money, especially with Agentic. And so they, uh, I think they have been losing money. Oh, I'm sure they've been losing money, but they're going to lose a lot more money. That's the problem. So I think, I think you're going to see everyone moving into consumption pretty heavily. Even Anthropic has talked about moving away from the monthly plans, just moving to pure consumption, because the reality is they're subsidizing your use of AI at those cost tiers. Then people get mad because they change the, you know, they change the models, they kick out OpenClaw, et cetera. And then people get mad and it's like, well, but if you were truly paying for the consumption you were using, then you wouldn't be as upset about this. [11:25] Jonathan: So I think it was like $220. I saw somebody like looked at the number of tokens and did the math out. It was like $220 for the $100 plan is essentially what you're getting equivalent to tokens. [11:36] Matt: Right. So we'll continue to see Anthropic get invested in, I'm sure. [11:42] Justin: Mm-hmm. [11:42] Matt: Uh, oh, next, you know, probably 2 or 3 more cash infusion. 'Cause their side of it is they are paying a lot of money to these cloud providers to provide TPUs and GPUs and NVIDIA itself. And so eventually that merry-go-round has to stop. So we'll see. SpaceX has apparently struck a deal with AI coding startup Cursor via either a $60 billion acquisition or a $10 billion partnership fee, giving Cursor access to XAI's Colossus supercomputer, which runs 200,000 NVIDIA H100 equivalent GPUs for model training. Azure has been compute constrained despite reaching $1 billion in annual recurring revenue and a $29.3 billion valuation. So this deal directly addresses their infrastructure bottleneck for scaling model intelligence. The partnership positions SpaceX to compete in the AI coding tool space against Anthropic and others, notably given XAI's Groq has publicly acknowledged falling behind competitors in coding capabilities. For developers and cloud users, this deal signals continued consolidation between compute vendors and AI coding tools. Which could influence future pricing. SpaceX's recent acquisition of xAI combined with their cursor deal suggests a vertical integration strategy connecting rocket company compute infrastructure directly to developer-facing AI products ahead of a potential IPO later this year. Uh, this is— this IPO is going to be a lot of funny money. This one might be one. [12:56] Justin: Oh my, yeah, like it's so weird. Like, you know, it's, it's already bad that they, you know, the, they sold this, the AI What was the, the XAI to the SpaceX? [13:07] Matt: To the SpaceX. [13:08] Justin: Like that's just giving money to yourself. [13:11] Matt: Yeah. I can't believe they didn't rename SpaceX to SpaceX AI. Wouldn't that have made more sense? Wouldn't, is that what you would've done? I, I, yeah. I mean, I think people saw this announcement come out right before NeXT and we're all, everyone was like, why is SpaceX buying Cursor? Then you had to remember the part they bought XAI and it was like, oh, okay, this makes perfect sense in the light of XAI, but still doesn't make sense though. So that's my problem with it. Yeah. I mean, the, the thing that I don't get is the $10 billion partnership versus the $60 billion acquisition. [13:36] A: Mm-hmm. [13:37] Matt: Like what's the triggering events on those things? Like when, like when is it a partnership versus when is it now an acquisition? And does that mean that these people who are working at Cursor, if it's a partnership, aren't getting equity? Because that's a bummer. [13:50] Justin: Mm-hmm. Yeah, I didn't catch that it was one or like it, so that's kind of interesting. [13:55] Matt: I mean, I think it starts as a $10 billion investment. And then it can turn into a $60 billion. It's like a $60 billion option on it is basically my understanding. [14:04] Justin: Yeah. [14:05] Jonathan: That's, uh, still feels strange. [14:07] Matt: It's a weird deal because I think reality is Elon doesn't have any money. Right. It's not liquid, not real money. Yeah. And so this is his way of like, I can basically tie you down at a, at a reasonable price. And then if the valuation goes down further, I can get you cheaper maybe later. Right. That's kind of my feeling. [14:23] Justin: Yeah. I mean, Tesla's no longer gonna make cars, they're just gonna make robots. SpaceX is doing AI, like they're, it's cats and dogs living together. It's mass hysteria. [14:32] Matt: Yeah, it's crazy talk. It's crazy talk. Uh, all right, well, AI is how ML makes a lot of money apparently. This week, uh, OpenAI updated its agent SDK to general availability, adding native sandbox execution, configurable memory, and file system tools modeled after Codex. Agents can now read and write files, run shell commands, install dependencies, and apply code patches within controlled environments without developers building infrastructure themselves. And Ryan says, kill it with fire. [15:00] Justin: Uh, no, as long as it also logs and has permissions and some sort of boundaries, I don't have to kill it. [15:10] Jonathan: Admin rights for everyone. [15:11] Justin: It's no, it's just terrifying that people, because we already, already have you know, people that are just throwing questions into any chat tool and just then running whatever command it spits out indiscriminately. And now this is gonna happen at a faster rate and people don't quite understand what's going on and in a lot of cases don't care to, right? Just do what I want. [15:34] Jonathan: I was gonna say, Ryan should never look at how I play with my dev environment for like my personal AWS account. Like, hey Claude, just go run fully autonomous and just draw whatever commands you want. Just, it's fine. Don't worry about it. [15:46] Matt: So you're, you're saying that I shouldn't go use this website that's by a company called Moonshot that's Chinese. I don't, and I, I don't know what my data is doing. I mean, I think the reality is most people don't actually know what's in their prompts. Yeah. And if you ever actually look at the prompt logs, you're, you should be like, oh my God. Cause it sends everything. [16:04] Justin: All the things. Yeah. I don't think people quite realize what data gets consumed. Yeah, credentials and environment files, all kinds of stuff gets sent somewhere and it doesn't necessarily mean that it's exposed, but it, it also doesn't mean that it's secure. And so like if there's ever kind of a, a leakage or any kind of breach at one of these AI companies where they can get ahold of that data, if somehow, that'd be a problem. [16:26] Jonathan: I've definitely seen people say, when it goes like insert key here, and they just go, no, no, no, give me the script. Here's my key, create the script for me. I'm like, You understand that like, one, you just killed like, I don't know, 3-some, I don't know, whales by you or this whole other prompt. And all you had to do was just copy and paste the key into this spot. And then your key is now probably not that, I mean, like you said, it's not unsecure, but it's not the most secure thing anymore. It's somewhere in the logs of one of these companies. [16:59] Justin: The truth is, it's not really the agent problem. The truth is, is that we don't have the correct authorization permi— permissions model for people, right? And so we give people the, a broad standing access and now that's a concern because there's gonna be a whole bunch of automation that's basically pretending to be me running on these local environments. And so like, that's really the issue. And so it, I think it'll change, it'll have to change the identity and access management plans that we use. So there'll be a lot more, um, You know, just-in-time permissions and accepting of that kind of thing. So a continuation of what you already have with AI of like, is this okay? Click approve. Is this okay? [17:38] Matt: I mean, I think the fear that you have to have in that idea is, okay, so we are asking for a lot of permissions all the time in like Claude code, right? Or Gemini CLI. It's like, I wanna run this script, I wanna do this thing. And you're like, yes, yes, always, always, always. Cause you get tired of, you get, it just becomes prompt fatigue. Which is kind of what happened with like Windows UAC back in the day. You know, people just got to the point of like, whatever, I just hit yes, I don't even read these things, when it grays out your screen and makes you do a thing. And I think that's the risk that you have even inside of some of these agentic users' identities, is that if you're going to constantly be having to approve things, then you don't actually get agentic, which is what the whole point was. And so the balance of security versus agentic versus least privilege is gonna be a, a really interesting friction. [18:26] Justin: Yeah. [18:27] Matt: That's gonna be going out the next year or two. Yeah. [18:29] Justin: And it's crazy 'cause I, like, I don't have great answers. Like, it's like, it's bad on both sides. Like, I want the productivity of Agentic. I use the, you know, so much more productive because of these things. And I want to enable, you know, agents on my work laptop where I can do much more sophisticated things. But, But I also am terrified of that and people being able to like execute directly against production environments, you know, because they're port forwarding into that environment. They can just run a command and the AI knows to do it because you left some other file that says how to do it. And so it's like, oh, this is useful information. I'll consume this and then just use that. [19:11] Matt: Yeah, it's just, eh. Well, you know, the things like, you know, Cowork and these other tools, like you can create whole agents that run on your laptop that are now, you know, invoking things as you because 'cause they have access to your identity and things like that. And so yeah, the boundaries are evolving and changing and the perimeterless enterprise has a suddenly new tone to it, doesn't it? More sinister than it used to be. [19:34] Justin: Yeah. [19:34] Jonathan: I mean, at one point you gotta have your end users be responsible for what they're doing and, you know, help them help themselves and help the company because they have to take some ownership. [19:46] Justin: They have to take some ownership, but I don't know how to give sane guardrails. And that's really the problem. As a security person, I don't want to say no to everything. That's the joke. But the goal really is to provide those same guardrails. And then not, you know, the problem is it's not everyone agrees what those guardrails should be, but that's fine. And this one, I don't know, other than like saying, yeah, if you do something bad, you're fired. I got nothing, right? Like, it's like, and I don't want to do that. [20:09] Jonathan: But some, at one level, at some point, It does come down to that. [20:14] Justin: Like, yeah, yeah, we all have to have some accountability, but it's also, you know, it's— it— when it's so quick to shoot yourself in the face, like, it's all— it has— there has to be some sort of level of protections as well. [20:28] Matt: And it doesn't have to be, we just fire people. [20:30] Justin: Yeah, okay, non-Calvin, we blame AI for it. [20:34] Matt: It's fine. Yeah, it's fine. That's what I mean. That's what they're doing now, AI layoffs. For business problems. [20:39] Justin: That's true. [20:39] Matt: Popular thing. [20:40] Justin: Kind of a different way to do it, but I guess, yeah. [20:44] Matt: Well, if you like to burn money, Claude Opus 4.7 is now generally available across Cloud products, the API, Amazon Bedrock, Google Cloud Vertex, and Microsoft Foundry, all at the same price as Opus 4.6. So you get $5 per million input tokens and 25 million per output tokens. Model targets complex, long-running agentic coding workflows with early testers reporting 13% higher resolution on a 93-task coding benchmark. And 3x more production task resolution on Rakuten software benchmark compared to the Opus 4.6. Vision capabilities received a notable upgrade with Opus 4.7 now supporting images up to 2,576 pixels on the long edge, more than 3 times the resolution of prior Claude models. This opens up cases like computer use agents reading dense screenshots and data extraction from complex technical diagrams through higher resolution images with some more tokens. Anthropic is using Opus 4.7 as a testbed for cybersecurity safeguards for any broader release of its more capable Mythos preview model. The model includes automatic detection and blocking of prohibited cybersecurity uses, with a new cyber verification program available for legitimate security professionals doing penetration testing or vulnerability research. A new extra high effort level to burn those tokens faster sits between the existing high and max settings, giving developers finer control over the reasoning versus latency trade-off. Developers migrating from OPUS 4.6 should note that updated tokenizer can increase token counts by roughly 1 to 1.3 to 5 times depending on content type, And a migration guide is available to you, uh, out there. So yeah, this is a fun trick. We just make the tokenizer be less efficient and then we make more revenue. It's beautiful. The last thing is file system-based memory improvements allow Opus 4.7 to retain notes across multi-session agentic work, reducing the need to reestablish context at the start of each task. This is particularly relevant for enterprise teams running parallel agent workflows, which are still not GA, but I use every day. [22:27] Justin: Well, there's a lot of other platforms in which you can sort of enable this, right? Where it's, you're sort of tying it together with with Glue ecosystem. It's funny that they, this is, I didn't realize it's the same price because every platform that I'm using this in, Opus 4.7 is so much more expensive. [22:43] Matt: Well, it's more expensive because of that tokenizer and, uh, other, you know, memory things and other reasons why it is a bit more expensive. And also you can put bigger images into it, which, you know, it, the thing is if you were pasting images in and they were too big, it would just automatically downsize them previously. Now it just takes a bigger image. So now it's more tokens it has to deal with. So there's lots of little, little things there that you need to be aware of in that conversation. [23:07] Jonathan: I mean, I definitely found it burning more tokens and I switched back to 4.6 for a few things, but also, you know, as I use AI more and more, I'm starting to try to be a little bit more conscious of what model I choose where, you know, and not just saying, ah, let me use Opus all the time. Sonnet does perfectly fine for 90% of what I need, you know, and except for when I'm doing deeper research and analysis and trying to fix this, you know, a pretty deep bug, then I flip over to Opus. [23:36] Matt: I mean, if you don't need reasoning, even Haiku can do a lot of work for you. So, uh, like if, if you don't need a lot of reasoning and you have a very specific clear ask, like I need you to take these secrets and convert them to parameter stores, save me money. Haiku can do that in its sleep. [23:53] Justin: Um, yeah. [23:54] Matt: You know, and so that's like, it's a very clearly defined this and that task where you need a thought and like red engineering and stuff. That's where you want the Sonnet or the Opus. But I've never been a big Opus user cuz A, it is outrageously expensive. I don't know if I'm in the middle of like a production bug and I need to figure out what's broken with this code real quickly. Yes, I might use Opus then. But I also like a lot of my coding stuff, I'm using more and more into like KEMI 2.5, uh, not via Moonshot, by the way, but, uh, through, you know, Ollama and other solutions that provide, you know, these, because those, those models are being rated near the same level as like an Opus or a Sonnet, uh, and they're like $3 or $4 per million input token, you know, it's so much cheaper. So the reality is like you definitely should shop around. So I just started using Bifrost in the last week just because I wanted an easier way to migrate between different model agents. And Bifrost is super cool way to like add multiple providers, be able to choose them just the way you would in any other tool. And you can use Claude code, you can use OpenCode, you can use Cortex or any of the other agents that are out there that if you, whatever one you prove you'd like. So you don't need to change your tool, you just change the backend. Everything else works the same. And then you just, you know, oh, this is working great. I'll keep using it. If not, then you switch, but there's lots of opportunities to save yourself a lot of token money. [25:10] Justin: Definitely. You do have to kind of invest and understand that whole ecosystem 'cause it's, it's fine once you get it up and running. Getting it set up is a little complex. [25:19] Jonathan: And that setup is what takes the time, you know, but the payoff on it is a lot if you're doing a lot of development. And that's kind of the balancing act here is if you are doing a lot, then it makes sense to spend the time. If you're not, then something, you know, off the shelf. [25:38] Justin: Yeah. It is kind of funny 'cause I, I go back and forth, uh, with my wife who's, you know, thankfully not in tech and like she goes, she, I walk her through some of my setups and she about halfway through start stops listening and then she, she understands that it's clearly like too complex and she just wants to ask a question somewhere. It's like, okay. [25:57] Jonathan: Yeah. [25:58] Matt: Yeah. The reality is that like, you know, don't give up your data center. Maybe buy some GPUs because running these models yourself might be the only way to save yourself a lot of money. I mean, cautionary tales. I don't think we talked about it on the show. Maybe we did, but you know, Uber, they spent their entire year budget for tokens in 4 months, like an entire budget. They now have to figure out something else. And so yeah, maybe don't use Opus, number one. But, uh, yeah, like those are big risks that businesses have. So I think the this area of being, how do we be more efficient, I think is gonna become a bigger area. And I'm, I'm hoping there's gonna be some tools coming out that'll actually like dynamically look at what you're asking and make the decision. 'Cause right now that exists sort of in some of the, the APIs out there. If you have specific rules like, oh, they're asking a question about this, or they're asking this, I will go to this model. So you can do like really static routing rules, but like, I really wish there was a more dynamic interpretation of like, oh, this is, this is a super easy question. This is a more complex question. And then it can automatically route between them. I think that's what we'll probably see is kind of the next, next phase of AI tooling will probably be a bunch of tools that help do that exact thing. I think you do. [27:06] Justin: I think that is an option for some of the coding stuff like Cursor. [27:09] Matt: Huh? [27:10] Justin: You have to configure it and then Copilot can also do that. I'd like to see it expand outside of the coding stuff for, you know, maybe more just a general. [27:18] Matt: But then you also risk, uh, you know, if it's like, I think Anthropic's working on something, an adaptive mode where basically it'll adapt between the 3 models. But then going back to this tokenizer, hey, we need more revenue this month. Change the adapter to use more Opus this month. Yeah. [27:34] Justin: But they are, you know, like on the flip side, they are doing granularity. You can now choose the amount of reasoning that you select even above the model. And so like there's other things that they're doing, giving more buttons, but you have to be flipping around. [27:47] Jonathan: A lot of these things is where I want that model router that like Azure and other cloud providers to really, that's the intelligence I want in that service. Like I want to, weirdly, I don't say I want to pay the cloud providers much in general in life, 'cause I give 'em way too much money either way. But I want to pay for something that will do that intelligence. And like Justin, you said like, hey, this is a simple query, send it to Haikyu. It's an if this, then that. Versus no, this is a, hey, I need to architect this entire thing, go to Opus and burn a few extra dollars, see a little bit more detail so we can use the cheaper things, these cheaper models later on because everything is broken down in a much more simplified way. [28:31] Matt: Agreed. Yep, model routers, hot topic for sure. All right, moving to Claude Design was released. This is their answer to Google Stitch. Basically Anthropic launched Claude Design in research preview for Pro, Max, Team, Enterprise Cut subscribers powered by Claude Opus 4.7 enables users to create interactive prototypes, pitch decks, wireframes, and marketing assets through conversational prompts and inline editing controls. The novel workflow features that Cloud Code handoff, where finished designs are packaged into a bundle that developers can pass directly to Cloud Code for implementation, creating a tighter loop between design and engineering. Cloud Design builds a team-specific design system during onboarding by reading codebases and design files that automatically applies brand colors, typography, and components to every subsequent project. Teams can maintain multiple design systems simultaneously. Earlier user data from Brilliant suggests complex pages that required 20+ prompts in other tools needed only 2 prompts in Claude Design, indicating a meaningful efficiency gain for interactive prototype creation. Export options include Canva, PDF, PPTX, and standalone HTML with organizational scope sharing and collaborative editing. And for enterprise customers, the feature is off by default and must be enabled by admins in the organizational settings. So yeah, designer. It's great. Love it. [29:44] Justin: Yeah. I haven't had a chance to play around with it. I mean, I, I really like these kind of tools just because, you know, big joke is my, my UX skills suck. So these are, I love, even, even with AI, my UX skills suck, which I find hilarious. [29:59] Matt: I mean, I mean, the reality is Claude was pretty good at design before this tool existed. Like it, you know, if you put the front end design, uh, plugin into Claude, it does a really good job on its own. So, you know, it's nice to see it does, I did unfortunately decide to use this and try it out in the middle of a coding thing and it opus ran my credits right out. So I had to wait, I had to wait 2 hours to get my next allotment of credits or use overages. But yeah, it's definitely impressive what it can do. [30:25] Jonathan: Now, is it as good as they say? 'Cause I feel like a lot of the Anthropic, you know, we talked about this, I guess 2, 3 weeks ago, a lot of their stuff feels always a little bit overhyped. [30:35] Matt: So, I mean, I haven't tried Mythos, but you know, the, the analysis of the market is that Mythos is a bit overhyped. I played with this personally and I was very impressed with its output. It's at par or better than Stitch. So that's the only thing I can compare it to. I don't use Canva, I don't use Figma, so I can't compare it to those tools. Maybe it's not as good as those tools. Again, those are specialty tools for designers. I am not a designer, nor do I ever wanna be one. And so, you know, for what I, what I need, it's pretty darn good. And I like it. But again, I'm not a designer. So yeah, I'll keep an eye out for some articles about people saying it's really shit and I'll let you guys know. [31:07] Jonathan: We're so optimistic. [31:09] Matt: Yeah, it was awesome. [31:13] Jonathan: There are a lot of cloud cost management tools out there, but only Archera provides insured commitments. It sounds fancy, but it's really simple. Archera gives you the cost savings of a 1 or 3-year AWS savings plan with a commitment as short as 30 days. If you do not use all the cloud resources you've committed to, Archera will literally cover the differences. Other cost management tools may say they offer insured commitments, but remember to ask, "Will you actually give me my rebate?" Archera will. Check out thecloudpod.net/archera to schedule a demo today. [31:55] Matt: Uh, Cloudflare, uh, held its first Agents Week, uh, which was not that exciting actually, but they had several things they released over the week. The first one up was the Agentic Cloud. So now you can run all your Agentic workloads on top of Cloudflare. The new environment supports both full operating system containers for package installation and terminal commands, and lightweight, which isolates the start in milliseconds for high-scale environments. They also shipped a Git-compatible workspace design for agent-generated code moving from prototype to production. Uh, security and identity were treated as built-in defaults rather than add-ons. Thanks. With new tools for connecting agents to private networks and managing autonomous actions taken on behalf of users across your organization. And the agent toolbox additions include inference, search, memory, voice, email, and a browser primitive, giving agents the ability to perceive, remember, and communicate without developers assembling separate third-party services. I look forward to Cloudflare taking down Cloudflare with us, uh, then writing an RCA with these great tools. [32:44] Justin: So it'll be a very well-documented RCA. [32:48] A: Indeed. [32:49] Matt: That's pretty funny. [32:51] Jonathan: Will the AI blame itself for it? [32:54] Justin: I mean, it does in my code all the time. [32:56] Matt: Yeah, it apologizes to me all the time. It says, you're right, Justin, you, uh, I was mistaken. I had to do that. [33:00] Justin: I'm like, cool, cool. [33:01] Matt: Appreciate you. Uh, that's why having personas is great. Cause even the same, even Claude as the developer versus Claude as the QA, and they argue with each other. It's fantastic. [33:10] Jonathan: Mm-hmm. [33:10] Matt: In addition to this, Azure GCP also launched Artifacts in private beta, a version file system built on Git that lets developers and agents programmatically create, fork, and manage Git repositories at scale via REST API and native workers API with public beta targeted for early May. System is built on durable objects with a Git server written in Zig and compiles a roughly 100-kilobyte WebAssembly binary, nailing tens of millions of isolated repos instances per namespace while handling the full Git smart HTTP protocol with zero external dependencies. Cloudflare is also open sourcing Artifact File System, a file system driver that mounts large Git repos using a blob-buttless clone and lazy file hydration, reducing startup times for multi-gigabyte repos from 2 minutes to 10 to 15 seconds. Uh, so yeah, I guess that's good too. Also another way that it's going to take down Cloudflare. So look forward to that. Yeah. Well, it's kind of interesting. [33:56] Justin: I didn't like, it solves a problem I didn't know existed, like, uh, in terms of, you know, being able to have an agent manage multiple Git repositories more quickly. [34:05] Matt: Oh, let me tell you how much I learned about worktrees. [34:07] Justin: Oh yeah. [34:08] Matt: Okay. In the last month, like dealing with multiple agents and wanting to work on the same codebase and you're like, well, your answer is worktrees. Worktrees are not designed for humans. So thank God there's AI because AI can make them work. Cause like, like the idea of worktree was, well, you're in the middle of doing code, but you need to do a hotfix against production, which is on the master branch. And so you need ability to do this quickly. And like, that was kind of the gist of when I read the documentation on this, it was like the gist of why they created it. But the idea was you only had 1 or 2 worktrees ever. [34:35] Justin: Right. [34:36] Matt: Um, and, uh, so now, uh, reading, uh, you know, with my AI, I can have like 10, 15, 20 of them and let AI deal with that. Nightmare, cuz I can't, it hurts my brain. [34:46] Jonathan: It's one of those Git features I didn't know existed and immediately solved the problem when I was trying to do like what Justin's saying, hey, I need 3 things going on at once that all are gonna collide. And yeah, I, I tried to do it manually myself at one point. For one thing, I was like, no, too hard, not worth the effort. You know, let's go have AI do this for this thing for me. [35:08] Justin: You know, just merge and rebase in like a cyclic fashion for these feature branches until You just end up ripping one out before committing to main. [35:17] Jonathan: No, I end up with a feature branch that just has too many things in it. That's not one feature. 'Cause I'm like, ooh, let me add this next thing as I'm doing it. Or I just commit and it's like, this code doesn't work and it's untested. And then I go back to my main branch, I check it out, I do the hotfix I need, and then I go back to my thing and I don't know where I am because I just have this commit that's called untested code. [35:39] Matt: Nice, brilliant. We'll work it out. We'll work on the QA process for your code. It's fine, Matt, we got you. [35:45] Jonathan: That's why I have a QA agent doing it for me. [35:47] Matt: Right, exactly. Snowflake is launching Cortex Agents as a full enterprise agent platform with several capabilities now generally available, including multi-tenancy with row-level data isolation, agent versioning with commit-based rollback, and resource budgets per, per agent and per team spending controls. So if you need your agents close to your data, this is a great way to do it. I definitely would look into your costs this one, 'cause Snowflake is not cheap for compute. [36:08] Justin: Nope. I, I do, I mean, I look forward to the, the MCP connector, which they announced as part of this. I think that that's gonna be a lot of fun. I think that's mostly because I don't like SQL interfaces in any kind of database. So like, but it, it, you know, like I do think that there's a lot of value in running tools directly in the data just because there's, it keeps it sort of contained and isolated and, and removes some of the concerns I have. Like running it locally with a direct session open to Snowflake, which is what everyone wants to do. So I do like this kind of model, but yeah, it's Snowflake, so how much can it cost me? A lot. [36:48] Matt: A lot for sure. Yeah. Well, you know, normally we expect Amazon to be a spoiler on Next, and they did their part, but OpenAI came in hard and heavy on this one too. So, First up, they announced GPT-5.5, generally available in both ChatGPT and Codex for Plus, Pro, Business, and Enterprise users, with API access priced at $5 per 1 million input tokens and $30 per 1 million output tokens. So a little bit more expensive than Opus. And a Pro variant at $30 input and $180 output for 1 million tokens. The model shows notable agentic coding improvements, scoring 82.7 on the TerminalBench 2.0 and 58.6 on SWE Bench Pro. By using fewer tokens than GPT-5.4 to complete the same task, which, you know, it's pretty nice. For cloud enterprise workflows, GPT-5.5 was co-designed with and served on NVIDIA GB200 and GB300 NVL72 systems with inference optimizations including dynamic load balancing heuristics that increase token generation speeds by over 20%. Print the money faster. Knowledge worker benchmarks are worth noting for enterprise buyers of 84.9 on GPT-VAL across 44 occupations. and a 78.7 on OSWorld Verified for Autonomous Computer Use, and 98.0 on TAO-2 Bench Telecom for customer service workflows. OpenAI is classifying GPT-5.5 as high under its prepared framework, preparedness framework for both cybersecurity and biological capabilities, and has introduced a Trusted Access for Cyber program through Codex that gives verified defenders expanded access with fewer restrictions, which has direct implications for security teams evaluating AI-assisted vulnerability management. [38:18] Justin: That's kind of cool. I mean, it's first I'm hearing of those kinds of frameworks where they're testing those things and sort of testing the, I don't know, the, not sustainable, but the safety AI sort of aspects of it and having a rating, which I like, but then having the ability to sort of give people access beforehand. It's like reporting a vulnerability, like directly before it go in public. It's kind of nice. [38:43] Matt: Yeah. To make Ryan grumpy, OpenAI then launched Workspace Agents in ChatGPT as a research preview for Business, Enterprise, EDU, and Teacher plans, positioning them as evolution of GPTs powered by Codex and designed for shared team workflows rather than individual use. These agents run persistently in the cloud, meaning they can continue working on long-running tasks without user interaction and can be triggered on a schedule or deployed directly in Slack. Practical use cases OpenAI highlights include a lead outreach agent that reduced 5 to 6 hours of weekly rep work to automated background processing. Processes, and an accounting agent that handles month-end close tasks including journal entries and variance analysis. On the enterprise control side, they've added several things including admins getting role-based access management, a compliance API for auditing every agent configuration and run, a built-in prompt injection safeguard, and the ability to suspend agents. And another feature, privacy filtering, an open-weight 1.5 billion parameter model detecting and redacting PII in text, available now on Hugging Face and GitHub under the Apache 2.0 license. So, uh, if you're looking for a lightweight built-in option inside of Codex to find privacy PII, uh, this little model sits on top of it and does great work. [39:47] Justin: Yeah, so there's nothing in here that makes me angry because it's not really running in your local work system pretending to be me, right? Like, it's— I think that running on these platforms like, you know, Gemini Enterprise or, or OpenAI Enterprise is the, is the right way to do it and the safe way to do it because they are sandboxed environments and they have access to so much. They're dedicated agent identities that you can control and, and, and give specific permissions to. So I do like, I do like that sort of model for these things. And you know, they're just, now they're making me really happy with the, like the PI filters on top of these things. 'Cause I think that running in those platforms is the only way you can sort of enable, at least that I know of, that you can enable that sort of sidecar AI watchdog. Which is cool. [40:35] Jonathan: And this is great also for not just that, but for— you could integrate this into your product, you know. So if you are taking any data from a customer, you could at least say we're not just sending it all from whatever they type into the LLM. We do have it run through some sort of filter that does attempt to catch it. With the caveat is it's not 100%, but we can at least say, hey, Matt's social security number, 11223344, is not actually going, and that will automatically should catch it and then at least obfuscate it. It's the old DLP coming back in a new way. [41:12] Matt: And it continues with GPT Images 2.0. They're apparently— it's way better at following instructions. It can now handle small text, UI elements, icons, and complex layouts at up to 2K resolution. No more getting something close enough. It actually delivers what you asked for. It works in non-English languages. Previous versions struggled outside of Latin-based text. Now it solidly supports Japanese, Korean, Chinese, Hindi, and Bengali, where language is baked into the design itself. It can think before it generates when paired with a reasoning model. It can search the web, plan the image structure, self-check its work, and even produce multiple distinct images from a single prompt. ImageGen meets Agentic AI. Flexible aspect ratios for any format. Supports everything from wide 3:1 banners to tall 1:3 mobile screens. Useful for social graphics, presentations, posters, and more without manual resizing. And it wants to be your entire creative workflow. The big picture play here is replacing the back and forth between prompting, designing, and editing. And as a person who does a lot of image generation, with ChatGPT and Gemini Nano Banana. Thank God. [42:07] Jonathan: I like you can actually do multiple at the same time. That's a pretty nice feature. 'Cause when you don't necessarily know what you want, you know, I create for less important things than Justin does, like, you know, artwork for the podcast. You know, mine are more like, let me just generate this to send to a coworker to troll them a little bit. Oh, trust me. [42:28] Justin: He does plenty of that as well. [42:29] Jonathan: Generate multiple. But it's nice that it will actually generate multiple of these right now. And then I could tweak it from there to make it as, you know, I don't know how to say it nicely, but you know, as Mattisms as possible for the person I'm sending it to. [42:46] Justin: Matt-tastic, maybe? [42:48] Jonathan: I think it was cynicism. Matt-cisicism? [42:56] Matt: If you were sad when Google got rid of the ability to register domains and you've been using something like Hover or GoDaddy, heaven forbid, or just using Route53 and just pointing it right on over to GCP, which, uh, that's what I do. [43:10] Jonathan: That's what I do. What most people do. [43:12] Matt: You can now use Cloudflare as they have a Cloudflare registrar API now in beta, allowing developers to search, check availability, and register domains programmatically through 3 straightforward API endpoints. Keep the entire workflow inside editors, terminals, or agent-driven tools, or via Terraform, which is also quite nice. API integrates directly with Cloudflare's MCP server, meaning agents and environments like Cursor or Claude code can already discover and call register endpoints without any need for additional integration or custom tool definitions. Also, the hacker MCPs can now automatically create phishing domains off your poorly spelled domain name. So maybe buy some extra domains this week. Yeah. [43:47] Justin: And definitely check your domain certification. [43:54] Matt: Okay, AWS, we're almost through guys with at least the AWS section. [44:01] Justin: We got through the AI section, so that's good. [44:03] Matt: We got through AI, so that was a lot. But first up, AWS Interconnect is now generally available. We previously talked about this, but basically it provides multi-cloud private layer 3 connections between AWS and other cloud providers starting with Google Cloud and with Azure coming later this year. Last mile, this is the last mile for anything on-premise locations at AWS through network providers like Lumen, AT&T, and Megaport as well. The multi-cloud option uses 802.1ae MACsec encryption by default on physical links, route traffic entirely over private backbones without touching the public internet, and includes built-in redundancy across at least two physical facilities. Pricing is a flat hourly rate based on bandwidth tier and region pair, so check the pricing page for sizing your connection. And provisioning is handling through the AWS Direct Connect console in a few clicks. Now, when I said it is supported on GCP, it is not supported on all regions yet. So currently it's US East, US West, and Europe for AWS. And for Google Cloud, it's US East— sorry, sorry, US East, US West, and Europe for Google Cloud. And it's US East North Virginia only currently for AWS. So thanks for that. But good to see it in GA. Hopefully it gets expanded out pretty quickly. We'll ignore all of the future region expansion announcements. Amazon will make for the next 6 months about this. [45:15] Jonathan: They have to hit their blog post quota. [45:17] Matt: Yeah, they do. Uh, so Amazon Quick, uh, came out in October, and I think we mercilessly mocked it for its dumb name, uh, and I had never used it, uh, but today they apparently came out with a new, uh, desktop tool as well as a bunch of features to help marketing intelligence capabilities. So it connects tools like HubSpot, Salesforce, Slack, and Adobe to create a unified knowledge graph from scattered marketing data. All available through the tool directly, as well as they now support multi-account sign-ins for the same browser, which is interesting because it uses QuickSight to support all of its logins. So it's not your normal AWS console account, it is your QuickSight account. If you don't have one of those, you can create one, which is easy to do. And this allows you to basically jump into Quick. And so I downloaded Quick because they also released a desktop client and I never used it. And it's a lot like ChatGPT with more integrations into QuickSight. So, but it, it has ability to schedule agents. It can do repetitive tasks. It can access all the other tools like I just mentioned for Adobe. All of them are out there and available. [46:16] Justin: Yeah, nothing makes me understand things less than something like this where it's Amazon Qwik, which makes sense, ties into QuickSight. Okay. That's, I can kind of get the tie in there. Wait, it's a local client that allows you to just run general AI. You know, agentic workflows. What does that have to do with a BI tool? [46:35] Jonathan: Well, they use the tool that added new marketing to figure out the name of it, which is the problem. [46:43] Justin: And they use the authentication mechanism for Amazon QuickSight, which has been a terrible, like, just wart on Amazon Identity for ages. [46:52] Jonathan: No, you can make this identity set up with your IAM when you set up— [46:57] Justin: Only 'cause you can do that with QuickSight now. [47:00] Matt: Right? [47:00] Jonathan: Right, but at least it's tied to your IAM. [47:04] Justin: Yeah, still have to have a QuickSight account. I don't understand this. Makes no sense to me. Quit doing things that don't make sense. [47:11] Jonathan: Well, I just think at one point they need to figure out how to better, QuickSight needs to go to like IAM or, you know, Amazon SSO identity, whatever they rebranded it to that my brain is not processing right now. Identity Center, you know, and this should tie into there or IAM. It just, it's weird that they have a whole other essentially thing, which what is it under the hood? And is it just Cognito? I'm not sure. [47:40] Justin: Yeah. [47:40] Matt: I don't know. [47:40] Jonathan: I kind of, for some reason, feel like it is like, it has like these little nuances when you play with QuickSight. They always felt like they like, like we're just gonna run Cognito under the hood for this. [47:49] Matt: I mean, it's, I'm, I'm 100% sure it's Cognito under the hood. And then someone said, well, to use this new tool, we have to ride another Cognito pool. And they were like, yeah, or my dead body, just use the QuickSight one. [47:59] Justin: Oh, you're right, that's exactly what happened. [48:06] Matt: They threatened him with Cognito. [48:08] Jonathan: I still remember for a customer we had to do DR for Cognito, and the answer was like, you can't, you can't extract, you can't do backups, you can't do anything. I think you can do a few of the things now, but, and you can't move it. So wherever you set it up, you damn well better hope you set that thing up correctly, 'cause there's no going back. [48:29] Matt: Just use Okta. That's the right answer. Or Ping or anybody who provides you a managed service for this. [48:35] Justin: Yeah, now that they support SAML, it makes it easier. [48:40] Matt: Amazon CloudWatch now lets customers audit telemetry configuration and enable telemetry from services like EC2, VPC, and CloudTrail across multiple regions from a single control point, reducing the operational overhead of managing observability at scale. Scale. IAM rules can be scoped to specific regions or all supported regions, and rules set to cover all regions automatically expand to include new regions as they become available, which is useful for organizations with growing AWS footprints. A practical use case is a central security team creating one organization-wide rule for VPC flow logs that consistently applies across every account and region, eliminating gaps in telemetry coverage that could create blind spots. The feature is available in all AWS commercial regions with standard CloudWatch pricing applying to telemetry ingestion, so costs will scale with the volume of logs and metrics, rather than the feature itself carrying on additional charge. For teams managing multi- account, AWS organization setups. This reduces the risk of misconfigured or missing telemetry in an individual account. [49:30] Justin: Yeah, I mean, this has always been a challenge across, you know, even before I was doing security and trying to do log governance across these things, trying to have, you know, you know, different serving, you know, farms basically in multiple regions and having to log into different webpages to view the metrics, you know, on each one. They, you know, they sort of fix that with the ability to sort of reference metrics in a foreign site a little while ago, but you could only do it for metrics. And so this is definitely something I'm glad to see that you— [50:03] Jonathan: I have a really good tool, Ryan, that you should try. Elasticsearch. [50:08] Matt: Is it? [50:09] Justin: Oh, he said the thing. [50:13] Matt: Why would you do that to him? That's just mean. Yeah, so mean. [50:17] Justin: All right, now I've got all twitchy. [50:19] Jonathan: See, I get to see Ryan's face when I said that, because he was all excited. I was going to say something useful. And then there was just pure sadness there. [50:28] Justin: He's gonna recommend Bindplane or one of these cool tools that actually does the thing. No, it's easy. [50:35] Jonathan: No, no, no. [50:36] Justin: A log aggregator. That's it. [50:38] Jonathan: Yeah. [50:40] Matt: But not. Amazon Bedrock is now automatically attributing inference costs to the IAM principal making the call. Novel thought. With data flowing into the CURR 2.0 format by a new line item IAM principal column. This works across all Bedrock models at no additional cost and requires no changes to your existing workflows. The feature supports 4 distinct access patterns, including direct IAM users or API keys, application roles on AWS Compute, federated identity through providers like Okta or Azure AD, and LLM gateway architectures. Each scenario has different configuration requirements, with the gateway scenario being the most complex since it requires a per-user assumed role session management to avoid all traffic appearing under a single identity. Cost allocation tags can be attached to IAM users or roles or passed dynamically as session tags through identity providers. And once activated in AWS Billing, they appear in Cost Explorer under IAM principal prefix. This enables chargeback reporting by team, project, cost center, or tenant without building custom tracking infrastructure. For organizations using LLM gateways like Lite LLM or Bifrost or custom proxies, the solution requires the gateway to call AssumeRole per user and cache credentials for up to 1 hour, which keeps STS call volume manageable but introduces architectural challenges. Uh, tags take 24 to 40 hours to appear in Cost Explorer and CURR, And I assume it's in the cur 2.0, it'll eventually end up in the focus spec as well. So that'll be nice when that comes. [51:55] Justin: Yeah. It wasn't that long ago that you were pleading for this feature, right? [51:59] Matt: Because you, I mean, I'm still pleading for it on Vertex, but, uh, at least you get it on Amazon Bedrock now. Yeah. [52:04] Justin: Because it is, it's really complicated to see these things and it's, you know, unless you're gonna give everyone their own dedicated Amazon account, like what are you supposed to do? So, yeah. [52:13] Matt: I mean, part of the reason why you'd use a Lambda LLM or Bifrost is to help with this problem of having a gateway do it. So it's sort of weird to me that you would then want to Also add the complexity of now trying to get to map to your billing data other than I'm sure all the billing tools don't have any understanding what an AI Gateway is. No. [52:29] Justin: So, and so it sort of abstracts the problem all over again, right? Like it's— [52:33] Matt: Right, exactly. Yeah. Uh, well we talked about S3 Files last week and I'm pleased to announce that AWS Lambda functions can now mount Amazon S3 buckets as a file system with S3 Files. Thanks. Could have announced that last week. Coming soon to you, Fargate support, I'm sure is next. [52:51] Justin: I mean, this is a really neat, you know, thing if you're, if you know, S3 file works out better than the fuse mounting does. I do really like this. I haven't used it yet, but I would, I would use this because I do think it's a great thing to have in those stateless application components. [53:08] Jonathan: Fargate, I'm trying to think of all the other tools they're going to add. [53:10] Matt: ECS, EKS. [53:11] Jonathan: Oh, nginx, ECS. Yeah, but those are all Fargate, like container solutions. I'm like, okay. [53:17] Matt: They'll add it to, what's that? They'll add it to like Lustre for some reason we don't understand, 'cause we don't know what Lustre really does. You know, things like that. Is it Lustre? It's a process on its own. Yeah, I'm like, I don't know. Or the real S3. They'll add official support in S3 Fuse to S3 files. And you'll be like, what? Isn't that kind of, it's already a bridge to that. So it, like, there's all kinds of ways they can mess with us. [53:42] Justin: That'd be kind of cool, actually. [53:44] Jonathan: It's NetApp will support— [53:46] Matt: oh no, you're right. [53:47] Justin: It will be NetApp. [53:47] Jonathan: It's gonna be NetApp supporting S3 files. [53:50] Matt: But S3 files is NetApp. Exactly. [53:54] Jonathan: I understand. [53:57] Matt: Uh, yeah, that's a mess. Cloud Cowork is a desktop application for Mac and Windows. Lets knowledge workers delegate research, document analysis, data processing, and report generation to Claude. And typically to get Co-Work, you have to be a direct Claude Anthropic customer. But now you can have all the model inference routed through Amazon Bedrock in your AWS account rather than over Anthropic's infrastructure, which could be super helpful. Uh, pricing is consumption-based through your existing AWS agreement with no per-seat licensing from Anthropic, which was a notable distinction from Cloud Enterprise and could make cost modeling more predictable for organizations with variable usage patterns. Enterprise security controls are central to the integration, including AWS IAM or Bedrock API key authentication, VPC endpoint network isolation, CloudTrail audit logging, and OpenTelemetry export to CloudWatch, with Anthropic receiving only aggregated telemetry that can be disabled. Setup relies on device management tools like Jamf, Intune, or Group Policy to push a managed configuration to Cloud Desktop, specifying the model ID, Bedrock inference profile, and auth method, which means IT teams control rollout rather than individual users configuring their own credentials. Organizations can use Cloud— are already using Cloud Code on Bedrock can reuse the same infrastructure setup for CodeWork, both in-region and cross-region inference profiles are supported to address data residency requirements across different geographies. [55:10] Jonathan: I mean, this goes back to with the prior one, you can actually now see where, who's doing what if you have people set up with IAM or anything else. Like this is, you know, I assume these were kind of one was waiting on the other because, and it's gonna be interesting to say, to see, 'cause you know, I've talked to friends and whatnot in the industry where people are like, 30% of your job needs to be using AI. And now with some of the buildback and everything else you can do, and if your company routes everything here, you could very easily see, okay, this person's using X number of tokens here, Y number of tokens, this person's using Y, and start to compare people. But then I think the real question is, how useful is that data? Because maybe they're just not efficient with their tokens and they're having conversations with BoltBot about Qwik versus Qwiksite. [55:59] Justin: No, I mean, it's, for me, it's more along the lines when you, when you use Anthropic's enterprise directly with Anthropic, you lose a lot of visibility into who's doing what. Like you can see tokens usage for users, but you can't really see, you know, what, what permissions are they, you know, doing? What, what are the agents doing? And I think using this, I haven't played around with it directly, but I think it would allow more visibility through the standard Amazon tooling. To, to see those IAM transactions, see what's going on. Calls to network VPC endpoints would be captured by flow logs. And so it, it would offer a lot more tools for, you know, a security org to, you know, route to a SIEM and, and be able to do investigations on anomalous traffic or any kind of learning rules, playbooks, which is not something that you can do with directly using Anthropic or Cloud Enterprise. [56:48] Jonathan: So you're saying Claude Enterprise isn't very enterprisey? [56:52] Matt: I mean, it, it is not in some ways and it is in others. [56:55] Justin: They're trying to make it. It will be with time, right? [56:57] Jonathan: They're working on those things. [56:58] Justin: They're, yeah, it's just not there yet. [57:00] Matt: The problem, the problem is, is that instead of building a proper enterprise backend that would do all the things they want, they partnered with WorkOS. And so while WorkOS has a bunch of things, it doesn't have all the things that you would want. And this is a problem also for OpenAI as well, 'cause they also partnered the same way and Snowflake partners with them. But you know, some have done a better job than others in how they lay out some of these tools. [57:21] Justin: I didn't know Snowflake, 'cause that does answer a lot of my questions 'cause I have a lot of same complaints with Snowflake. [57:28] Matt: Uh-huh. So all these, all these people are using basically this backend platform that helps them manage SaaS apps and it, they're sort of limited to what's available in the tool, which is interesting. Yeah. [57:40] Jonathan: That does make a little bit more sense though. I then go back to the, Well, if Force6 is a SaaS killer, why isn't one of these vendors just develop their own tooling for it? [57:49] Matt: Well, I mean, that's, that's the thing. I was like, Anthropic hasn't killed WorkOS yet and they're using it all over the place. You'd think they just built the right thing. SaaSpageddon be damned. All right. Uh, our final Amazon story, then we get into the fun of GCP. Amazon Bedrock Agent Core now includes a managed agent harness feature that lets developers define the agent's model, tools, and instructions by API calls without writing orchestration code, reducing initial setup from days to minutes. It supports popular frameworks like LangGraph, LLMIndex, CrewAI, and Strands Agents. The new Agent Core CLI available on GitHub keeps the full agent lifecycle in one workflow, covering local prototyping, deployment, and operations from a single terminal with CDK support and Terraform coming very soon. Agent Core now includes persistent session state via durable file systems, enabling agents to suspend mid-task and resume where they left off, which makes human-in-the-loop workflows practical without custom storage plumbing. Prebuilt coding agent skills give tools like Cloud Code and Quro curated knowledge of agent core, uh, best practices rather than just raw API access with plugins for Codex and Cursor coming by the end of April. The managed agent harness is in preview across 4 regions with no additional charges for the CLI, harness, or skills beyond standard resource consumption. [58:56] Justin: I mean, this is a great feature. This is now makes it competitive with Vertex AI's agent builder. And so like, I, you know, now it's a usable option in Amazon. [59:07] Matt: Awesome. I mean, other than it requires you to use npm install, which I hate. Uh, why? [59:13] Justin: Yeah, I know. Me too. [59:14] Matt: But, uh, but you know, maybe I'll end up in Homebrew. A lot of these end up in Homebrew pretty quickly after. [59:19] Justin: I mean, Claude Code, maybe you use npm because, you know, it's evil too. [59:23] Matt: So, well, they at least did move it to Homebrew as well. So you don't have to do that. [59:28] Justin: Oh, I didn't. Yeah, I don't, I haven't installed it since. Cool. [59:31] Matt: All right, GCP, they had several announcements pre-next, so— [59:34] Justin: which is just mean. [59:36] Matt: We already got to cover a conference, come on. We already got to cover your conference yet. So we'll go through these kind of quick. First up, they have a new text-to-speech AI model, Gemini 3.1 Flash TTS, now available in preview and available in all the googly things. The model scored an ELO of 1,211 on Marshall Townes' TTS leaderboard, which I have no idea what that is, but sounds impressive. And so there you go. If you need text-to-speech, Here you go. Solution A. If you've desperately wanted an app in your computer that's a client to access Gemini, which I have wanted, you can now have that wish fulfilled as there's now a desktop client that works on Mac and Windows. So you can now run Gemini natively in your computer without having to go to a webpage because webpages are dumb. Now, if they'd only learn this about Gmail and make an actually good Gmail client, we could all be much happier with our lives. [60:22] Justin: That's really funny. [60:23] Jonathan: Why haven't they? That doesn't— yeah. Yeah. I, like, as you said, I was like, well, they're building it out. Why didn't they do it for other things and take out, you know, Apple Mail or Thunderbird or any of these other ones? [60:36] Matt: Because none of the other people like— no one likes Outlook. So like, you know, everyone else, like, we love Claude, we love Claude Chat, we love ChatGPT, we love all these agents. And like, so Google's like, well, everyone else loves those things. But like, when you talk about email with Gmail and Outlook, everyone hates Outlook. So like, well, we'll give you something different. I guess that's my, that's my theory at least. [60:55] Justin: That's funny actually, cuz I use Outlook cuz I prefer it over the Gmail web interface. Yeah. [61:02] Matt: So I actually found a Gmail client that I actually like finally that talks native Gmail on the backend. So it doesn't have all the weird IMAP BS. Uh, and I, I do like it. [61:11] Justin: That's the only problem with Outlook and Gmail. [61:13] Matt: Yeah. Yeah. [61:14] Jonathan: That's the big problem. And that's why I've tried them multiple times and I'm always like, nah, I'm good. [61:19] Matt: Yeah, so it's a, it's a product called MimeStream, which is a terrible name. And it's like, I think it's $20 a year and I've been using it and I love it. So, oh, cool. [61:29] Justin: Looking for it. That's a good option. [61:31] Matt: Next up is they created expert content with an ability to deploy multi-agent system with Terraform and Cloud Run. Google's Dev Signal is a 4-part tutorial series showing how to build and deploy production multi-agent system using Google ADK, MCP, Vertex, IAM, Memory Bank, and Cloud Run. So if you've been confused how to do this, they've given you a full white paper on how to do it. Which only because of everything else they announced next, I thought we should talk about it. [61:55] Jonathan: So you're welcome. [61:58] Matt: All right, let's start with the most important thing, the conference. We're here, 32,000 people, 3 keynotes, 25 spotlights, 700 breakouts, 260 announcements. 260 announcements, not all on the main stage, thank God. But, you know, the most important thing is that points were awarded. And, uh, I had to say, guys, we were competitive this year. Yeah. So, uh, first up, uh, I went first and I nailed Wiz and GCP security, Anthropic defense, and a Wiz AI app, and Azure, of course, they acknowledged on stage. All hit beautifully in the first day. I was very happy. I did hope for a new Antigravity plus Gemini CLI capability. I didn't get it. They did give us Data Agent Kit. I am not taking a point for it. I don't think it's legit as a point, but basically they give you inside of Gemini CLI access to the data agent kit, which makes it easier to talk to BigQuery and to other tooling on the data side. It's a nice tool. I just didn't think it was really fitting what I intended of this. And since I was being nice and I didn't wanna win over technicality, I said no. [63:00] Justin: I would argue that one to the death too. [63:02] Matt: Oh yeah. As you should have, cuz I don't think it would be fair. And then I got an Ironwood successor, which they overdelivered on with a TPU, two TPUs, one for training, a TPU 8T. And a TPU 8i for inference. So I got 2 TPUs for the price of 1. So pretty nice. Next up, Ryan came in strong with Gemini 3.1 Pro general availability and the future model teasers. So they talked about several new models coming later this year. So nicely done. And they gave you Agentic enhancements. Basically the entire keynote was Ryan. The entire conference, the whole thing. The age of Agentic is definitely here. Yeah. Would you be— [63:39] Jonathan: I feel like we should have made him define that one more next time. [63:44] Matt: I mean, he was looking for an agentic capability and they delivered a bunch of agentic capabilities. So I was at Tuesday, I was at an infrastructure summit that I was invited to and they go, and we're gonna GA GKE agent. I was like, son of a bitch. Because I'm like, that's gonna help Ryan right there. And then the next day I saw the opening of the keynote and I was like, oh no, Ryan's gonna own this. Yeah, uh, but they did not give you a VMware or Kubernetes interruption play. The closest thing comes to it is maybe Spanner Omni, if I were to really stretch at it. But even then, not even— I won't argue that. Yeah. And then Matt, you did get, uh, 2 points as well. So default guardrails for agent identity, agent gateway, and model armor. Uh, I mean, they named all 3 of those on this page. So that was a— that's a clean— cleanest point on here. There's no wiggle on that one. Uh, and then they did give a Gentec SDLC through Data Agent Kit, a Gentec Task Force, and, uh, Task Force was really the key piece you needed for the SDLC. So congratulations on that. I will technically tell you that you did also win 3 non-AI announcements at the conference. It just did not mention them on the stage other than Virgo Network was the one thing. And even Virgo Network, I would argue, is pretty close to basically being a, uh, a Gentec feature. So, you know, but, uh, you're close on that one. So, you know, nice job everyone. Congratulations. A round of applause. Now it takes us to the tiebreaker to see who wins the game. Now I just want to point out that, you know, we chose how many times to say AI or artificial intelligence on stage. Now I will tell you, they did not say artificial intelligence very much. It was mostly AI. But in the first keynote, they said it 132 times. Yeah, yeah, double what they said in last year's. As I went back and look, I was like, this is crazy. Uh, cause like I won this last year too with like 90 or something, which was, you know, still crazy high. And so 132 times in the first keynote and the second keynote 55 times. And so because of that, I went high, which I wasn't going to go high. I was going to go between Matt and Ryan. Uh, I stole the win from Matt with 115, which is Price is Right rules. I did not go over. And so therefore I won this one by 7. I was within 17, pretty close. [65:54] Justin: Yeah, I remember I was listening to the keynote and I wasn't exactly counting, but it was just AI, AI, AI. Like, I knew in my head that if it came down to a tiebreaker, I didn't win. And I, I really did think of how much they said it the year before, that, that maybe there'd be some sort of, you know, I mean, at least Amazon took some of the feedback from last year and they added non-AI announcements to their keynote. [66:17] Matt: I— Google needs to take a similar playbook. Yeah. [66:21] Justin: I don't remember anything. Like, even the number of announcements you said, I'm sort of surprised to hear that number because it didn't feel that listening to it. [66:28] Matt: Well, it definitely wasn't on stage. I mean, the reality is they announced a ton of stuff in sessions and all over the conference. But, you know, from a pure, like, for the main crux of the conference, like, unless you cared about Agentic, you cared nothing about the keynotes. No, exactly. Yeah. Which was kind of a shame because I feel like it really diminishes the power of cloud. And like, they announced a lot of other cool stuff and we'll talk about some of them here in a little bit. But you know, in general it's just like, you gotta mix it up. I, I just, you know, AI's cool. I get that you have a huge need to please your investors. I know everyone's excited about Agentic. I get all that. Just don't make it everything. [67:05] Justin: Everything. I'm exhausted. [67:07] Matt: Yeah. I mean, to the point where I'm, I'm not even considering going to Next next year. I'm just like, if it's just gonna be an Agentic or AI disaster like this, I don't know that I want to. And it just like the, and the energy was so off this year, I felt like, because A, The reality is a lot of people are worried about Agentic killing their job. So we're literally at the conference to learn how Agentic's gonna destroy your job. [67:27] Justin: Mm-hmm. [67:27] Matt: Which is kind of this depressive loop cycle that you're at. Like, we should be having a good time at this conference celebrating what we do and learning a bunch of new cool stuff. And it's like, no, I'm learning about how my job's gonna be destroyed. [67:36] Justin: Cool. [67:36] Jonathan: Thanks. [67:36] Matt: Yeah, I think that's part— [67:38] Justin: I, I've been really thinking about this since the end of the conference is trying to figure out like a lot of the Agentic stuff that they talked about, like I have a hard time applying directly. And so like, it does kind of feel that way. Like You know, it is all of the things that you can do with the Gentec Workflows now are generally replacing the people in the middle. Like it's not about new capabilities, it's not about, it's just about doing the same things faster. And it, it left me like sort of uninspired, which usually for these cloud conferences, I come out and I wanna do all the things and play with all the new toys and I can't wait to introduce 'em to the rest of the business and try to, you know, get the, get the, you know, budget and time to roll 'em out. Um, but this one I really didn't. [68:18] Matt: Yeah, it was kind of a shame. All right. So then we had to deal with the 260 announcements and how do we even tackle them? And so with Claude, uh, we, we assessed, uh, and we tried to rank them by buzz, which we, we analyzed Twitter and we saw what people were talking about and tweeting about, uh, with a strategic significance to AI and all those things. And then of course the impact to you, the practitioner. And so we've, we put them into an S tier, uh, sorry, into a tier ranking, basically S, A, B, and C. We didn't have any Ds or Fs. Which is not normal for ranking, but it's kind of hard to do on new tools. I don't even know what they are yet to say they're F tier. We should probably maybe do an, do a bonus episode when we do bonus shows, uh, on just like ranking our favorite services or something. Yeah, that'd be kind of fun. It might take us hours, but it'd be exciting. So we'll jump into this. So tier S, uh, first up is the story of the keynote. We just talked about it, uh, Agentic. And the most Agentic part of the whole thing is that they're repositioning Vertex AI as Gemini Enterprise agent platform. Uh, so basically they're killing the Vertex branding in favor of Gemini Enterprise, and then basically something after that. So there's 16 named sub-features in this: Build, which includes your ADK, Agent Studio, and Agent Designer. You got Run for agent runtime, agent sandbox, memory bank, sessions, and long-running agents. And the memory feature they built out is pretty impressive. Uh, it's in the developer keynote. They have a pretty good demo of it. Uh, Govern and Optimize for agent identity, uh, which is a cryptographic ID per agent, agent registry, agent gateway. Anomaly detection, security dashboard, simulation, evaluation, and optimization all inside of that. So basically their basic strategic frame was agent is now the unit of work, not the model call itself. And it'll be interesting to see how this continues to drive over the next year as this continues to evolve. [70:02] Justin: Yeah, no, I was excited by especially the identity stuff 'cause I'm a security nerd. It was very cool. [70:09] Matt: The next one, you know, customer scale. There's lots of mentions and this is one of the things that the 260 is a bit of a cheat. If you look at the blog post, there's a lot of customer successes, which is great. And I, I love good customers, but the most important thing in this is that there's actually a lot of customers doing a lot of things with Gemini Enterprise, uh, the GenTech platforming. So Mars was highlighted, Merck was highlighted, GE had 800+ agents across manufacturing, logistics, and supply chains. Just some really good case studies. So if you're looking for how do I use this thing, there's probably a case study for your industry, which I think is always helpful to know, like, what are other people doing? They did in the keynote have Home Depot's Magic Apron basically talking about how they use it at Home Depot. It's kind of an interesting idea. Again, Virgin Voyages was the other one with Rovi where they kind of COVID a bunch of the things they're doing with these type of capabilities. And so if you're looking for real life examples of how to use agents in your day to day, there's a case study for you. I guarantee it. [71:01] Justin: They never get specific enough for me on these, these case studies. [71:04] Matt: I want a lot of times you just need ideation, right? Yeah. No, I get, I get enough of that gist of it. Like, oh, this is cool, I could see how I could apply that to my thing. That's basically what you're trying to do. [71:14] Jonathan: That's kind of the way I use them is like, okay, oh, there's this thing over here. Oh, I could do this and tweak in these 16 ways and then it's useful for my business. [71:23] Matt: TPU 8T and 8I, definitely an S-tier item. You know, 3x compute versus Ironwood on the training side and the inference being purpose-built, 80% better per dollar optimized for MOE and agentic workloads. The Torch TPU native PyTorch full eager mode kills the JAX-only friction they were having before. And they're the only hyperscaler shipping dedicated inference silicon in this generation, which is a bit of a stretch because there is the Inferentia chips over at Amazon. But since they didn't get updated, they have a chance to say these are the most current generation. So well done marketing at Google for calling that out. Those are the S-tier items from the announcement. Going into tier A, uh, Wiz. You know, of course the acquisition formally closed right before the conference. The Wiz AI app, which is code-to-cloud-to-runtime AI application protection platform, uh, Wiz now supports AWS Agent Core, Azure Copilot Studio, Salesforce AgentForce, and Databricks. And I assume we'll also have Snowflake coming as well. So if you're looking for agent control, Wiz has a bunch of stuff for you there. Other Wiz news across the conference: inline AI security hooks in IDEs, Wiz Skills, or virtual validated attack surface findings exposed to coding agents for auto-remediation, An AI bill of materials for auto inventory of every AI framework model, IDE extension used across your environment to kill all the shadow AI, and Lovable Vibe coding integration for security scanning inside of Lovable. I mean, this is going to be super interesting how it pays off over the next year or two or three or five or decade, but lots of cool opportunities in Wiz. And my interest in Wiz suddenly increased a lot after the conference. [72:53] Justin: Mine too. Yeah, specifically the agents in all these ecosystems, right? That's a, that's a big problem. Agent sprawl is a thing. So exactly. [73:02] Jonathan: You don't trust agents? Come on, what's wrong with you? [73:04] Justin: I do not. [73:07] Matt: Uh, so one of the things that we've been hearing a lot, if you've been tracking some of these things, is ChatGPT or OpenAI, they'll actually put a forward-facing engineering team or forward-deployed engineering team. Same thing with Anthropic. So they'll basically get a big customer like an Accenture or a a McKesson or somebody, they'll come in and basically say, hey, we'll put a bunch of engineers into your business to help accelerate your AI agent story and really help drive forward. And so Google has gotten into this now with the Partner Fund, a $750 million innovation fund for partner agent development. This allows you to get agents built into the agent marketplace and the agent gallery with already 70+ partner-built agents out there at launch from big companies like Atlassian and Lovable and Palo Alto. Salesforce, ServiceNow, Workday, etc. Uh, they do have forward deployed engineers at Accenture, Deloitte, McKinsey, and Google is making its own engineers available through the partner go-to-market strategy. Very Palantir-style move. So well done. Yeah, that's good. You're looking for help, talk to your Google rep. [74:07] Jonathan: I think AWS has been pushing this for a while too, so I'm kind of surprised that they're just doing it, or are they just like— do they just expand it and are highlighting— [74:14] Matt: I mean, they're committing $750 million to it. They were doing it 4, so, oh, that's what I figured. More money to it. Yep. Gemini, uh, or sorry, in the Antigravity Data Agent Kit, Gemini 3.1. So 3.1 Pro and preview across Vertex Gemini, Antigravity, Android Studio Gemini, and AI Studio. The Data Agent Kit, which is a portable suite of skills, MCV tools, plugins, and tools for VS Code and Gemini CLI into native data workspaces like BigQuery. Full stack vibe coding from AI Studio to Cloud Run, which that one, uh, Ryan, you should kill. And so, you know, from an engineering developer side, this is all about, you know, competing with Cursor, Claude, and Replit out there. If you're excited about data and you're looking for more Dentic in your data cloud, you've got Knowledge Catalog, Cross-Cloud Lakehouse, and Spanner Omni. So the Knowledge Catalog is a universal context engine maps business meaning across the data state. and is a great foundation for accurate agent execution. The Cross-Cloud Lakehouse, formerly named Big Lake, is the Iceberg REST catalog and now federates with AWS Glue, Databricks, Snowflake, SAP, and cross-cloud caching through the, the interconnect that we talked about will actually cut your egress costs with this tool. And then Spanner Omni is their Spanner running multi-cloud, on-prem, or even on a laptop. This is the most underrated announcement of the keynote in my opinion. You can now run Spanner on your laptop, which is great from a local development perspective. [75:39] Jonathan: Yeah, that's pretty cool. [75:40] Justin: Yep. [75:41] Matt: So glad to see that one. And then Lakehouse Federation for DynamoDB, live joins between transactional and analytical workload without doing the ETL first. So you don't have to do things like AWS Glue to move your data. [75:53] A: Thank God. [75:54] Matt: But the, the cross-cloud lakehouse, that's a big one. Being able to access your data inside of Snowflake or inside of of Databricks from Cloud Lakehouse means you don't have to move that data around, which is a big sustainability thing. It also allows you to take advantage of different tools. So hey, maybe you bought Snowflake first and now you're looking at some Databricks agent capabilities. You're like, well, I don't want to move all my data. Well, you don't have to with this federation. So the advantage of having Iceberg as a standard across the industry now makes this a powerful thing that you can make Iceberg endpoints available to everything and pull it all together with a single lakehouse type configuration. [76:29] Justin: Yeah, I, I couldn't love this more. I mean, there's so many datasets that are splattered out like that, then to— if you have questions that need to be answered across them, like, it's such a pain. So it's so great. [76:41] Matt: Uh, I did see some of these features rolling out in Tier B here, uh, on Google Workspace, uh, in the admin. I was kind of like, oh, I see some of the things coming together. But, uh, if you're a Workspace admin, which is probably not most of our listeners, They've given you AI capabilities, unified semantic understanding across Docs, Slides, Gmail, Projects, and Org domains. Workspace Studio, no-code agent builder skills deployable across your workspace. Microsoft 365 to Workspace migration tool to make it easier to move than ever before. Sovereign controls plus client-side encryption for, uh, all your US and EU sovereignty concerns. And Autobrowse with Gemini now available to you in Chrome Enterprise. So that's all available to you out of the box. Kubernetes. Cloud Run got a bunch of features, including the ability to deploy vibe-coded apps from AI Studio. Ooh. NVIDIA RTX Pro 6000 Blackwell support. So now it can run 70 billion parameter models without managing GPU infrastructure through Cloud Run. Billing caps. You can set max monthly spend, resources deactivate when hit. Cloud Run sandboxes for ephemeral isolated agent execution. SSH into running containers now in preview. And hot take, Cloud Run, is going to be the default run agent runtime behind after GKE, in my opinion. [77:51] Justin: Yeah, I, I saw many, uh, labs on how to run multi-agent workflows on Cloud Run. Um, and it's, it definitely looked very alluring. I, I'm going to play with it. I've been using Cloud Run a lot. So, um, I'm excited by this. You know, we make jokes about the full stack vibe coding, but, uh, if this is running in my cloud ecosystem and where I have visibility in that, it's a lot better than running than whatever's running in Ripplet or Lovable, which I have no visibility or any insight into what it's doing. [78:18] Matt: So. Agreed. [78:20] Jonathan: Mm-hmm. [78:22] Matt: And then finally, BigQuery AI was our last in this tier. Uh, AI parse document, single SQL function for OCR plus layout plus chunking via Gemini layout parser, tabular FM, BigQuery graph support, reverse ETL, connected sheets with Times FM, BigQuery hybrid searching, and 35% year-over-year performance improvement with lower processing in cost. So that's not bad. Uh, lots of work for data teams to get into though. And then, uh, our Tier C, uh, lightning round, uh, the Virgo Network. If you're doing any type of training of any kind, this is amazing for you. It's a custom interconnect with 134,000 TPUs can be bridged together into a single Fabric across, uh, 1 million plus across sites. A5X with NVIDIA VERA Rubin NVL72 with up to 960,000 GPUs cross-site. And they can scale further than anybody else, they said, which is impressive. [79:12] Justin: 130,000 TPUs is a lot for me to just ask, you know, AI what color I should paint my living room. [79:20] Matt: Yeah, it's a lot, a lot of water to destroy. Rapid storage capabilities, Rapid Buckets, Rapid Cache, and Managed Lustre all available to you. So you can get faster snapshotting, faster capabilities there. [79:32] Jonathan: The Rapid Buckets is kind of cool, you know, especially with these datasets, you know, 15 terabytes per second bandwidth. Is one, just an impressive number to say. Right, that's what I thought. But you know, forget that, 'cause as I was looking at the notes, I swore it was a typo. You know, 15 gigabytes a second, I was like, okay, that's cool, but 15 terabytes, you know. But I assume that's competing with the S3, like one zone. [79:57] Matt: Correct. [79:58] Jonathan: Single zone, whatever it is, like the, where it's tied to a single zone for speed. [80:01] Justin: I don't remember what they called it. [80:03] Jonathan: Express zone, I think. [80:04] Matt: I think Express, yeah. [80:05] Jonathan: It's too hard to keep all the naming conventions. [80:07] Matt: It is. [80:08] Justin: It is. [80:09] Matt: Uh, Axion, which is of course their custom ARM silicon. The N4a is now generally available, uh, which gives you 2x price performance versus the x86 hardware and 30% better performance for GKE agent sandbox versus other hyperscalers. There's a new C4a metal, uh, instance available to you in preview. So if you wanted, uh, Axion, but you need the bare metal for some reason, you have that. and now you can get, uh, confidential computing on the G4, the C4, uh, for all your confidential AI workloads for the spooks if you need it. Uh, reCAPTCHA, uh, which is probably being killed by AI every moment of the day now because of all the image recognition stuff, I can't imagine that that's lasting, has been rebranded into the new Fraud Defense, uh, package, which is now a platform that distinguishes bots, humans, and agents as agents, their capabilities coming in for the digital commerce journey for account to payment to checkout. And this is the closest, uh, of anything that got to our AP2 idea, but, uh, no mention of AP2 anywhere at the conference. So, bummer. Post-quantum crypto had a little bit of appearance with KMS quantum safe key imports capability, PQC, and cross-cloud networking, which is great. And then GKE got several, uh, quality of life improvements too, including a 4x faster node startup, 80% faster pod startup, and a 5x faster model loading. The GKE hypercluster, uh, is now available with a single control plane, millions of accelerators, and multiple region. In private general availability. Predictive latency boost in GKE inference gateway, up to 7% lower time to first token. A KV cache tiering across RAM and local SSD cloud storage for Lustre and RL scheduler, RL sandbox and RL observability for reinforcement learning workloads on top of GKE. So nice. So in general, I think the big thing, agent platform is the new operating system. Vertex, now Gemini Enterprise agent platform isn't cosmetic. Wiz is now going to be a huge part of their push with Mandiant into the security of your entire organization, and they're gonna— they're showing it already with already heavy integration into agents. It was pretty impressive, and it's not all vaporware. Lots of good customers up there. Yeah, uh, you know, things that were kind of missing— I didn't see anything for robots. There was a Gemini Robotics area on the, on the conference floor. which was interesting. I didn't really see a nano banana update. It was mentioned on stage, but no update to it. No answer to Glasswing or Mythos, no TurboQuant and Vertex yet. So maybe those will come later this year or next year's Next, and out there. And then these are things, a couple of things here below the cut we didn't think were worth even putting on the chart. So there you go. That's, that was Claude and I's work on trying to make sense of this massive amount of announcements. [82:43] Justin: It was, yeah, it is not easy. [82:46] Jonathan: Bravo, Justin, for going through all that. [82:48] Justin: Bravo. [82:48] Matt: Yeah, it was a lot. We have all of the deep dive articles on all of these in our show notes. So if you're curious to read more about any of these capabilities, you can find all of the articles from the conference. There are a couple that I'd like to jump into that didn't make the cut up there, but are cool. Google has announced an official agent skill repository. This basically is launching with 13 skills covering products like BigQuery, Cloud Run, GKE, Firebase, and Gemini API. This teaches your agent how not to be stupid with Google services, which I really appreciate. So the fact that this is now available to you as officially sanctioned with lots more coming later this year, I think that's a really great enhancement and something that you should be plugging into all of your agents if you're doing something with, with Google Cloud for sure. Gemini Cloud Assistant, which last year we kind of panned a little bit. It's kind of your SRE agent, if you will. Is moving from a reactive system to a proactive operational platform, using agentic architecture to handle tasks like infrastructure troubleshooting, cost anomaly detection, and application design without waiting for prompts. The redesigned Application Design Center lets teams describe infrastructure goals in plain language and get back visual architectures with deployable Terraform templates integrated with Security Command Center to enforce organizational policies from the start. A 24/7 FinOps agent monitoring for cost anomalies and correlates spending spikes with specific triggers like auto-scaling events, GCP server support extends Gemini Cloud Assist beyond the Google Cloud Console into IDEs, CLIs, and third-party tools like ServiceNow and Slack. And, uh, Petco was their case study on this one, reported a 60% reduction in Google Cloud-related questions to their cloud team after adopting Gemini Cloud Assist. Would have loved to have had this at a prior life where we had a lot of Amazon accounts. Oh yeah. [84:26] Justin: I, I missed that this sort of tied into the Security Center Enterprise, so I'll have to play around with that. [84:31] Matt: Pretty cool. Well, the fact that you can have it, you know, designed through Application Design Center, you describe what you want and then have it basically give you a diagram and how it needs to tie into the command center. It's a pretty nice— [84:40] Justin: it's pretty sweet. Yeah, yeah, yeah. They are going to replace my job. Sweet. [84:48] Matt: It will be as grumpy as you though, that's the question. [84:51] Justin: Yeah, will it be, will it be as entertaining? No, no, no, no, no. Well, no. [84:57] Matt: All right, that's Google Next. That was a lot. We got there. [85:01] Justin: Yeah, that was crazy. I know I was definitely exhausted by the end of the week, as I usually am. I'm staying an extra day. [85:08] Matt: I mean, that might be the staying up too late every night problem. [85:10] Jonathan: I was gonna say, is that definitely the Outspin, or is that the extracurricular activities? [85:14] Matt: Well, it's the running around. [85:15] Justin: You're doing miles of walking all day, and then yes, you're going out and you're meeting with people. [85:19] Matt: I average like 15,000 steps every day, so just running around, you know, all the places. Um, it, it is a tire— and then, you know, it's dry and I think it was like the driest trip to Vegas I've ever had. It was like, you know, I just was dying. My sciences were killing me all week. So yeah, it was a, it was, I was glad to go home, to be honest. Yeah. Hit the, hit the moisture of the bay. You got off the plane like, ah, I can breathe again. So it's nice. [85:45] Justin: I do think the sessions were well organized. I didn't, you know, like things I wanted to go to, I could get, I had time in between. They weren't spread out everywhere in different wings of the conference center, with the one exception of some of the EBCs being in two locations. [86:00] Matt: Oh yeah, you had to find the secret compute one that's in the corner of the show floor. Oh great, thank you for that. I mean, the one thing I, I continue to feel like they've outgrown Mandalay. I thought that last year, it felt that way again this year. They did a lot of things to try to help crowd control and to try to make it easier, but You know, like the, they had how many breakout overflow rooms for the, for the arena? I mean, like, I think I counted at least 6 plus the FinOps X had their own, you know, viewing party they were doing in their restaurant they ran out there at the event. [86:35] Justin: I didn't know that. [86:35] Matt: That's fun. So, you know, I think it's just, you know, they are gonna be there again next year. They've already announced dates and everything and they'll be back. But I, I can't see them continuing to do Mandalay for more than a year or two more. I mean, especially if, they're as successful as they want to be with Wiz and Agentic, I, I think there's gonna be a lot more interest in this conference going forward. [86:54] Justin: So, I mean, it's definitely full, but I don't know if it's big enough to take over 2 yet. [86:58] Matt: I mean, I don't think, I don't think it's a takeover 2, but I like move it to the actual convention center or to the SANS convention center. Like those are much bigger facilities than Mandalay. I mean, Mandalay is a great facility, but it's not, it's not the size of Venetian or, or the actual convention center in Vegas. [87:14] Jonathan: So the problem is once you have more than one spot moving, like I remember the first year Amazon, they still were at SANS and they had like this one thing at the other one. You just, you don't make it over there in time. There's no time between stuff. So it's like you don't want to outgrow it because as soon as you do, it's too big. I like what Amazon started doing, which is splitting up the conferences a little bit into like the security conference. [87:38] Matt: And then they stopped doing that. They, they've not, they're not doing a re:Inforce anymore. They stopped. So like, right. [87:43] Jonathan: But I like that idea because then you can leave this for general and maybe shrink it down. But I think no one went to the other conferences, so what's the problem? Yeah. [87:51] Justin: I think they should do that at re:Invent. They should have like all the, the different hotels and stuff they spread out to. They should, and they try to do it a little bit, like they try to keep all the security breakouts in one section, all the compute. Yeah. But I think they should like double down on that model and, and figure out how to like have, you know, uh, overflow rooms per, per center for all the, the main keynotes and, and really have them be a little bit more self-contained would be kind of neat. [88:15] Matt: My, my thing about the re:Invent was the Sands, if you go out the back door, is right next to the Link, uh, which has a monorail station and the monorail goes right to the big convention center. And so when they decided to, they were outgrown, you know, Sands, why they didn't just go to the convention center and get both? Like, very similar to our Convex would do, and then just have, you know, free tickets on the monorail to take you over there. Instead, they said, well, we're gonna do MGM, and we're gonna do, uh, Aria, and we're gonna do the former Mirage, and we're gonna have all these things and all these buses. And that was a nightmare. That was terrible. So like, if you could figure out how to simplify between two locations, I think two locations is workable. Seven locations that year, that was the year I swore off re:Invent. [89:01] Justin: Yeah, that was my The last two. [89:03] Jonathan: Was it the first year they went to like more than two? [89:06] Justin: It was the second. [89:07] Matt: The first year we kind of gave the benefit of the doubt that they'll figure this out. And then the second year— [89:10] Jonathan: first year was terrible. Yeah, yeah, yeah. I just remember I went to go do something and I saw the line and it must have been like 2,000 or 3,000 people deep to go from one place to the other. I was like, I'm just walking, like, I'll make it there faster. [89:24] Justin: And he did. Probably. [89:27] Jonathan: Might have stopped for a beer on the way, so probably not. But, you know, it's fine. [89:30] Matt: Well, there's all kinds of frozen drink machines. We joked with Ryan about that. He should have one. [89:36] Justin: All right. [89:37] Matt: Well, people are getting tired of us talking. Seriously. So let's get through Azure here to wrap this up for the week. Azure has announced optimized object storage costs automatically with Smart Tier. This is a smart tier for blob and data lake storage is generally available, automatically moving objects between hot, cool, and cold tiers based on actual access patterns. Data inactive for 30 days shifts to cool, then cold after 160 days, and immediately returns to hot upon reaccess with no retrieval or early deletion charges. The feature eliminates the need to manually configure and maintain lifecycle rules, which is particularly useful for organizations managing large analytics workloads, telemetry data, or data lakes with unpredictable data access patterns. So pricing includes standard hot, cool, and cold capacities with no tier transition fees, but a per-object monthly monitoring fee applies to objects managed by Smart Tier. So thanks, you finally got what Amazon's had for a while. [90:25] Jonathan: Like 10 years, I feel like intelligent tiering. [90:27] Matt: I mean, it might be 10 years. I don't feel like it's been that long, but it could be. [90:32] Jonathan: You can ask Bullpot as we're doing it live. [90:35] Justin: Yeah. [90:35] Matt: What's this? Microsoft Entra ID is adding synced passkeys, passkey profiles, and phish-resistant MFA support for Linux SSO, giving organizations more options to move away from passwords while meeting compliance requirements for stronger authentication. Hey, can we get this, Ryan? Yes. Immediately. [90:49] Justin: Yeah. [90:49] Matt: Yeah. Last passwords, last password changes. I would like. [90:53] Justin: You might have to fix AD. [90:55] Jonathan: Yeah. [90:55] Matt: Starting June 1st, 2026, Entra Connect Sync and Cloud Sync will block hard match operations for users with assigned Entra roles, closing a potential attack path where on-premise AD attribute manipulation could be used to take over privileged cloud accounts. Well, that's a fun bug. Admins should review their hybrid configurations before that date. Why is it? Why not till June 1st? Yeah. [91:14] Justin: What? [91:14] Matt: Do this today. [91:15] Jonathan: Told everyone the problem. I think first they're turning it on by default. So maybe it's that. Maybe people time to do it themselves. [91:25] Matt: Yeah, that could be. I was gonna turn— you should turn this on too, Ryan. Microsoft Authenticator app now includes jailbreak and root detection for Android with a phased rollout moving from warning to blocking to wipe mode. The users on non-compliant devices will eventually lose access to Entra credentials entirely. I'm sure people are going to love that. [91:41] Justin: Yeah. Yeah. [91:42] Matt: That, that one's gonna make people happy. That one's not gonna go over. [91:44] Jonathan: Yeah. Ryan, you should turn this one on. [91:47] Justin: Uh, I haven't jailbroke my phone. [91:49] Matt: It's not my problem. Yeah. I mean, I, I'm an iOS user. I don't do that. You might with your Android phone. [91:55] Jonathan: Mine's not. No, I'm just on the straight Pixel. It's easier. [91:58] Matt: All right. [91:58] Jonathan: I used to jailbreak and it's not worth it anymore. [92:01] Matt: Yeah. I tried jailbreaking an Apple phone one time and I was like, this is terrible. I'm never doing this again. And I undid it. Agent management is consolidating under Agent 365 as the single control plane, with existing Entra Admin Center agent registry and collection blaze retiring May 1st, 2026, and current registry graph API being deprecated and replaced. I mean, I don't know what any of that meant, but happy retirement. Uh, and then finally, Entra ID governance adds support, several notable features this quarter, including SCIM 2.0 API support, delegated workflow management and lifecycle workflows, and a new billing meter for guest users. Which organizations rely on governance features for external identities should review for potential cost impact. Does SCIM 2.0 include ability to sync now? Because that's my one big complaint about SCIM is it's always like, eventually I'll get your accounts provisioned. It's never instantaneous and it never works as fast as I need it to. [92:50] Justin: And you always have to choose between SCIM and on-the-fly provisioning, right? [92:53] Matt: You can't have both. [92:54] Justin: That's always so frustrating. So I hope it does. I don't actually know what's in SCIM 2.0, but if anything's better. but I am really excited about, even if it's not the increased workflow management of that, means you could, means you can do it. [93:06] Matt: Agreed. Well, thanks. I normally don't like Entra updates, but those ones seem good. Azure S3 agent now supports Log Analytics and Application Insights. What did you do before? As native connectors, allowing the agent to run KQL queries directly against WorkSpaces and App Insights resources during incident investigations, replacing the previous approach of shelling out to Azure CLI. Really? You shelled out to Azure CLI before? Yep. [93:29] Justin: Pipe grip. [93:32] Jonathan: Uh, yep, same things that everyone's always done. [93:37] Matt: Great, thanks, that's awesome. No wonder why your tokens get used so fast. Yeah, exactly. Setup is simplified compared to manual RBAC approach, selecting a resource from the dropdown, unprovisioned grants to agent managed identity log analytics reader and monitor reader on the target resource group, but manual entry fallback to resource server fails. Then the AI that removes the permissions or move those later, but that's fun, right? Uh, the feature is backed by the Azure MCP server using the monitor namespace, giving the agent read-only tools like monitor workspace log query and monitor table list. Practical use cases include EKS cluster investigations where the agent can automatically query container log queue events and application traces. So they really want you to burn tokens. [94:11] Justin: Well, but that said, that's the only way you'll ever understand Kubernetes logging. [94:15] Matt: That's true. That's true as well. [94:16] Jonathan: Yeah. [94:19] Matt: Uh, all right. Well, I mean, I don't log on Azure ever, so good luck to you, Matt. I hope those work out well for you. Thank you. [94:26] Jonathan: I think I'm not really sure. I mean, if I'm going to shell, it's gonna be better. [94:31] Matt: Yeah. And finally, I look forward to hearing how your Azure Key Vault retirement plan is gonna go, as apparently Azure Key Vault is retiring its legacy HSM Platform 1 on September 15th, 2028, and customers using Microsoft Purview Information Protection will bring your own key, need to migrate their tenant root keys to the modern FIPS 140-2 Level 3 certified HSM platform before that date. 'Cause who loves a good HSM migration? I don't. [94:55] Justin: Yeah, no, those, I can't imagine that's easy. I've never actually done one, which, because why would you? [95:00] Matt: I don't ever want to do it if I don't have to. [95:04] Justin: Exactly. [95:04] Jonathan: I feel like that's one of those, I may go get a new job before I have to go deal with this. [95:09] Matt: Yeah. [95:09] Jonathan: Yeah. [95:09] Matt: 2028, you'll get your time. So that's good. [95:12] Jonathan: The thing is Microsoft does give you normally a decent amount of time to do stuff, but what's always fun is if you buy I don't know, a 3-year reservation, you're stuck with it and you have to deal with returning it right now because otherwise you would have negative time once it's there. You know, unrelated to, I don't know if they do it for HSM1 platforms or not, but I've been burned by that a few times now. [95:36] Matt: Lovely. Well, gentlemen, we made it to the end of this, uh, Google marathon. [95:41] Jonathan: Wow. [95:41] Justin: I didn't think it was gonna happen. [95:43] Jonathan: And AI marathon. I feel like it was very top and middle heavy this week. Yeah. [95:47] Matt: It was, yes. Well, AI, I mean, all the AI players wanted to make, get their announcements out before Google dropped all theirs. So, you know, I, I didn't see that coming, but now in hindsight I'm like, oh yeah, that makes sense. [95:57] Justin: If we had sponsors, you know, trying to throw money at us, we'd probably have The Cloud Pod and the AI Pod. We'd have to break it all out. [96:04] Matt: I don't know if I could do, we can't get you guys together for one podcast recording every week. [96:07] Justin: Oh, I mean, if that's what I'm saying, they'd be happy to throw a bunch of money. [96:10] Matt: The AI Pod would have to be just AI. Just AI. That's true. Right. All right, gentlemen, see you next week, hopefully with LLM stories. All right. [96:21] Justin: Bye, everybody. [96:22] Matt: Bye, everyone. [96:23] Justin: Another week of cloud news wrapped up. [96:26] Jonathan: Boat will collect the news. [96:28] Justin: Justin will get the notes. Jonathan will write some code. Ryan will watch the perimeter and Matt will reluctantly watch Azure. Till next week for AI Amazon. On Google Cloud and Azure, and hey, maybe even Oracle, who knows? Check out thecloudpod.net for our newsletter. Join our Slack, message us on socials, or leave a review. I got new headphones, so that sounded amazing. [96:57] Matt: It does. Sounds good. But you want to talk about shoes, Ryan. [97:01] Justin: Well, you know, why not talk about shoes in a technology podcast? [97:05] Matt: Makes, you know, I mean, I, the technology of shoes is really impressive. There's a lot of rubber involved. There's a lot of, you know, 3D printing of shoes now they do at Nike. I mean, there's, there's definitely some cool technology angles, but I don't think this is that. [97:17] Justin: Oh, it's not? [97:19] Matt: Oh, I mean, I don't know. You tell me. [97:21] Justin: No, it's not. This is a shoe company named Allbirds announcing a $50 million deal to rebrand as New Bird AI. So shifting from making shoes to GPU compute infrastructure on-demand cloud service built to host your AI workloads, which, okay, sure, that's a, that's a, that's a turnaround. [97:45] Matt: I mean, like, did they just have a lot of warehouse space? They just know what to do with it. They're like, well, if we turn this into data center, we can, you know, do a lot better for ourselves. I mean, it's, it's a, it's a crazy pivot. I mean, they sold off the shoe part of the business for like pennies, like nothing. [97:58] Justin: Like they didn't try to get any money back on that. Like, I don't know if it was like going downhill for years and this is how they tried to pivot something. [98:05] Matt: I mean, that's my understanding is that Allbirds as a brand had kind of fallen outta favor and they hadn't innovated any new shoes. And so while there's a very loyal tech bro following of Allbirds in the Bay, it wasn't really sustainable as a growth business anymore. So they were gonna be in a situation where they would be, you know, optimizing margins, which is never a fun business to be in. And so, uh, yeah, pivot to AI, I guess that's the way it is. That's what it, it's the new hotness, right? [98:30] Justin: So it's Crazy. The thing that really annoyed me is that on this announcement, their stocks surged 580%, which I would be mad if I invested in a shoe company and they pivoted to some random other business. [98:46] Jonathan: But apparently that's not the, uh, you would actually be very happy, I feel like. [98:51] Matt: Well, just saying. [98:52] Justin: That's true. Yeah, I'd be angry, but I'd have money, which softens the blow a bit. [98:58] Jonathan: $2.50 to just shy of $17. I mean, it's now sitting You know, that was unsustainable, but it's now sitting at like $7, $6 and change, high $6 and change. So it's definitely staying higher than they were at with just saying, we're doing AI, we're doing GPUs. And I don't understand. [99:17] Matt: I mean, this company has never done well. It's kind of like IPO was November 1st, 2021 at $520 a share, which is crazy to me for a shoe company. Right. And basically it's it's never, never retained the heights of its IPO. You know, within a year, let's see, November '22, it was down to $56 a share. So they literally lost like 90% of their stock price. And then it's basically, their peak was February 20th, '23 where it was $54. Now it's after '24 and it's just basically been slowly, it's been under $10 a share now for almost a full year. So I mean, they had to do something if you want to keep the stock price going well. But yeah, I mean, what are they going to become? A Cloudflare? Are they going to be a neoscaler? Like, you know, what if they're going to buy all these GPUs from Nvidia and then they're going to put them somewhere and then they're going to do what with them? That's what I don't understand. [100:14] Justin: And with GPU scarcity, like, they're, they're going to be a small fish in this larger pond of competition. So they're got anything new, they're gonna— it's gonna be delayed. Like what, and what business like acumen did they have that they could capitalize to offer something new in this market? Like, I don't, I really don't understand. [100:33] Jonathan: There's something else going on, whether it's like a back, backroom deal or something like that. [100:39] Justin: Oh, I, I think the CEO had a nervous breakdown and was just like, AI. [100:44] Jonathan: And then, yeah, I told you you needed to use AI, so I'm gonna buy AI for you. [100:50] Matt: We're gonna sell the shoes and we're just gonna do AI. But yeah, basically they're saying they're gonna lease AI GPU capacity to customers who need dedicated AI access. So I mean, like, this is a complete business change of the model. I mean, I, I'm glad the stock came back down. It was like that original pop was so ridiculous. Yeah. And I, I would be shocked to see that this doesn't return to $2 or $3 a share unless they can really show momentum in the space. But like, like, you know, when are they going to get their GPUs? When are they going to get all these things? Because, you know, they're on the clock now. They got to prove that they can do something. [101:25] Justin: Yeah. I mean, I, yeah, I don't suspect we'll see this company around long-term. I just don't think they're going to be competitive in the space. They're not offering anything new to me, like in the, and you know, they're not, there's no narrative on why the shoe company is going to be good at this. [101:42] Matt: So. [101:43] Justin: I'm out. [101:45] Jonathan: So there, there's an article that, you know, not the one I think we referenced in here, but you know, I saw, I just found it again. It's like a lot of companies today that you know didn't start anywhere near what they do nowadays. Like Nokia was the Finnish telco company, started as a paper pulp factory, and Nintendo was originally playing cards. And Samsung was exporting dried fish and produce. [102:12] Matt: I mean, in theory, Nintendo's still playing cards. Have you seen Yu-Gi-Oh and Pokémon? [102:17] Jonathan: Yeah, I know. [102:18] Justin: But I've never seen someone pivot like completely. Like Amazon was a bookseller company, right? But they didn't give up the commercial business. [102:24] Matt: Yeah. I mean, like even in the, like this, they just dumped, you know, back in like the '20s and, you know, 1910s, like there was pivots there. But you know, again, like they were typically somewhat tangential, like, oh, well I thought I was gonna build cabinets and then I realized that the real money was in countertops. So I moved to countertop, you know, I was like, or like 3M, you know, I invented a bunch of chemicals. And so then I realized, I stumbled across Post-it notes and I made a bajillion dollars on Post-it notes. So I became a paper company. Like they're very, they're very clear transition points. Like if Allbirds became a 3D printing shoe company who, you know, sold designs for 3D printing shoes, I could say that makes sense. It's related to what you did before, but it's different and maybe it's making more money than the original business. [103:04] Jonathan: I could see that. [103:05] Matt: But this, this is like, you know, oranges and apples. [103:09] A: Mm-hmm. [103:11] Jonathan: Well, you're in the fruit family. [103:12] Matt: And like, what, I mean, unless I'm desperate for GPU capacity and I can't get it from any other place in the entire world, why would I go to Alberts? Yeah. What's the, what's the play for me is what's the value prop? Like, why am I going to, are you telling it to me for pennies on the dollar? 'Cause that's not gonna be good for your business either. Right. [103:26] Justin: It's so strange. [103:27] Jonathan: Well, you'll get the people using the product at least. Won't be, cost-effective for you, but will you— people would use it. [103:34] Justin: Yeah, not for long. [103:36] Jonathan: Not for long. [103:38] Matt: All right. Yeah, well, uh, we'll keep an eye on this one. I don't have a lot of hopes for them. I think this is just a slow, slow decline into nothingness, but maybe, maybe Allbirds AI will be, uh, amazing. I mean, do they have a new website? Can I sign up for interest? Yeah, do they have a blog that we can follow on the podcast? [103:55] Jonathan: What happened to, um, ai.com after Super Bolt. [103:58] Matt: Yeah, I have no idea what happened. [104:00] Justin: Good call. [104:01] Jonathan: Yeah, we need, um, we need— maybe I'll look at adding a feature to Bolt of like these random things that like, you know, kind of remind her like one year from now to do something like this. Like, where is ai.com? Where is allbirds.com? [104:14] Matt: I would tell you to— you should write it into, uh, Bolt, but I don't know if you have the time for that. [104:22] Jonathan: No, I'm working on that. It'll be a future project. I'll put it on my to-do list. [104:27] Matt: There you go. Uh, so it looks like it's still the same thing it was on the, on the Super Bowl. It's still just a landing page where you can sign up for a handle. My handle is still reserved 'cause I was trying to sign up for it again and it said, nope, it's still, still, it's already used up. [104:39] Jonathan: But, uh, yeah, so they dropped their database. Got it. [104:43] Matt: I mean, it's still, uh, it's still, it looks like X is still involved somehow, or I don't know, terms and conditions, like there's ai.com. Is this, I mean, yeah, like when are they gonna launch something? Do something with it. [104:54] Justin: Yeah. [104:55] Matt: Next Super Bowl they just spend another $10 billion. [104:57] Justin: Just a new one. Yeah. [104:58] Jonathan: Yeah. [104:59] Justin: It's just a laundry, you know, a money laundering scheme. Yeah. [105:01] Matt: I'm gonna post this on my Twitter, see if people anyway. ai.com. Where'd it go? [105:08] Jonathan: Well, this is fascinating for our viewers. [105:09] Matt: I know, it's great. Sorry. [105:10] Jonathan: We'll edit. [105:11] Justin: It's the end of the show. [105:12] Matt: We'll edit all this out, right? Yeah. That's, that's why this is after show. It's great. You're a great time. Yeah, well, I think on that note, we'll let Allbirds, uh, become an AI company. Yeah, see how that goes with them. And, uh, in the meantime, we'll keep following the cloud. So gentlemen, I'll see you, uh, next week. [105:27] Justin: Yeah, till next week. Bye now. Bye.