AI is moving faster than most people realize—and it’s continuing to reshape the workforce. Paul Roetzer and Mike Kaput dig into Microsoft’s 6,000 job cuts and what they signal about the future of AI-powered automation, they also explain the major copyright report that triggered a high-level firing and they break down new data from the 2025 State of Marketing AI Report.
The episode also covers OpenAI’s autonomous coding agent, TikTok’s new AI video tool, the rise of AI baby podcasters, whether or not chatbots are replacing search, and more in our rapid fire section.
Listen or watch below—and see below for show notes and the transcript.
Listen Now
Watch the Video
Timestamps
00:00:00 — Intro
00:06:49 —More Quiet AI Layoffs, Including at Microsoft
- Microsoft Is Cutting Thousands of Employees Across the Company - Bloomberg
- X Post from Arman
- Microsoft Slashing Thousands of Workers, Including Management Jobs - The Wall Street Journal
- Microsoft Layoffs Hit Coders Hardest With AI Costs on the Rise - Bloomberg
- Company Regrets Replacing All Those Pesky Human Workers With AI, Just Wants Its Humans Back - Futurism
- Klarna Slows AI-Driven Job Cuts With Call for Real People - Bloomberg
- Microsoft CEO says up to 30% of the company’s code was written by AI - TechCrunch
00:19:24 — Bombshell Copyright Decision and Drama
- Copyright and Artificial Intelligence Part 3: Generative AI Training - Copyright.gov
- X Post from Ed Newton-Rex
- Trump fires director of U.S. Copyright Office, sources say - CBS News
- Elon Musk’s apparent power play at the Copyright Office completely backfired - The Verge
- Copyright and Artificial Intelligence - Copyright.gov
00:30:01 — 2025 State of Marketing AI Report Findings
00:39:18 — OpenAI Releases Codex
- Introducing Codex - OpenAI
- AI AGENTS EMERGENCY DEBATE: These Jobs Won't Exist In 24 Months! We Must Prepare For What's Coming! - The Diary of A CEO
00:41:40 — Altman Wants to Build “Core AI Subscription” for Your Life
00:56:20 — Altman, Musk, and Grok Drama
- X Post from Sam Altman on Feud with Musk
- X Post from Kaitlan Collins
- Grok’s white genocide fixation caused by ‘unauthorized modification’ - The Verge
- X Post from xAI
- X Post from Sam Altman on Grok Incident
01:01:22 — Are Chatbots Replacing Search?
- Are AI Chatbots Replacing Search Engines? [New Research] - Andy Crestodina
- X Post from Cloudflare CEO
- X Post from Similarweb
01:05:36 — AI in Education Updates
- BGSU to be first in the nation to offer AI + X bachelor’s degree - BGSU
- The effect of ChatGPT on students’ learning performance, learning perception, and higher-order thinking: insights from a meta-analysis - Nature
01:11:15 — The Cost of AI
01:14:29 — AI Product and Funding Updates
- Perplexity:
- TikTok:
- Meta:
- Higgsfield:
- Hedra:
- Google Deepmind:
- OpenAI to Z Challenge:
- HeyGen:
01:20:04 — Listener Question
- Do you see agencies in the future being mostly agents with a client facing team that runs the agents?
Summary:
More Quiet Layoffs
Microsoft is cutting about 6,000 jobs—roughly 3% of its workforce—in a move the company says is aimed at “removing layers of management,” in another example of the quiet AI layoffs we’ve been tracking.
The cuts are hitting software engineers the hardest. In Washington state, nearly half of all layoffs were engineers. Product managers and technical program leads were also disproportionately affected. Meanwhile, frontline sales and customer support roles were largely untouched.
Microsoft said the cuts were motivated by the need to shed layers of management and build a more efficient business. And AI played a big role in that.
Microsoft is spending big on AI—an estimated $80 billion this fiscal year on data centers alone. And the tools it’s building are already reshaping internal workflows. CEO Satya Nadella recently said AI now writes up to 30% of the code on some projects.
Bombshell Copyright Decision and Drama
President Donald Trump has fired the head of the US Copyright Office, just days after the office released a report questioning how AI companies use copyrighted material to train models. The report, notably, pushed back on the idea that scraping vast amounts of copyrighted content qualifies as fair use.
The firing of Shira Perlmutter, who had held the role since 2020, came right after Trump also ousted the Librarian of Congress. Some industry sources speculate this was a tech-industry coup, possibly linked to Elon Musk’s vocal disdain for intellectual property.
After all, the new report was seen as a huge win for copyright advocates. While it isn’t legally binding, it does provide much-needed guidance from the US Copyright Office on how it views AI and copyright.
At the core of the report—and the overall copyright debate—is fair use doctrine, which major AI labs use as justification for using copyrighted material to train their models and produce outputs.
Basically, if your use of the work falls under “fair use” rules, you are not infringing on the work’s copyright. Over 113 pages, the report appears to challenge the idea that all AI training is fair use, like labs would like to argue.
According to AI copyright expert Ed Newton-Rex, the report concludes that some uses of copyrighted work by AI aren’t “transformative” enough, or changed enough in their usage from the original material. It also concludes that AI use of these materials in their training may seriously dilute the markets for similar works, which is a key consideration when judging whether or not something constitutes fair use.
2025 State of Marketing AI Report Findings
This past week, at Marketing AI Institute and SmarterX, we released our fifth-annual State of Marketing AI Report.
In this year’s report, we had almost 1,900 marketing and business leaders answer 23 different questions about AI usage and adoption. And we found tons of useful insights as a result about where the marketing industry is really at when it comes to AI.
You can go download the full report at www.stateofmarketingai.com.
This episode is also brought to you by the AI for B2B Marketers Summit. Join us on Thursday, June 5th at 12 PM ET, and learn real-world strategies on how to use AI to grow better, create smarter content, build stronger customer relationships, and much more.
Thanks to our sponsors, there’s even a free ticket option. See the full lineup and register now at www.b2bsummit.ai.
This week’s episode is also brought to you by MAICON, our 6th annual Marketing AI Conference, happening in Cleveland, Oct. 14-16. The code POD100 saves $100 on all pass types.
For more information on MAICON and to register for this year’s conference, visit www.MAICON.ai.
Read the Transcription
Disclaimer: This transcription was written by AI, thanks to Descript, and has not been edited for content.
[00:00:00] Paul Roetzer: I just feel like people generally have no concept of how fast stuff's gonna change because this change already lives in their labs, and that's usually we're getting things that's 12 plus months old of what they've already been able to unlock internally.
[00:00:15] So there's always smarter versions.
[00:00:17] So Welcome to the Artificial Intelligence Show, the podcast that helps your business grow smarter by making AI approachable and actionable. My name is Paul Roetzer. I'm the founder and CEO of Smarter X and Marketing AI Institute, and I'm your host. Each week I'm joined by my co-host. And Marketing AI Institute Chief Content Officer Mike Kaput.
[00:00:39] As we break down all the AI news that matters and give you insights and perspectives that you can use to advance your company and your career, join us as we accelerate AI literacy for all. Welcome to episode 1 48 of the Artificial Intelligence Show. I'm your host, Paul Roetzer, [00:01:00] along with my co-host Mike Kaput,
[00:01:02] who is fresh off his stint at Saster this week. So recording, uh, Sunday morning. This is an unusual recording day, so if you're new to the show, we normally record on Mondays, drops on Tuesdays.
[00:01:15] But we are recording Sunday morning, May 18th, because, I have to leave for London tonight, so Mike and I are like passing ships in the night here. so Mike was at Saster. Gimme the quick rundown. Mike, you and I haven't really caught up. How was Saster? What, what was that like? I've never gone to the conference.
[00:01:33] Mike Kaput: Incredible event. It was my first time being there. even in the short time I was there kind of helping out moderate a panel with Google Cloud. Just made a ton of great connections.
[00:01:43] I mean, show Floor
[00:01:44] is amazing. Tons of thousands of people there. definitely as far as I can tell, all in on ai, AI is everywhere.
[00:01:52] so yeah, I got to moderate a panel on their CMO Summit Track, which was kind of a private event within the event.
[00:01:59] Got to [00:02:00] talk to leaders at Google Cloud and Demandbase about the future of AI for CMOs and marketing leaders. and some awesome panel, awesome conversation. Had some fans of the podcast come up afterwards and take photos and stuff. It was fun.
[00:02:16] Paul Roetzer: I I'll say, like I know you get out and do some events as well. it has become. I haven't quite adjusted yet to the fact that anytime I go to events there, there are like a dozen or so people that come up and like, just wanna talk about the podcast. So shout out to all of our listeners.
[00:02:34] I mean, it's, it's still wild to me that that this, podcast has grown so, so big. And, we have so many people who are regular listeners to the show. I like, I was at one, I think it was a week or two ago, it was on a Tuesday, so the episode had just come out like earlier. And someone came up to me and started like, talking to me about something from that episode.
[00:02:56] I was like, how do they know this? Like, this isn't even out yet. And then I realized, like, they must [00:03:00] have listened to it in their hotel room before they like came down for
[00:03:02] Mike Kaput: Oh my gosh.
[00:03:03] Paul Roetzer: Just wild. yeah, that, that's awesome. glad you got to do that. Google Cloud has been an amazing partner of ours.
[00:03:11] Uh, so it's, it's always fun to get to do events with them. yeah. And then I'm heading to London. I have a series of talks this week on. I it's Tuesday and Wednesday maybe. So I'm, I'll, I'll be back in town later this week. But, uh, yeah, this is my, my last like, big trip before the summer and then I only, I only have like a couple of engagements for, for the summer, so I get to be home with the kids and so that should be nice.
[00:03:35] Mike Kaput: That's great.
[00:03:37] Paul Roetzer: all right, so, before I jump on my plane To London, let's get into episode 1 48. It is gonna be a wild week ahead. We're going to, we'll give a little preview of some of the stuff that's likely coming, but we've got Google io. where we know some stuff's gonna be dropping. that's happening May 20 to 21, which means that OpenAI [00:04:00] is probably gonna drop another model.
[00:04:02] We may get the Anthropic model this week. Uh, it's, it's gonna, it's gonna be a wild end to May. I think. I think we're gonna go into June with a bang when it comes to model releases and news. but we've got some bigger items to talk about going into this week. So this week's episode is brought to us by the AI for B2B Marketers Summit.
[00:04:21] This is one of our marquee virtual summits through Marketing AI Institute. This year's event is presented by Intercept and it is packed with incredible sessions from top B2B marketing experts. It's all happening virtually on Thursday, June 5th, starting at noon Eastern time. You'll learn real world strategies to use AI to grow better, create smarter content, build stronger customer relationships, and much more.
[00:04:44] So thanks to our sponsors, there is a free ticket option. So, as with a number of our virtual events, you, you have the choice to do the free ticket option. There is a private registration option. I think that one's $99 if I'm not mistaken. And that means your contact [00:05:00] information is not shared with the sponsors.
[00:05:01] And then there is an on-demand option as well. So if you happen to be at a different time zone or just can't make it, there is an on-demand option there. So that is B2Bsummit.ai. Again, that is B2BSummit.AI to learn more. And. This episode is also brought to us by Macon 2025. So this is our flagship in-person event through Marketing Institute, the Marketing AI Conference.
[00:05:27] We started this in 2019. So if you are ready to drive personal and business AI transformation, join us in Cleveland, October 14th to the 16th on the shores of Lake Erie. Right across from the Rock and Roll Hall of Fame is where the convention center is that we'll be at. Um, with dozens of speakers already announced dozens of breakout and main state sessions you can check out.
[00:05:50] And four hands-on workshops, including one. Mike is teaching and one I am leading. this is our sixth year for the Marketing Institute to bring forward thinking marketers together [00:06:00] to share, learn, and connect. We are, we had about 1100 last year, so we're expecting more than that this year. Prices go up May 31st, so grab your ticket now at MAICON.ai
[00:06:12] That is MAICON.ai. Okay, Let. Let's jump in again if you're new. I know every week we have new listeners. So what we do with this weekly is three main topics is the normal format. So we kind of pick the three main topics, you know, usually we go 5, 7, 10 minutes on each of those topics. And then we go through rapid fire items and there's usually seven to 10 rapid fire items.
[00:06:34] And we do it all in about, I don't know, I always think it's gonna be an hour. It never is an hour. always like an hour 15. but you can go up to two x on your podcast listening, network of choice and, and get done in under an hour. So, alright Mike, lets go.
[00:06:49] More Quiet AI Layoffs, Including at Microsoft
[00:06:49] Mike Kaput: All right, so first up, Microsoft is cutting about 6,000 jobs, roughly 3% of its workforce, in a move that the company says is aimed at [00:07:00] removing layers of management.
[00:07:01] And we're kind of seeing this as another example of the quiet AI layoffs that we've been tracking. Now, it sounds like from some of the reporting, the cuts are hitting software
[00:07:13] engineers the hardest. In Washington state alone, nearly half of all the layoffs that Microsoft did, there were engineers, product managers, and technical program leads were also disproportionately affected.
[00:07:26] It seems like frontline sales and customer support roles were. Somewhat untouched. And Microsoft is kind of leaning into this messaging about saying we need to shed layers of management and build a more efficient business. And AI is playing a big role in that in a few ways. Microsoft is spending big on AI and estimated $80 billion this fiscal year on data centers alone.
[00:07:49] And the tools it's building seem to already be reshaping internal workflows and expectations. For instance, CEO, Satya Nadella recently said AI now writes up to 30% of [00:08:00] the code on some projects. Now some are looking at this a little more starkly. So we found a post on X from someone we determined is an employee at Microsoft.
[00:08:10] a software engineer at Microsoft AI named Armand Conker posted about this development saying, quote, my company just laid off thousands of tech workers to invest in ai. This is end game.
[00:08:25] So Paul, starting back on episode 1 45. We began really talking in earnest about this phenomenon. You had indicated at the time, it was your belief that quiet AI layoffs had been happening for the last six to 12 months, that they're now accelerating.
[00:08:41] Sometimes they're under the guise of other cost cutting. Certainly seems like Microsoft is an example of more of this. I mean, AI is a key factor in the efficiency plays they're making, yet the stated reason for the cuts is about shedding layers of management. how are you? [00:09:00] Looking at this and reading between the lines here.
[00:09:02] Paul Roetzer: Yeah, I mean, I definitely think that that's what's happening now. We have to accept, like big tech companies lay people off all the time. Like there's always the cyclical nature of these layoffs. But given the environment of where we are and the advancements of the technology and the comments from executives in other, venues, other, uh, you know, news outlets, it's not hard to connect the dots of, you know, kind of what's going on and what they think is coming.
[00:09:29] just to like, call it a couple of elements from the Bloomberg article, Mike, that you were, uh, citing. So we continue, this is a quote from their spokesperson. We continue to implement organizational changes necessary to to best position the company for success in a dynamic marketplace in Bloomberg. they go on to say Microsoft, which employed 228,000 people in June, 2024.
[00:09:51] Deploys periodic layoffs, often to re reorient its headcount towards priority areas. So for example, they laid off 10,000 people in [00:10:00] January, 2023. does talk about the company being under pressure in recent years to keep a lid on costs amid the massive spending for the data centers like you had highlighted, the 80 billion they have committed this year in CapEx.
[00:10:12] But then I thought was, what was telling was, there was a excerpt from the article where they said last year, CEO Satya Nadela said AI was helping the company save on labor costs. The theme came up again on Tuesday during a JP Morgan conference when Microsoft Finance Executive Bill Duff said the company is quote, saving hundreds of millions of dollars a year.
[00:10:36] By
[00:10:36] using AI for customer support and reducing the need for human interaction. said Microsoft is deploying AI across multiple divisions to help personnel analyze deals for compliance issues, write marketing materials and other tasks. The company routinely reorients its workforce. That is a very popular way of explaining what happens away from legacy products and toward growth initiatives.
[00:10:58] So again, it is [00:11:00] common for sure, but the thing we've talked about previously is this idea that that software engineers who are writing code are basically the canaries in the coal mine right now for AI job disruption, that's the one that all the executives are talking about, how much code is being written by the ai.
[00:11:19] And as they look forward, they expect that to become dramatically more so. drop a link in the show notes to a TechCrunch article in which Mike Mark Zuckerberg was actually interviewing Satya Satya Nadella last week. might've been the week before on, yeah, this was at the end of April, at the Lama Con Conference, which we talked about on, the, on previous episode.
[00:11:43] And so Satya said in that one, which you had just cited, Mike, the 20 to 30% of code inside the company's reposit repositories was written by software meaning ai, Microsoft CTO. Kevin Scott previously said that he expects 95% of all code to be AI generated [00:12:00] by two. 2030. So it's not like we're drawing some difficult to find conclusion here.
[00:12:06] If the CTO of the company is saying that within five years we expect 95% of all code to be written by the ai, then what do you need a bunch of engineers for? this is just a prelude to, and it's basically looking on saying, okay, like as it gets better and better, more accurate, we're just gonna need fewer software engineers.
[00:12:24] Sundar Phai, CEO of of Google Alphabet said AI was generating more than 30% of their company's code. So now it's not clear exactly like the impacts it can have, but Zuckerberg's on record saying they're gonna have a junior AI engineer, software engineer by the end of this year. so, you know, it's just, that's why I say the canary in the coal mine is like, you look at this and it's the one they're all building for because it has compounding value to automate with ai.
[00:12:50] the writing of code and the creation of products and the improvement of research and all these things. But if you then take that out and start looking at other industries and say, well. [00:13:00] There's no reason they couldn't do this in a bunch of other professions. They're just choosing to first do it with engineering.
[00:13:07] Mm-hmm. So remember this Mike, like back in 2014 when I wrote the Marketing Performance Blueprint, which was the first time I publicly shared any of my research on ai, and Mike worked with me on the research project to write the section of the book that was related to ai. So there's a section in my 2014 book called Algorithms and Artificial Intelligence.
[00:13:28] Now, in that section I cited a study from the University of Oxford. So thi this was a study done in 2013 called The Future of Employment, how Susceptible Our Jobs to Computerization. Now the That study was before Gen ai, so this was just looking at machine learning. Deep learning was emerging like. We had had the breakthrough in image recognition in 2011, 2012.
[00:13:55] So like they were starting to think we could see deep learning become a thing, but it [00:14:00] wasn't a thing by any means yet. And at that time, that study found 47% of total US employment was at risk of computerization, which they called job automation by means of computer controlled equipment. And at the time, the high risk jobs were telemarketers, insurance, sales agents, real estate agents, market research analysts, marketing specialists, PR specialists, graphic designers, again before Gen ai.
[00:14:24] so if you now look and say, okay, if we look at what's happening with software engineering and we know the labs are pushing on automating coding, what happens when. the startup world emerges and says, well, let's go do this for accounting and bookkeeping. Let's go do this for legal assistance.
[00:14:39] Let's go do this for HR and recruiting. Let's go do it for di diagnostic medical roles and, and all the content creation. It's all it takes now is just like someone to say, okay, like, let's push on this one. Let's get those to 30 40% with a view to 75 plus percent just is happening with engineering.
[00:14:56] And so that's, I don't know, I, that's [00:15:00] again, why I think that's just so important. We have these conversations and we start being realistic. I get the, that people are like, I. Doubt that this is gonna happen. Like, I understand that it's hard to like comprehend, but we're watching it happen. Like it's, it's literally happening right now,
[00:15:17] Mike Kaput: right?
[00:15:18] Paul Roetzer: people still wanna pretend like this isn't gonna come and affect jobs in the economy. And I just don't understand that perspective. yeah, so don't, I don't, I don't know. I, it, again, it's a topic we'll probably keep talking about because I think we, we just need more conversation in different industries.
[00:15:39] So like, if you're in the legal industry, look at this and then try and like start thinking ahead. What does it mean to you? it's just critical that we're doing that,
[00:15:48] Mike Kaput: you know, kind of as a counterweight to all that. We've kind of also talked about the fact that. In the rush to potentially automate or augment workforces that there is a [00:16:00] danger that you're using AI for things that it can't adequately do, yet you're getting rid of people too fast.
[00:16:05] Kinda that difference between AI first versus AI forward. and we kind of did see an example of that this week because, you know, as early as late last year, we've been talking about the company on and off Klarna, it's a FinTech company, and we've been talking about them because their CEO has been like all over the headlines, bragging about how the company is freezing, hiring and automating away hundreds of customer service positions thanks to ai.
[00:16:32] While it came out this past week, he's now reversing course going on a hiring spree because AI was not producing the quality needed for the roles that he needed it for. He actually said to Bloomberg as cost unfortunately seems to have been a two predominant evaluation factor when organizing this, meaning his big push to have AI automate away humans.
[00:16:54] What you end up having is lower quality. Really investing in the quality of the human support is the [00:17:00] way of the future for us. So Paul, maybe you could just give me a quick bit of context there. Like, it seems like this isn't, as black and white perhaps as some of the, AI futurists want us to believe.
[00:17:12] Paul Roetzer: Yeah. So, you know, we talk about all this automation and job disruption, but what we also have to do is have business leaders within these enterprises that are being realistic about what the capabilities of the AI truly is. Because we do see too much like rapid, uh, like. Disruption happening because assumptions are being made that it can replace people.
[00:17:32] And one, like, as we've said many times, I'm not an advocate of that approach at all. Like, I'm, I'm fine with like, maybe we don't need to hire as many people moving forward, but we shouldn't be thinking first and foremost, like, let's just get rid of all the people. but I think that a lot of companies are just assuming they hear all these AI agent talk and everything, and they just assume it can actually do the job of people.
[00:17:54] So in this case, yeah, as a high profile person, they got out over their skis a little bit, [00:18:00] you know, in terms of making all these claims. And now they're, you know. Kind of circling back around, realizing people aren't all that bad. I did love the futurism. We'll put a link in the show notes, the headline I don't know if AI wrote it or if there's actually like a human writing these headlines, but the headline was, company Regrets replacing all those pesky human workers with AI just wants, its humans back.
[00:18:21] It's like that's so perfect. And I can imagine seeing a lot of that. I think there could be these situations where almost like you see with Doge right now in the US government Where it's like, oh, let's just get rid of all these pesky people. Oh wait, they were like fundamentally important to what we were doing and critical to our operations.
[00:18:37] Let's bring all those pesky humans back. And so I I think you could probably see something like that occur and maybe tech is, again, tech's always gonna be kind of leading on this good or bad. where maybe six months, 12 months now it's like, oh, actually we do kind of need a bunch of people and customer success.
[00:18:53] 'cause humans actually like. Talking to other humans sometimes. So I don't know. I think it's gonna ebb and flow. I don't think there's this [00:19:00] like cutoff point. We just fall off this cliff and all the jobs go away. I think it's just gonna ebb and flow by industry and some people are gonna move too fast and a bunch are gonna move too slow.
[00:19:08] And that's probably actually the greater risk, honestly, right now, is that you just sit around for three years and wait and see what happens with all these other companies that are figuring this out.
[00:19:16] Mike Kaput: Right.
[00:19:17] Paul Roetzer: trial and error is gonna lead to the breakthroughs and the innovations that are gonna obsolete your company if you do nothing.
[00:19:24] Bombshell Copyright Decision and Drama
[00:19:24] Mike Kaput: Our next big topic this week is that President Donald trump has fired the head of the US Copyright Office just days after the office released a report questioning how AI companies use copyrighted material to train models.
[00:19:39] Now, this report notably pushed back on the idea that scraping vast amounts of copyrighted content always qualifies as fairy use, which is what big AI labs tend to argue in this debate.
[00:19:53] Now, the firing of shear up. Pearl Mutter who had held this role since 2020, came right after Trump. Also, [00:20:00] as did the Librarian of Congress. Some industry sources have speculated that this was a bit of a tech industry coup, possibly linked to Elon Musk's vocal disdain for intellectual property. that's because this new report is seen as a pretty significant, win or, you know, planting a flag in the ground for copyright advocates because it's not legally binding, but it provides a bunch of guidance from the US Copyright Office on how it's viewing AI companies using copyrighted materials.
[00:20:32] And at the core of this, like I mentioned, and at the core of this overall debate is fair use doctrine, which major AI labs use to justify the fact they've scraped copyrighted material to train models and produce outputs. Basically the thinking is if your use of. The copyrighted work falls under these fair use rules that, that are enshrined in US law,
[00:20:54] You are not infringing on the works copyright. So this report is [00:21:00] 113 pages and over it appears to challenge the idea. That all this AI training is fair use. So according to ai copyright expert Ed Newton Rex, who we've talked about before, it basically concludes that some uses of copyrighted work by AI aren't what they would call transformative enough or changed enough in their usage from the original material.
[00:21:22] It also concludes that AI use of these materials in their training may seriously dilute the market for similar works. That's
[00:21:30] a big consideration when lawmakers judge whether or not something constitutes fair use. So, Newton Rex actually said, quote, many of us have been saying this for a long time. Shouted down by a chorus of AI boosters. It's great
[00:21:45] to see the US copyright office, the government body responsible for providing guidance on copyright law. Agreeing. So Paul, let's maybe first talk about the report's conclusions, then get a little bit into the political fallout here. Again, [00:22:00] this is not legally binding, but it does serve as.
[00:22:03] Pretty strong guidance on what they're thinking related to copyright concerns related to ai. So how big of a deal are the conclusions in this report?
[00:22:12] Paul Roetzer: Yeah, I don't know what kind of long-term impact they're gonna have, you know, in terms of is it gonna change anything? I mean, obviously big enough that within 24 hours they fired him. Yeah. like the people involved. so, you know, again, just to rewind here, the, all the AI labs building these models. Absolutely. Used copyrighted materials. They absolutely knew it was a legal gray area and most likely illegal at the time based on current US law. But they knew the other labs were doing it, and if they didn't train on the copyrighted materials, their models would not be on par with their competitors.
[00:22:55] And so they basically took a financial and legal gamble [00:23:00] that whatever the d the of this is, even if we end up losing in courts one, it's gonna take five to 10 years to all play out and the game's over by then if we wait. is, know, if we end up having to pay billions of dollars in fines, we're pursuing a market worth trillions of dollars.
[00:23:20] So I. We gotta go. And so they all kind of had this window to figure this out. You know, and this is part of what happened back in 2022 when chat GBT came out and did something Google what hadn't been willing to do to that point, was under too much regulatory scrutiny. They, they couldn't have basically done what Chad, GBT and OpenAI did.
[00:23:41] But Once that was, you know, that happened in the market, it triggered this massive competition between Google and Meta and all the other players to just go do it too. Now again, they had all probably already been training models on these things, but they hadn't released a high profile consumer product with these models [00:24:00] that had been trained on this stuff.
[00:24:01] So I think, I'll, I'll call it a few excerpts that I assume is what caused for the firing. So, this is straight from the document. We'll put the PDF in the show notes, throw it in notebook, lm, and, you know, explore it. It's like 113 pages, if I'm not mistaken. So there's a lot to it, but I'll try and call out.
[00:24:21] The things I think are, are the triggering points. So, in the docket says, in providing this analysis, the office rejects two common arguments about the transformative nature of AI training. noted above, some argue that the use of copyrighted works to train a models is inherently transformative because it is not for expressive purposes.
[00:24:39] We view this argument as mistaken. Again, they didn't mess around with their language. They're very direct. language models are trained on examples that are hundreds of thousands of tokens and length, absorbing not just the meaning and parts of speech of words, but how they are selected and arranged at the sentence, paragraph and document level.
[00:24:58] The essence of linguistics [00:25:00] expression. Uh, image models are trained on, oh wait, then we go into image models. Image models are trained on curated datasets of aesthetic images because those images lead to aesthetic outputs where the resulting model is used to generate expressive content or potentially reproduce copyrighted expression.
[00:25:14] The training use cannot be fairly characterized as non expressive. It goes on to say. agree that AI training is inherently transformative because it is like human learning to begin with. 'cause that's when you hear all the time, well this is how humans learn. We go online, we absorb things, and then we like create things.
[00:25:30] So it's just like imitating how humans learn. So they're saying that is not true. they say to begin with, the analogy rests on a faulty premise as fair use does not excuse all human acts done for the purpose of learning. A student could not rely on fair use to copy all the books at the library to facilitate personal education.
[00:25:47] Rather, they would have to purchase or borrow a copy that was lawfully acquired typically through a sale or license. Copyright law should not afford greater latitude for copying simply because it is done by a computer. [00:26:00] Moreover, AI learning is different from human learning in ways that are material to the copyright analysis.
[00:26:05] Humans retain only imperfect impressions of the works they have experienced filtered through their own unique personalities, histories, memories, and worldviews. Generative AI training involves the creation of perfect copies with the ability to analyze works near nearly instantaneously. The result is a model that can create at superhuman speed and scale.
[00:26:26] In other words, of Professor Robert Broadness. Generative model training transcends the human limitations that underlie the structure of the exclusive rights. And then two quick excerpts from the conclusion. Making commercial use of vast troves of copyrighted works to produce expressive content that competes with them in existing markets, especially where there is a, com. this is accomplished through illegal access, goes beyond established fair use boundaries.
[00:26:53] In our view, American leadership in the AI space would best be furthered by supporting both of these [00:27:00] world-class industries that contribute so much to our economic and cultural advancement. Effective licensing options can en ensure that innovation continues to advance without undermining intellectual property rights.
[00:27:11] These groundbreaking technologies should benefit both the innovators who design them and the creators whose content fuels them, as well as the general public. Now, imagine when this goes to the Supreme Court, that that is basically the argument being made, on the side of fair use or, or on the side of the creators.
[00:27:30] And so that's why the administration's pissed. They basically know there's a bunch of legal battles in involving Mark Zuckerberg, Sam Altman, Elon Musk, all the people who are giving the money to the administration, who are fighting for, you know, American innovation. They don't want this to be the outcome.
[00:27:46] If it does, it impacts their ability to drive innovation in America. And the US Copyright Office just laid out the argument that could go against all of that. why they get fired. Now, what happens? I have no [00:28:00] idea what my best guess is. Someone gets installed into this position that rewrites this paper so that it does
[00:28:06] Paul Roetzer: When it goes to the courts, this is not the position of the US government. 'cause as of right now, this is now the official position of the US government that can be, you know, used in legal proceedings.
[00:28:14] Mike Kaput: So at first I was excited we had more clarity on this, but it seems like that might not be the case.
[00:28:21] Paul Roetzer: Yeah, I don't know. I mean, we'll, we'll see what hap I did see an article, I'll, I'll try and find it with her on the show notes. That was like the firings backfired because.
[00:28:30] Paul Roetzer: somebody got installed into the position who's actually like anti-tech and it might actually be worse for the tech companies.
[00:28:36] Mike Kaput: Yeah. There was some reporting from the Verge that actually what happened, as of right now is that people who are installed in these positions are actually still part of the Trump Administration's Coalition, but less on the tech accelerationist side and more on the kind of anti-big tech side.
[00:28:54] So they're a bit hard liner on those.
[00:28:56] We'll see how that actually plays out, but not the [00:29:00] ideal people you would expect to be in these positions if you're. Against copyright law.
[00:29:04] Paul Roetzer: Yeah. And we'll try and keep an eye. I'm very curious to see how fast before we see these exact words quoted in an ongoing case. My guess is the briefs have already been written and it's like been submitted as, you know, evidence or supporting materials in existing court cases.
[00:29:22] Like I can't imagine this wasn't within a day already being injected into actual briefings.
[00:29:27] Mike Kaput: Yeah. We had a, a obviously won't name names, but we had a, a partner of ours who was, sent me an email the other day after we had done some work together and was like, Hey, could you just like, what's the latest on AI copyright?
[00:29:39] We were thinking of these questions and wondering about some nuances here, and I just like, wanted to cry.
[00:29:44] I was like, I'm so sorry. I don't have a good answer for you here. This is a, a highly paid lawyer's question because I can't answer this.
[00:29:52] Paul Roetzer: We are not giving legal advice. We say this all the time. We are not lawyers.
[00:29:55] There's no legal advice and we're not providing investing advice, We are providing commentary on what is [00:30:00] happening in the world.
[00:30:01] 2025 State of Marketing AI Report Findings
[00:30:01] Mike Kaput: All right, so our third topic this week is about something we actually released, which is, we released our fifth annual state of marketing AI report from Marketing AI Institute and Smarter X.
[00:30:13] And in this year's report, we had almost 1900 marketing and business leaders answer 23 different questions about AI usage and adoption. And we actually found a ton of really useful insights as a result about where the marketing industry is actually at when it comes to ai. So. Paul in this segment I was gonna share just a few of the key findings from the report.
[00:30:36] Kind of get your thoughts, where you have them on certain stats here.
[00:30:41] I would highly encourage everyone go download the full report at stateofmarketingai.com. There's a ton more in here than what we'll go through
[00:30:50] here, so I'm just gonna dive in.
[00:30:52] Paul Roetzer: And I think the webinar's on demand too, like they could probably, so if you, go to marketing
[00:30:56] Mike Kaput: ai institute.com under resources and click webinars, [00:31:00] you'll be able to quickly see where we have the webinar on demand diving further into this. Definitely. Also, worth a lot. We will drop a direct link for that in the show notes too.
[00:31:08] Absolutely. So we did find, you know, to kick things off that marketers are racing ahead with ai. We ask about various aspects of the marketing team's AI usage every year. And this year actually, 60%
[00:31:24] Together, either said they were piloting or scaling ai.
[00:31:28] you could choose from three phases, understanding, piloting or scaling. And 60% are now either prioritizing
[00:31:35] and running AI pilot projects or say they are. Pursuing wide scale adoption of ai. And what's really cool is because we've started to benchmark this over several years, we can see that that's actually an 18 percentage point jump since
[00:31:49] 2023, where only 42% were in either phase, which is super interesting to see that kind of growth and adoption. Uh, when [00:32:00] marketers, when we ask about what are you trying to achieve with AI marketers say more, they're more likely than ever to say that they're trying to actually save time. So this year, 82% said they're, reducing the time they spend on repetitive data-driven tasks was the primary outcome they're trying to achieve with ai.
[00:32:19] Not only was this the top answer, it was the highest percentage of responses for this question that we've recorded so far.
[00:32:27] There was some cool data on tools,
[00:32:29] respondents overwhelmingly said ChatGPT was their favorite tool. When asked, it was like 57% of people. We also asked if their organization provides licenses to either chatGPT Team, or Enterprise, Microsoft Co-Pilot, or Google Gemini, or none of the above.
[00:32:46] ChatGPT was also a leader here with 61% saying that their company was most likely to provide them with a chatGPT license.
[00:32:55] Now, what's interesting is that actually steadily declines in popularity as firms [00:33:00] get bigger, so it's much higher on the low end, one to $10 million,
[00:33:03] For instance, in revenue add 70% of people saying this, it falls to 37%
[00:33:09] when you look at firms with a billion plus, and that gap is often filled by Microsoft copilot. Now a big finding. We cite this every single year. A lack of training is a huge barrier to AI adoption. So once again, marketers cited a lack of education and training as their top barrier. When we ask about this, that's been the case every single year.
[00:33:32] We've published the report this year, 62% said it was a top barrier. Not to mention, we also ask, do you have AI training at your company? 68% of respondents say no, and that consists of 44% saying they don't have it. 21% saying it's in development and 3% unsure if
[00:33:53] it exists. I mean, Paul, we talk about that stat all the time because it's really not changing fast [00:34:00] enough.
[00:34:00] Paul Roetzer: Yeah, and and just a, a reminder, you know, we've touched on this before, but like. This is almost 1900 people from our audience. So this is like marketing AI Institute subscribers. It is people who listen to the podcast. This is like likely the leading edge in terms of adoption. Yeah. So when you're hearing this data, it's not necessarily representative of the entire market because people who aren't doing this stuff likely aren't listening to the podcast or subscribing to Marketing Institute, coming to our events following Smarter X.
[00:34:36] So this is like in, in my assumption, this is probably the best of the best. This
[00:34:41] Mike Kaput: Right.
[00:34:42] Paul Roetzer: more indicative of what innovators are doing than. The laggards in the market.
[00:34:47] Mike Kaput: Yeah, that's so true.
[00:34:48] Paul Roetzer: the data is like skewed and has bias because it's a forward thinking audience already.
[00:34:54] Mike Kaput: Yeah.
[00:34:55] And if you check out the full report, we go in depth into the methodology, the [00:35:00] respondents, so you can see exactly kind of who is answering these questions. Couple more data points I want to call out here. So. Unfortunately, marketers seem more pessimistic than ever about AI's impact on marketing jobs. So this year, 53% said they believe more marketing jobs will be eliminated by AI than will be created.
[00:35:20] We started asking this question in 2023, and it's the highest number on record. It's risen by 13 percentage points in just two years. Now, it should be noted marketers are a bit more nuanced when we also asked this year. What do you think the impact will be on your job specifically? the highest proportion, about 33% say they're neutral in their concern about AI's impact on their job.
[00:35:45] They see both risks and opportunities. That was the highest percentage answer.
[00:35:49] Now, last but not least, we actually found through a series of data points that. Well, when you look at the data, at the end of the day, it's clear that things like AI roadmaps, AI [00:36:00] infrastructure is the unlock, because we ask about, do you have an AI roadmap or strategy for the next one to two years that kind of outlines your use cases and direction for ai, 75% lack one.
[00:36:14] But when we look at who does have one. The people who do have an AI roadmap are also twice as likely across several data points to have training available, to provide prompt training, to have an AI council and to have generative AI and AI ethics
[00:36:29] policy. So Paul, I was just wondering to kinda get your thoughts on that. It really does seem like the way forward is just more formal AI infrastructure within your organization.
[00:36:40] Paul Roetzer: Yeah, I mean, there's so many interesting data points in here, and the fact that we now have years of benchmark data is always so fascinating to me. To go back through this, I'll just kind of highlight and a couple of these you hit on Mike, but when we, you know, every month I do this, free scaling AI class where I teach these five steps for scaling ai and, and it's actually the [00:37:00] basis for the entire Scaling AI course,
[00:37:01] series.
[00:37:01] You can go
[00:37:03] Paul Roetzer: to scaling ai.com and, and see more about that. But we teach these five steps and so one of the ways I actually assess, AI transformation within an organization is to look at how far along they are with each of these steps. So the first is education and training. Do you have an internal AI academy?
[00:37:17] 32%. yes. So that's o only a third habit. Do you have an AI council? Only a third,
[00:37:24] Paul Roetzer: Mm-hmm. Do you have generative AI policies? 38% AI ethics policies, 41%. we don't ask directly, have you done AI impact assessments, but to show the need for those 53% now believe that more jobs will be eliminated because of AI in the next few years?
[00:37:43] Interestingly, when we asked the question, which I think was a newer question this year, how do you personally
[00:37:48] Mike Kaput: Yeah.
[00:37:48] Paul Roetzer: AI and the impact it's having? 70% said positive. So
[00:37:52] Mike Kaput: Yep.
[00:37:53] Paul Roetzer: this, like people are actually optimistic. And then the last one, Mike, that you had really zoomed in on is this idea of an AI roadmap.
[00:37:59] Only a quarter [00:38:00] of 'em, 25%. So to me, this screens opportunity, 'cause again, these are probably. More likely to be forward thinking. AI forward professionals and leaders.
[00:38:12] Mike Kaput: Yeah.
[00:38:13] Paul Roetzer: And this is still where we're at. Like a third are doing the things that they actually need to be doing. So for you, regardless of your job, regardless of what company or in what industry where you're at in your career, the opportunity is so massive to like move, like stop being overwhelmed by it and having all this anxiety about it.
[00:38:31] Just go do something because this data just keeps showing. We're not seeing leaps in any of this stuff. We're
[00:38:37] Mike Kaput: Yep.
[00:38:38] Paul Roetzer: you know, five percentage 0.8 percentage point, which, you know, maybe 10%, 20% improvement each year. are not seeing this, this, the switch flipped where it's like, oh, okay, now 80% of people have generated policies and
[00:38:50] Paul Roetzer: of roadmap, and I don't think we're anywhere close.
[00:38:52] Like we've been doing this for, what, five years did you say? Six
[00:38:54] Mike Kaput: Yeah. Five years. Yeah.
[00:38:56] Paul Roetzer: And like. It's incremental every year. [00:39:00] So I, to me, it just screams opportunity and I would, I would race ahead to integrate these five components that I just highlighted.
[00:39:07] Mike Kaput: Yeah, clearly doing something is better than what the vast majority of people appear to have been doing, which is nothing at the company level.
[00:39:15] Paul Roetzer: Yeah, wild.
[00:39:16] Mike Kaput: All right, let's dive into this week's rapid fire.
[00:39:18] OpenAI Releases Codex
[00:39:18] Mike Kaput: So first up,
[00:39:20] OpenAI has launched a major new tool for developers called Codex. This is a cloud-based software engineering agent that writes code, fixes, bugs and submits pull requests. All on its own. Unlike a chatbot assistant Codex appears to work more like a full-fledged coworker.
[00:39:38] It takes assignments, run tests, and returns. Finished work inside a secure cloud sandbox. It's powered by a model called Codex One. This is fine tuned with reinforcement learning to match real world developer habits and can handle multiple tasks at once. Work with actual repositories and follow instructions from pro project specific [00:40:00] agents.md files that guide how it navigates
[00:40:02] and tests code. So OpenAI says
[00:40:05] that companies like Cisco are already starting to test out this tool in real world environments. OpenAI engineers say they rely on it daily to reduce context switching and handle repetitive tasks. Right now, codex is available to ChatGPT, pro Team and enterprise users, and apparently plus and EDU access is coming
[00:40:26] soon. So Paul, I wanted to get your initial thoughts on Codex. Obviously you
[00:40:30] and I are not programmers, but this definitely comes at an interesting time as we're seeing more and more people start to wake up to the fact that. Truly agentic ai like this could have a pretty serious impact on knowledge work.
[00:40:45] Paul Roetzer: Yeah, I mean, a few quick takeaways here.
[00:40:48] People who build things already are gonna build smarter and faster, so people who can code are gonna code faster. Just as we talked about with the Microsoft story upfront. Probably not, unrelated [00:41:00] that these kind of advancements are happening. Microsoft has a unique relationship with OpenAI, likely has had access to this kind of tooling for months.
[00:41:08] and so, yeah, I mean, I think things that get built are gonna get built faster. So as a consumer of these products, as a, a business leader, as a practitioner, a knowledge worker. You, it's just gonna explode. Like the apps that it can be built, the things these apps can do. The smart, the tech's gonna get smarter.
[00:41:25] That's what this means to most of us. To, the non-developer audience. It just means developers are gonna build faster and smarter. this is not a product that you or I are gonna go play around
[00:41:36] Mike Kaput: Yeah.
[00:41:36] Paul Roetzer: It's, it's a product for developers.
[00:41:40] Altman Wants to Build “Core AI Subscription” for Your Life
[00:41:40] Mike Kaput: Next up in a recent chat with Sequi Capital, Sam Altman reflected
[00:41:45] on Open AI's journey so far as a company, and as part of this discussion highlighted a really interesting aspect of the company's larger ambition.
[00:41:55] He talked about how they want. To become the core AI [00:42:00] subscription. Those are their words for your life. Basically, the central interface you use to interact with
[00:42:06] the world powered by ai, hyper-personalized to you and your preferences. Now related to this, in this interview, he also notes that kind of crazily, there's already this, what he calls a generational divide when it comes to how different age groups use ai.
[00:42:22] He said he was seeing that older people tend to use ChatGPT as a Google replacement. People in their twenties and thirties are using it as a life advisor,
[00:42:33] and people in college seem to be using it like a full fledged operating system. Now Paul, taking together a thought the points about becoming a core AI subscription for your life and the fact that these different groups are already using this in different ways, in that context, I found all that really fascinating to kind of think about where we're headed.
[00:42:53] Like is this the direction we can expect AI development to go?
[00:42:58] Paul Roetzer: This is [00:43:00] definitely meant to be a rapid fire item. However, as I was kind of prepping this morning to, to go through the notes, I actually realize this connects very deeply to another podcast I'd listened to earlier this week. So I watched this, interview with Sam and it's like a 30 minute q and a.
[00:43:14] It's a q and a with the moderator, but then largely with the audience who's made up of Sequoia portfolio companies. I assume, like, you know, CEOs, founders of companies funded by Sequoia. So they get this inside access to Sam as part of that portfolio. so I would recommend watching it. Uh, he tells an interesting story of chat GBT that I actually hadn't heard told in the way he told it, like the origin of it and their own doubts around it and things like that.
[00:43:39] So it's kind of like an interesting first like seven minutes where he sort of goes into that. I wanna, I wanna call out a few quotes he had and, and provide them under categories. 'cause I think this is actually, It ends up being really important. Like, I think six months from now, we could probably come back to this episode.
[00:43:56] What I'm about to tell you I think will make a lot more sense. So I'll do my best [00:44:00] to explain why I think it's really important right now. So first he got asked, asked about enter enterprise adoption. and he did say it's gonna reshape everything, but that basically companies move really slow. He then he, he really nailed on this like 20, 30-year-old thing.
[00:44:13] He kept coming back to this theme, but he basically said, A bunch of enterprises are just gonna get obsoleted that the startups are just gonna move so fast and these companies aren't gonna move quick enough. And they're, they're basically all screwed, is kind of what he was saying. but he said, I, you know, this basically happens to people too.
[00:44:27] That's where he said, talk to a 20-year-old about how they use chat GBT and then go look at a 30 5-year-old. And it's just completely different how the uses is. And this actually ties back to an episode or two ago when we talked about them having to pull back the updated four oh model. Because they didn't realize how the younger generation was using it fully.
[00:44:46] And then they realized once they did. That like they hadn't really accounted for that usage as like therapists and, and friends and. Relationships and things like that. when he talked about that young people use, he, you know, the quote [00:45:00] that you kind of alluded to, he said gross oversimplification, but older people use chat GT as a Google replacement.
[00:45:05] People in their twenties and thirties use it as a life advisor, and people in college use it as an operating system. The operating system is the key phrase here.
[00:45:13] Paul Roetzer: back to that. I thought it was really interesting. They asked him about the importance of voice, and we had all this, you know, drama last year when the voice agent came out and it started, it sounded like Scarlett Johansen from her.
[00:45:25] And, and it like, made you think that they'd totally cracked the code on voice. Well, he said, quote, we just haven't made a good enough voice product yet, and that's fine. It took us a while to make a good enough text model. We'll crack that code eventually. And when we do, I think a lot of people are going to want to use voice interaction a lot more.
[00:45:41] When we first launched our current voice mode, the thing that was most interesting to me was it introduced a new stream on top of the touch interface. You could talk and be clicking around on your phone at the same time. I continue to think there's something amazing to do with the voice, uh, user interface.
[00:45:56] We just haven't cracked it yet. Before that we'll make voice really [00:46:00] great. And when we do not only will it be cool with existing devices, but I think a voice will enable a totally new class of devices if you can make it feel truly human level. So that's a prelude to them getting into the device business, which we know that they're looking at.
[00:46:13] he, they talked about future innovation and he was basically like, okay, so is it algorithms, is it data? Is it these massive data centers? Is it more Nvidia chips with compute? Like what is the thing? And he did say that he thinks there's still big algorithmic breakthroughs to be tad that'll unlock 10 x to 100 x improvements.
[00:46:30] He said, I don't think there's very many, but there might be one or two, like the invention of the transformer, like these truly, you know, breakthrough things that change the future of development. And this is the one I wanted to center on when he got asked about customization. So we've talked many times about how open AI's vision is truly personalized ChatGPT experiences to where, you know, Mike, you and I have totally different experiences with chat, right?
[00:46:54] Paul Roetzer: So he said, I, the interviewer said, I'm curious how you're thinking about customization. You mentioned [00:47:00] the federated sign in with OpenAI bringing your memories and your context. I'm just wondering if you see customization in these different post training applications, specific efforts as a kind of bandaid or a step toward making model, core models better.
[00:47:12] How are you thinking about that? And he said, in the Plato platonic ideal state is a very tiny reasoning model. So imagine like oh oh three with trillions of tokens of context. You put your whole life into the model. Never retrains. The weights never customize, meaning the underlying model just doesn't evolve.
[00:47:32] They just have a great reasoning model. But that thing can reason across your whole context and do it efficiently. Every conversation. Again, this is still saying, Every conversation you've ever had in your life, every book you've ever read, every email you've ever read, everything you've ever looked at is in there.
[00:47:49] Plus it's connected to all your data from other sources and your life just keeps appending to the context Your company does the same with all its data. We can't get there today, [00:48:00] but I think of everything else as a compromise off of that platonic ideal, and that's eventually how I hope we do customization.
[00:48:06] That is the operating system vision. So when he talks about it being the operating system for your life, for your business, that's it. So when when we get to the ability to have a trillion tokens of context back to that moment and he just laid out for this is like I almost, when I heard him say this, I was like, shit, this is like.
[00:48:24] When he wrote the Moore's Law for everything in March, 2021, and people just like
[00:48:28] Mike Kaput: and nobody paid attention.
[00:48:30] Paul Roetzer: This is it.
[00:48:30] This is the same thing, like every, when they introduced memory three weeks ago when they introduced the genic abilities, all of it is for this. So then that led me to this other interview that I didn't even realize it was gonna be connected, but on a flight last week or whenever it was.
[00:48:45] I was listening to Logan Kilpatrick, who we've talked about on the show before Google DeepMind. He was interviewing like a legendary I researcher, Nikolai Sav at, at Google DeepMind. And the whole premise of the podcast episode was about long context windows. [00:49:00] And so I'll, I'll call out a couple of things because it's extremely relevant here.
[00:49:03] So he talked about, like in the early days, how like 200,000 tokens of context. Now, again, if, if you're new to this stuff. A token is a piece of a word. And so a given word may have like two tokens in it. Three tokens in it. So we're talking about 200,000 tokens. You're, you're looking at it, it's like roughly 70, 80% of that is like the number of words that it can remember basically.
[00:49:25] So 200,000, you're talking about like 170,000 words. So imagine you're having a conversation with ChatGPT, it remembers things about you. It can roughly remember 170,000 words and talk coherently based on the knowledge that exists within that word. So our last book, Mike, was 50,000 words. So basically like three of our books.
[00:49:44] Is like the context window of these models originally. Right? So then ships one that does a million contexts. So, and I think 2.5 Pro is two 2 million
[00:49:55] Mike Kaput: I think at the moment, Yeah.
[00:49:56] Paul Roetzer: Yeah. So imagine like if you wanted to use this in your [00:50:00] business and you want to give it like a bunch of proposals you've written, you can give, you know, up to say 1.5 million words of and it's gonna like remember, so the way Notebook LM works, you throw a bunch of stuff in there.
[00:50:12] The reason it can cite things, it's because it has a really strong context window. So here's where it gets interesting though. When they did the research release in spring of 24, it might've been at the IO conference, I don't remember when they previewed this, but they said in the lab, we tested this on 10 million.
[00:50:29] Tokens of context and it works still. So they told us a year ago they could already do 10 million. Hmm. So this is the first time I heard this explained though, he said when we released 1.5 Pro, we actually ran some inference tests, meaning testing it, like actual chat usage at 10 million. And we got some quality numbers and single needle retrieval, like we're trying to find a really small thing within there.
[00:50:53] It was almost perfect for the whole 10 million we could have shipped this model. So they're saying like, we had a market [00:51:00] ready 10 million context window a year ago, which means they probably have a hundred million right now, 10 x, is, you know. but he said it's pretty expensive to run it. And basically no one had a use case for it.
[00:51:11] So like if we built this thing we put it out in the world, like no developer had anything that they would use it for. So we're like, ah, let's just like ship some other stuff. So I was like, okay, so this starts getting really interesting. But then he says, what I think is going to happen is when we achieve close to perfect million context, then it's going to unlock incredible applications, like something we could never imagine, ability to process information and connect the dots.
[00:51:34] It'll increase dramatically. What I think is going to happen, and this is Nikolai again, what I think is going to happen is these superhuman abilities, they're going to be more persuasive. the better the context we have, the more capabilities that we could never imagine are going to be on lock. So that's, uh, step number one.
[00:51:50] The quality is going to increase. and We're going to get nearly perfect retrieval after that. What's going to happen is the cost of the long context is going to decrease. So to do the 10 [00:52:00] million is gonna cost as the same as the 1 million today, and that's going to every 12 months this will happen. So to get to a billion in like two years, it'll basically cost them to do what they're doing today.
[00:52:10] So I think, so the cost will increase. So I think reasonably soon we will get that 10 million context window. I actually, I have no of this game. I would not be shocked at all if that actually gets announced at Iowa this week
[00:52:22] Paul Roetzer: That the 10 million is available to build on. that would be really logical.
[00:52:26] Actually, now that I'm thinking about it. so the 10 million becomes the commodity. So open AI chatt BT today is. 300,000. I don't even know what CHATT bet is. I don't even think
[00:52:34] Mike Kaput: It's a lot. It's a lot lower. because yeah, it's Yeah. Yeah.
[00:52:40] Paul Roetzer: window thing. so he says, when this happens and we get basically the super cheap 10 million, uh, that's going to be a deal breaker for some applications like coding.
[00:52:48] So go back to the coders. Do we really need people writing code And by 2030? He's saying no, because the flaw right now is code databases are massive. Hmm.
[00:52:58] Paul Roetzer: once you can put an [00:53:00] entire code base into it and it's perfect at retrieval, then you don't even, you don't need the humans to do the code. So that gets interesting.
[00:53:07] And he says, we're going to hold all this information in the memory at once and they're going to be reproducing any part of this information. Precisely. Imagine this in the legal industry. Not only that, they will also be able to connect the dots. They'll find the connections between files, better than coders.
[00:53:22] Basically, I imagine we will very soon get superhuman coding systems. They'll be totally unrivaled and they will basically become the new tool for every coder in the world. And so when this 10 million happens, that's the second step.
[00:53:34] Paul Roetzer: gets into agents that can operate effectively because now they have basically infinite context.
[00:53:38] they can keep track of everything. They don't lose, you know, previous actions, all these things. and then he says, so to keep all these previous interactions in memory, so we get back to memory, you need longer context. So that's where long context needed. so to, to realize, to wrap this up, to realize.
[00:53:54] The Sam is laying out for an operating system, which Google probably has the same vision. It's why they're turning on memory within your [00:54:00] emails and calendar and all these things and your search history. Once they have perfect retrieval of everything and everything you've ever done that you've allowed them access to is in the memory, in the context.
[00:54:10] 'cause it's basically infinite and they have perfect retrieval ability Instantly everything changes. And so like the key with this interview is this is two tech people. Like Nikolai is not the CEO trying to juice the stock price or raise the next 40 billion or whatever. This is a dude living within the labs building long context.
[00:54:30] He's like the expert on long context. He has no reason to hype this stuff. He's, he's just talking about the reality of what they're building. And that to me is like, so. I guess demonstrative of what is likely to come and why. I just feel like people generally have no concept of how fast stuff's gonna change because this change already lives in their labs, and that's usually we're getting things that's 12 plus months old of what they've already been able to [00:55:00] unlock internally.
[00:55:00] So there's always smarter versions So we always say like, this is the dumbest form of ai. we're gonna ever have notebook. LM is incredible as it is. Deep research is incredible as it is. They have smarter versions of it, a hundred percent within their labs that they're testing right now.
[00:55:14] Mike Kaput: Yeah, it's fascinating to that last
[00:55:16] point, because I don't think this applies to our audience, but there's so many people out there that I feel like just stopped updating their mental models.
[00:55:26] Like I still hear so much about, well, hallucinates well, it can't do this well, it can't, hold these documents in memory, whatever. And it's like, sure, I'm not denying any of that, but you really need to start paying attention to what's already existing.
[00:55:40] Paul Roetzer: Yeah. I just being realistic, like I don't, I, again, I don't think it's ignorance most of the time.
[00:55:46] Like I, yeah. I think it's denial, honestly. Like, I Yeah. think that the majority of people who aren't willing to like, realize that this is a distinct possibility that the next few years looks nothing [00:56:00] like today. I think it's just denial that like nothing can change that fast. It's not possible that it can do what I do.
[00:56:07] Like I think it's, and I'm not a psychologist by trade. I think it's like largely rooted in that.
[00:56:13] Mike Kaput: Yeah.
[00:56:14] Paul Roetzer: can't accept it, that it could possibly be that different that fast.
[00:56:18] Mike Kaput: I agree.
[00:56:20] Altman, Musk, and Grok Drama
[00:56:20] Mike Kaput: All right. Our next topic there is some more drama shaping up between Sam Altman and Elon Musk and Musk's Grok AI chatbot is in the middle of this, so a new
[00:56:32] spat between the two. Started off when Musk posted about Altman Online, he reposted something and appeared to call out Altman for past tweets critical of the Trump administration. Way back in 2016, Altman responded saying, Hey, I got it wrong back then.
[00:56:48] And then posted an article
[00:56:50] from 2022 that reported Musk was being Trump or critical of Trump at the time.
[00:56:57] Then Altman posted, Hey, anyway, see you next [00:57:00] week. In apparent reference to both of them being in the Middle East for a state visit by Trump, and he then said, quote, let's be friends. AGIis too important to let a little feud get in the way. Now must. With this.
[00:57:13] Paul Roetzer: playing.
[00:57:14] Mike Kaput: Yeah. Yeah. it's It's crazy. These are gonna be like the pieces of content that are like enshrined in history here, you know?
[00:57:22] is done. Yeah. And.
[00:57:25] Musk, I don't think responded to that one, but Altman took another opportunity to needle him a bit because it came out just after this, that grok spent several hours over this week for some reason, pushing a bizarre fixation on quote white genocide in South Africa.
[00:57:44] It was injecting this topic into all sorts of unrelated posts and responses. XAI comes out blaming the incident on an, what they call an unauthorized modification to its system prompt at 3:00 AM according to [00:58:00] XAI, someone altered rock's instructions to force this politically charged response, indirect violation apparently, of their internal policy.
[00:58:08] Altman then kind of gets the last laugh here
[00:58:11] saying he posted about this
[00:58:13] and said, quote, there are many ways this could have happened. I'm sure Xai will provide a full and transparent explanation soon, but this can only be properly understood in the context of white genocide in South Africa as an AI program to be maximally truth.
[00:58:27] And they, he trails off like totally trolling. so Paul, I kinda wanna focus on like there's obviously always this drama, this back and forth,
[00:58:38] This is not the first time something like this has happened with Grock and like, I guess I just have to ask because both times it's been seemingly like issues that one person in particular really cares about.
[00:58:52] And that person is Elon Musk, like, is he just going in and making these changes?
[00:58:56] Paul Roetzer: I, I don't know that is certainly the, what is being [00:59:00] implied by Sam and other people online is, so, you recall I, don't, like two months ago there was this issue again where GR got changed by an employee. Mm-hmm. they supposedly put a whole bunch of like, guardrails in place, so that couldn't happen.
[00:59:14] So I would imagine that the number of people who can at 3:00 AM make a change to the system prompt. Guides rock's, behavior is quite limited. And uh, there's certainly one person in particular you would assume has that ability. So I have no idea. But for me, at the bigger picture, level one, it's funny, obviously in, in the Middle East trip where they were there for like these energy deals and data center deals, which, you know, we're not gonna get into in this episode, I don't think, but like there was, massive deals being made, to, drive AI infrastructure basically.
[00:59:54] And so they were there together, which is what Sam had alluded to. And then this tweet kind of taking these [01:00:00] shots at, at grok happened six days later. So we're guessing they didn't make peace while they were, while they were there. It was not a peace summit. I would say, yeah, I, the fundamental problem to me that this is demonstrated of is the bias of the models we are relying on to power society.
[01:00:15] So like. basically five models right now that the vast majority of users use, and one of them, is gonna be embedded into humanoid robots cars, x all these places, apparently someone can just go in and change them and it completely changes their behavior in the way that the human is trying to change
[01:00:40] Paul Roetzer: So, grok, I gotta give him credit. The thing is pretty honest, like it's, it's training data has caused it to say all kinds of things that Elon Musk would not want it to say. There does just seem to be certain topics that really bother him, that then somehow magically, the system prompt is a, is [01:01:00] changed to not do that thing.
[01:01:02] This is the con, this is why there's all this push for, like, they need to be democratic,
[01:01:06] Mike Kaput: right,
[01:01:07] Paul Roetzer: because the humans decide they're democratic, they provide the training data to make them democratic in nature. So So this, it's just illustrative of the fact that we are reliant on human decisions and bias to actually build these supposedly unbiased models.
[01:01:22] Are Chatbots Replacing Search?
[01:01:22] Mike Kaput: Next up, a new survey came out from our friend, SEO expert, Andy Crestodina and the research firm Question Pro, and it has some compelling new data on AI and search. Their survey pulls over a thousand people in the US across a range
[01:01:37] of questions related to ai. I. Search. We'll put
[01:01:40] the link to the full recap that Andy posted in the show notes, but a couple interesting call outs they found, 72% of searchers use Google's AI overview when it appears, which has big implications for click-throughs and organic traffic.
[01:01:56] 62% of people use an AI chat bot every day. [01:02:00] 51% of chat Bott users
[01:02:01] plan to use them more.
[01:02:03] and 49% of people think AI will eventually replace search engines. additionally it talks about how Google's AI overviews now appear at the top of half of all search results and chatGPT, which obviously has barely existed for three years, is now the dominant AI tool.
[01:02:22] It is outpacing. By a long shot, Microsoft and Google's AI tools. And it's not always clear when people are reaching for chatbots over search based on the data. but this data does provide some more nuance. So for local info or reviews, it seems Google is still winning there. But for messy kind of multi-part questions, think something like find me venues in Chicago with vegan food and Riverview chatbots seem to increasingly be the default for people on those types of searches.
[01:02:54] Now, among all this, they point out very clearly search usage is actually up, [01:03:00] but the ways in which we seem to be searching are changing quite a bit, thanks to the existence of AI chatbots, bots.
[01:03:08] So Paul, I found this to be pretty helpful and nuanced.
[01:03:12] Like I like that we're not just saying, you know, search or SEO is dead or alive or whatever.
[01:03:17] We're just getting a more balanced picture of the changes ahead. But I do think we just have to accept that this is going to change a little faster than we anticipate, don't we?
[01:03:27] Paul Roetzer: Yeah. I'll just give a quick, like, personal anecdote here. So I was planning a trip to New York, so for, for Mother's Day, for my wife.
[01:03:34] She's, you know, loves libraries and bookstores and, rare books in particular. And I was like, I don't, like, that'd be a fun Mother's Day gift. Like, let me like plan a, like a book tour basically in New York. So I actually went to ChatGPT first, I was like, I'd love to like plan this like, unique getaway for my wife or for Mother's Day.
[01:03:50] I want, you know, what are some cities we could do it at? it I started with cities and actually came up with New York. and then I happened to have a talk in New York. So I was like, oh, perfect, like, let's do this. [01:04:00] So So it found the library hotel for me, which I did not know existed, that has Dewey decimal system floors and like rooms like it
[01:04:07] Mike Kaput: That's awesome.
[01:04:08] Paul Roetzer: That's awesome. It's across from the New York, library. So I was like, perfect. And I was like, okay, what else would you see if you were gonna go on a multi-day book tour New York book? What are the places you would have to see? And it wrote up a summary of like six bookstores at why they would be important.
[01:04:22] I gave a little context of like the kinds of books she likes, things like that. So this is all living in ChatGPT so far. Everything's happening. It's giving me URLs, whatever. So then I was like, all right, well let me map this out. Like how is this in relation now? I went to Google Maps and I actually like popped in each one, like, where are these things in the maps?
[01:04:37] So that gave me a function for, for that. And then the only time I ever actually went into Google was to, to look at the actual websites of each of the places. it was not the search function itself. And then I think I did one Google search just to verify of like, oh, what are the, pla I basically the same as I gave you the prompt, like what are the places we would have to see what are the bookstores?
[01:04:59] And it just. [01:05:00] Verified that, yeah. The six that I already had were like the places to go. So I mean, that's, that's an evolved search for me. That is a, like a total change in how I would've planned that trip previously from a year ago
[01:05:12] Mike Kaput: Right.
[01:05:12] Paul Roetzer: And honestly, it was amazing. And then I had cheche like, help me write descriptions of each of them and created like a, you know, a.
[01:05:19] Presentation with like about each one and Nice. It was perfect. So yeah, I think that more and more that's what's gonna happen and I do think the next generation, like, it's just, they're just never gonna go to,
[01:05:30] Mike Kaput: well, like we talked about with them, treating it as an operating system that changes everything.
[01:05:34] Paul Roetzer: Yeah, for sure.
[01:05:36] AI in Education Updates
[01:05:36] Mike Kaput: All right. Our next topic is related
[01:05:39] to kind of a hot topic we talked about last week. then we talked about a New York magazine article that went viral about how AI is causing a cheating epidemic in higher education. You know, in that article the educators interviewed were kind of almost at the point of despair about AI's impact on classrooms.
[01:05:58] they're [01:06:00] really worried it was replacing wholesale the need for students to do critical thinking and engage
[01:06:05] with the learning process. So
[01:06:07] we actually wanted to continue the discussion of AI in education this week with a couple other updates
[01:06:13] that provide some more context on this topic now.
[01:06:16] This week, it's actually about how AI could be impacting education in positive ways.
[01:06:21] So a couple developments that we've been tracking, bowling Green State University, which is in our backyard here in Ohio, is launching the first undergraduate degree of its kind in the us, which is a Bachelor of Science in what they call AI plus X. This is set to begin fall of this year. The program pairs core AI training
[01:06:40] with a second field of study.
[01:06:43] So you, IT is offering students the chance to apply AI to one of six disciplines, computer science, math, physics, history, journalism, or public relations. So the goal is to prepare graduates, not just as people who are comfortable around AI technology, [01:07:00] but also as domain aware technologists. Ready for a diverse set of careers. And then second, some more positive news. A new study seems to indicate that ChatGPT may not be having as adverse and effect on students as some people think it does. A new meta-analysis of 51 studies found that chat
[01:07:21] GPT significantly enhanced student learning, performance perception and higher order thinking.
[01:07:27] The strongest effect was on learning performance. Students benefited most when ChatGPT was integrated into problem-based learning models and skills development courses. The impact on learning perception, which is students' attitudes and motivations, was moderate and highly dependent on duration. Longer engagements beyond eight weeks or so led to better emotional engagement, possibly due to getting increasingly comfortable with using ChatGPT in an education context.
[01:07:56] and then short-term use, unfortunately showed limited [01:08:00] gains, but they
[01:08:00] say this is likely because students hadn't yet developed the prompting skills to use JGBT effectively.
[01:08:07] So Paul, we don't wanna like diminish the very real disruptive impact that AI is having on education right now, but it's also important to highlight the positive, which is when used in the
[01:08:18] right way, it seems like AI can have a really
[01:08:20] Good impact on learners. How do you weigh these positives and negatives when it comes to this? Because we know, we just see like the dark side so often here as well.
[01:08:29] Paul Roetzer: it just comes down to responsible use and teaching of responsible use. And you know, I think even my LinkedIn post about, That article, I had some wonderful comments from
[01:08:40] Mike Kaput: Yep.
[01:08:41] Paul Roetzer: education professionals who were like, yeah, here's what we're doing in our classroom. And some of it was amazing stuff. So there are absolutely outliers here of, you know, entire schools and, and more specifically into like the professors and teachers, administrators who are racing ahead to solve for this.
[01:08:56] And I think we're hopefully gonna hear more and more stories like [01:09:00] that. I've been an advocate since 2018, I think it was the first time I proposed this to a university that AI 1 0 1 should be required for every freshman. Like
[01:09:08] Hmm. It should be the absolute thing when you come in and then you have the fundamentals and then you like what BGS U is doing, which is phenomenal.
[01:09:16] Like You know, kudos to them for doing that. And thanks to my dad for sending me that article because I actually hadn't seen that one. I think we're going to Continue to see that. Now, I will say in the meantime, because this is not gonna be evenly distributed, I get asked all the time about advice from parents who have students who are in high school or college.
[01:09:34] And so I had, someone actually reached out to me last week about this. So here's the general advice I I give just for people who are trying to figure what to do. is like, regardless of major get whatever AI experience and education you can to compliment the major that they choose. each university is handling this differently, you know, in terms of what they offer as we see with BGSU.
[01:09:56] But I would look for any additional learning opportunities that are focused on the business [01:10:00] side of ai, not necessarily the computer science side. So, as we talked about earlier, we don't know the importance of coding. I'm a believer that learning to code has far. far beyond just building software.
[01:10:12] You know, I think I mentioned this recently, but it teaches you to do hard things, repetitive things. It teaches you to problem solve. It teaches you how to, you know, be, be a critical thinker. Like there's value to learning to code regardless of if you use it in your career. but there's tons of opportunities.
[01:10:26] Free low cost from Coursera, LinkedIn Learning, Google, Microsoft Open ai, like all the tech partners. HubSpot has stuff. So if you're looking for education, you can go, you know, add to your degrees and your studies with e-learning platforms. Our AI academy that we're, you know, we've had for five years, but like we're reimagining and rebuilding now, we're doing a lot more work with things that'll be additive to, we actually have partnerships with some universities who buy licenses for students in classes and then they take our classes as like complimentary to what they're learning in schools.
[01:10:57] And you know, we're working on some much [01:11:00] larger initiatives there under our AI literacy project to expand. our efforts in, in education. So yeah, we're not helpless here. There's good stuff being done and there's more that we each can do individually and then, you know, in partnership with some of these bigger organizations.
[01:11:15] The Cost of AI
[01:11:15] Mike Kaput: Our next topic
[01:11:18] is from AI expert Allie Miller, and she recently shared a somewhat disturbing anecdote about where AI might be headed. So she recently posted on x quote, A famous person wore a secret hidden wire while speaking with me about ai according to her account. She was at an event backstage talking about someone talking with someone about her private views on ai, and she joked at one point the conversation should have been recorded.
[01:11:45] The person I. Responded saying they were always micd up and Miller said, quote, he had a secret mic under his shirt the whole time. No disclosure. No consent, just recording. I felt completely violated.
[01:11:56] She then pointed out the connection here to AI tools [01:12:00] coming down the line. We have things like meta smart glasses, AI recording necklaces like limitless, that she says may lead to increased social friction if they become the norm because they duplicate this very uncomfortable environment, they might erode trust and compel people to be more on their guard.
[01:12:18] Now Paul, this is a wild story. It doesn't really, I guess, unfortunately surprise me, but is this the future we're headed towards where everyone records everything?
[01:12:28] Paul Roetzer: Yeah, this is go back to like the conversation around the limitless pendant and all that stuff, and I was not a big fan of it and I remain not a big fan of it.
[01:12:37] I do worry a lot about this. I don't know the answer. I mean, just in the last week we saw that Google or Apple is, you know, supposedly planning on integrating, you know, recording devices, cameras into watches. They're working on glasses. Apparently Meta already has glasses. Google is gonna have glasses.
[01:12:55] I'm guessing that we're actually gonna see a demonstration of Google's vision for glasses at, at io [01:13:00] this week. I
[01:13:00] Mike Kaput: Hmm.
[01:13:00] Paul Roetzer: a distinct possibility they reintroduce that project into the world. I don't know how we avoid it. I do honestly, like, I find myself uneasy around people wearing metal glasses already.
[01:13:12] Right. no idea if they're recording what we're saying. And it's not I'm saying something that I wouldn't want recorded, but like, I don't want to have to think about it. Like, it's just, you just, you're just having conversations and you, you, you assume some level of. Privacy if it's a one-on-one conversation and that it's not gonna end up somewhere.
[01:13:30] And so I feel very uneasy a about that stuff. And if I saw someone wearing a penant, I would naturally be guarded around them. I don't know. Like, I don't know how this plays out, and I think I'm more conservative than some people when it comes to this stuff. I know a lot of people are like, who cares?
[01:13:46] Everybody's got their phones out anyway. Everybody's recording everything. Like, I get it, but it's not the same thing. Like,
[01:13:50] Mike Kaput: Right.
[01:13:51] Paul Roetzer: That's a bad counter argument. so yeah, this is a very, very slippery slope. I think it's a direction we go. I don't [01:14:00] think the current administration does anything from, a regulatory standpoint to stop it.
[01:14:05] I could absolutely see this though, being a major legal thing,
[01:14:09] Hmm. you know, a couple years down the road,
[01:14:11] Mike Kaput: It's like a landmark case or two about this. Yeah. Yeah.
[01:14:16] All right, so as we come to the end of today's episode, I am going to quickly run through some AI product and funding updates, Paul, and then we're gonna kind of have you wrap things up by answering our listener question of the week.
[01:14:29] AI Product and Funding Updates
[01:14:29] Mike Kaput: Cool. All right. So a bunch of updates this week. First, perplexity,
[01:14:33] they're reportedly closing in on a $500 million funding round that would value them at $14 billion. That's a huge jump from last year's, $3 billion valuation, but notably lower than the 18 billion it initially targeted. Next up, TikTok has launched a new AI powered feature called AI Alive that turns still photos into short animated videos directly inside TikTok Stories.[01:15:00]
[01:15:00] This is the platform's first image to video creation tool.
[01:15:05] Meta is not having a great week. Unfortunately for them. They are quietly pushing back the release of their flagship AI model. LAMA four behemoth originally slated for spring, then bumped to summer. The launch has now been delayed again, likely into the fall or beyond generative AI video platform.
[01:15:25] Higgs Field AI has released a feature called Higgs Field Ads, which turns a single product picture into
[01:15:33] a studio ad in seconds. You literally just upload a product photo. Pick from 40 different design templates,
[01:15:41] and the tool will instantly create a cinematic ad spot. In some other video news video generation startup Hera just raised $32 million in a series A round led by Andreessen Horowitz.
[01:15:55] The company focuses on character driven video, which you may have seen in [01:16:00] one instance on display. Lately with trends like taking
[01:16:03] the internet by storm, like the AI generated videos of talking babies hosting podcasts, seems like there's at least a lot of money in talking babies hosting podcasts based on this. Google DeepMind has
[01:16:17] unveiled Alpha evolve an AI coding agent that designs entirely new algorithms, powered by Gemini models. It will generate thousands of code variations, test them, and evolve the best performers to solve tough problems in math and computing. Interestingly, it's already saved Google real compute.
[01:16:37] It found a scheduling trick that freed up 0.7% of capacity across data centers, which is huge at global scale. Open AI just
[01:16:46] launched a unusual and ambitious initiative. It's called the Open AI two Z Challenge, and it's literally a real world archeological treasure hunt. Powered by ai. Participants are invited to search for [01:17:00] undiscovered ancient sites hidden deep in the Amazon rainforest using
[01:17:05] satellite imagery, LIDAR scans, colonial diaries, and indigenous oral histories. Entrants must use open AI's latest models
[01:17:13] to analyze open source data, identify likely settlement patterns and proposed novel discoveries, and it kind of centers on finding this legendary lost city of Z, which is rumored to lie somewhere in Northern Brazil, the winning team gets $250,000 and the opportunity to collaborate with archeologists to verify their findings in the field.
[01:17:35] A couple other pieces of news here. Anthropic is about to release two upgraded AI models, Claude Sonnet and Claude Opus, which will take a bold step forward in how machines think. This seems to be from the early commentary Anthropics entry into
[01:17:51] test time compute, which is used by other models on the market to produce deeper reasoning by taking more time to think about [01:18:00] problems.
[01:18:00] Now, some rivals claim what Anthropic is. Teasing is not that different from what's already out there today, but Anthropic insists their approach will be more robust and adaptive. And last on my end, and Paul, I know you've gotten an update or two. Google io. We've mentioned a couple times. The company's developer conference is happening May 20th to 21st.
[01:18:21] So definitely expect some announcements from Google incoming days, which we will also cover on next week's episode
[01:18:28] Paul Roetzer: on the Baby Trend. HeyGen,
[01:18:30] Sorry,
[01:18:30] just, I think this was just yesterday. I saw this, HeyGen announced that in a Tweet He said, baby podcasts are everywhere and our intern just made it a feature. Now you can start a pod, even if they're in diapers.
[01:18:40] live now on HeyGen comment, who's these baby hosts remind you of was their tweet. So we'll put that tweet in the link. there is a rumor that Mike and I may be soon turned into baby version of the Artificial Intelligence Show, so stay tuned for that. I am sure our team will have some [01:19:00] fun, with that.
[01:19:01] And then, just a couple of quick notes. The alpha Evolve Things A big Deal, like we're not gonna get into it on this show in depth, but it's something we may come back around to. Like like like I said before, I think if you go back to AlphaGo and Alpha Zero, yeah. every alpha fold, is the distinct advantage that Google has that has yet to be played out into the models we all have access to.
[01:19:24] But I think if you go study what they've done there and what they're doing with Alpha Evolve, you start to see how they may actually have like a very strong competitive advantage over the other labs and then the A to Z challenge from OpenAI. Man, I love
[01:19:36] Mike Kaput: That's so cool
[01:19:37] Paul Roetzer: I wish I had like hours free that I could go.
[01:19:39] I would a hundred percent go play around with that. That is like, I love that there was a. Lost City of the Monkey God or something. There was a book I read about like how they used Lidar to find these. Oh man, I I could go all day on that stuff. Like space stuff and like hidden, you know, cultures and stuff.
[01:19:58] It's just so cool. Lost [01:20:00] cities and civilizations. I love that stuff. okay, listener question then we'll wrap up.
[01:20:04] Listener Question
[01:20:04] Mike Kaput: All right. So final segment. We're answering listener questions every week and this week's question is related to the topic of agents, which we talk about quite a bit. This one specifically about how agencies
[01:20:17] navigate agents. Do you see agencies in the future being mostly agents with a client facing team that runs the agents? So what are agents going to do to the agency model?
[01:20:28] Paul Roetzer: I dunno how far in the future we're talking. I think, you know, we have an AI for Agency Summit and this year my keynote, this was November of 24, was the last one.
[01:20:39] We do it annually and my keynote for that one was. I think it was like a AI agents in the future of agencies or something like that. So we'll put a link in the show notes. You can go get this on demand. this is the only place that talk is available, I think is on demand. yeah, I mean, I think agencies are gonna be fundamentally reimagined.
[01:20:58] In some cases [01:21:00] it might be, you know, totally AI native from the ground up, and you just build it with as few employees as possible. And if you focus on distinct services, like landing page optimization, media buying, like if it's a very specific repetitive tasks, and that's a your As, as a, as a whole, that's what your agency does is this collection of tasks that are repetitive and data driven and don't have long time horizon, you know, solutions to them.
[01:21:26] like, you know, more like consulting work then Sure. Like I could see these distinct, narrowly focused agencies probably being automated. Pretty quickly. Like we're Mo most of the work is probably done by the agents in, you know, say like a two to three year time period. If you're a bigger consulting firm and you know, you're doing more strategic work, then it's gonna totally change the way you do it through like, deep research products and things like that.
[01:21:54] But you're not eliminating the humans. You just need fewer humans. So I think agencies are [01:22:00] just gonna look fundamentally different in the next few years. Some will be highly automated, and some will just be way, way more efficient, create way more outputs, higher quality outputs through the integration of these tools.
[01:22:15] But I mean, they're, they're fundamentally gonna transform for sure.
[01:22:19] Mike Kaput: All right,
[01:22:21] Paul, that's a wrap on another busy week in ai. We've got a lot of updates coming down the line, so looking forward to next week as well. Thanks as always.
[01:22:29] Paul Roetzer: Yeah. And we, So Memorial day's, the following Monday. We'll, we'll have a regular episode.
[01:22:34] Yeah. So we'll do the regular weekly and we'll cover all the model updates and app news and all that good stuff. So, thanks Mike. And again, thanks As we started the show off, like Mike and I, you know, are truly grateful to have, you know, you all listening each week. We love hearing from you on LinkedIn or in person and
[01:22:50] Mike Kaput: Yep.
[01:22:50] Paul Roetzer: thanks for, you know, listening and being a part of this growing community.
[01:22:54] Thanks for listening to the Artificial Intelligence Show. Visit smarter x.ai to continue on [01:23:00] your AI learning journey and join more than 100,000 professionals and business leaders who have subscribed to our weekly newsletters, downloaded AI blueprints, attended virtual and in-person events, taken online AI courses, and earn professional certificates from our AI Academy and engaged in the Marketing AI Institute Slack community.
[01:23:19] Until next time, stay curious and explore ai.
Claire Prudhomme
Claire Prudhomme is the Marketing Manager of Media and Content at the Marketing AI Institute. With a background in content marketing, video production and a deep interest in AI public policy, Claire brings a broad skill set to her role. Claire combines her skills, passion for storytelling, and dedication to lifelong learning to drive the Marketing AI Institute's mission forward.