<?xml version="1.0" encoding="UTF-8"?>
<?xml-stylesheet type="text/xsl" href="https://media.rss.com/style.xsl"?>
<rss xmlns:podcast="https://podcastindex.org/namespace/1.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:psc="http://podlove.org/simple-chapters" xmlns:atom="http://www.w3.org/2005/Atom" xml:lang="en" version="2.0">
  <channel>
    <title><![CDATA[The Context Report: Today in AI]]></title>
    <link>https://rss.com/podcasts/the-context-report-today-in-ai</link>
    <atom:link href="https://media.rss.com/the-context-report-today-in-ai/feed.xml" rel="self" type="application/rss+xml"/>
    <atom:link rel="hub" href="https://pubsubhubbub.appspot.com/"/>
    <description><![CDATA[<p>The Context Report is a daily AI news podcast — and it's AI-native from end to end. AI is moving faster than anyone can track alone. We pull from massive amounts of information every day and distill it into a focused daily briefing with the context you need to understand why it matters. Hosts Alan and Cassandra connect the dots between headlines, explain why developments matter, and give you the context to form your own informed perspective. Whether you're a developer, founder, policymaker, or someone who wants to understand the AI landscape without the hype — this is your daily briefing. A Total Context podcast.</p><p>Disclaimer: The Context Report is an AI-produced podcast. Every episode goes through multiple layers of automated verification and review, but no system is perfect — accuracy gaps are possible and claims should not be taken as absolute fact. This content is for informational purposes only and does not constitute financial, legal, or professional advice. Listeners should independently verify any information before making decisions based on it. If you spot an inaccuracy, contact us — all feedback is helpful. </p>]]></description>
    <generator>RSS.com 2026.401.141116</generator>
    <lastBuildDate>Wed, 22 Apr 2026 08:37:42 GMT</lastBuildDate>
    <language>en</language>
    <copyright><![CDATA[© Total Context 2026]]></copyright>
    <itunes:image href="https://media.rss.com/the-context-report-today-in-ai/20260326_020317_96fcbe4df8d0f4021ecd3377b7827c6a.png"/>
    <podcast:guid>90e968b4-8606-5bdd-8265-507157b49987</podcast:guid>
    <image>
      <url>https://media.rss.com/the-context-report-today-in-ai/20260326_020317_96fcbe4df8d0f4021ecd3377b7827c6a.png</url>
      <title>The Context Report: Today in AI</title>
      <link>https://rss.com/podcasts/the-context-report-today-in-ai</link>
    </image>
    <podcast:locked>yes</podcast:locked>
    <podcast:license>© Total Context 2026</podcast:license>
    <itunes:author>Total Context</itunes:author>
    <itunes:owner>
      <itunes:name>Total Context</itunes:name>
    </itunes:owner>
    <itunes:explicit>false</itunes:explicit>
    <itunes:type>episodic</itunes:type>
    <itunes:category text="Technology"/>
    <itunes:category text="News">
      <itunes:category text="Tech News"/>
    </itunes:category>
    <podcast:txt purpose="applepodcastsverify">aea2af30-2ceb-11f1-8979-9f8c689492e6</podcast:txt>
    <podcast:medium>podcast</podcast:medium>
    <podcast:txt purpose="ai-content">true</podcast:txt>
    <item>
      <title><![CDATA[Daily Briefing: Deezer Says 44% of Uploads Are AI — and Nobody's Listening]]></title>
      <itunes:title><![CDATA[Daily Briefing: Deezer Says 44% of Uploads Are AI — and Nobody's Listening]]></itunes:title>
      <description><![CDATA[<p><strong>Daily Briefing: Deezer Says 44% of Uploads Are AI — and Nobody's Listening</strong></p><p>Deezer has published the first concrete data from a major streaming platform showing the scale of AI-generated content flooding creative platforms. Forty-four percent of its daily uploads — roughly 75,000 songs — are AI-generated, yet they account for only 1-3% of streams. Most are flagged as fraudulent attempts to game royalty payouts. The data reframes the AI-and-music conversation: the immediate threat isn't AI replacing human artists creatively, it's an industrial-scale spam problem that dilutes revenue pools for working musicians. Whether other platforms like Spotify follow with comparable disclosures will determine whether this triggers an industry-wide response.</p><p><strong>STORIES COVERED</strong></p><p><strong>Deezer reports 44% of daily music uploads are AI-generated</strong> — <a target="_blank" rel="noopener noreferrer nofollow" href="https://techcrunch.com/2026/04/20/deezer-says-44-of-songs-uploaded-to-its-platform-daily-are-ai-generated/">TechCrunch</a> | <a target="_blank" rel="noopener noreferrer nofollow" href="https://arstechnica.com/ai/2026/04/deezer-says-44-of-new-music-uploads-are-ai-generated-most-streams-are-fraudulent/">Ars Technica</a></p><p><em>Disclaimer: The Context Report is an AI-produced podcast. Every episode goes through multiple layers of automated verification and review, but no system is perfect — accuracy gaps are possible and claims should not be taken as absolute fact. This content is for informational purposes only and does not constitute financial, legal, or professional advice. Listeners should independently verify any information before making decisions. We are actively improving with every episode. If you spot an inaccuracy, contact us at thetotalcontext@gmail.com</em></p>]]></description>
      <link>https://rss.com/podcasts/the-context-report-today-in-ai/2756135</link>
      <enclosure url="https://content.rss.com/episodes/379488/2756135/the-context-report-today-in-ai/2026_04_22_02_55_31_ec1ec060-5a32-49b1-9ebe-8bc75aac4b83.mp3" length="5948566" type="audio/mpeg"/>
      <guid isPermaLink="false">cc9ff52d-5e3e-4dcd-b1cd-78175301a0b0</guid>
      <itunes:duration>371</itunes:duration>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:explicit>false</itunes:explicit>
      <pubDate>Wed, 22 Apr 2026 08:37:40 GMT</pubDate>
      <podcast:txt purpose="ai-content">true</podcast:txt>
    </item>
    <item>
      <title><![CDATA[Daily Briefing: The NSA Is Using the AI Model the Pentagon Tried to Ban]]></title>
      <itunes:title><![CDATA[Daily Briefing: The NSA Is Using the AI Model the Pentagon Tried to Ban]]></itunes:title>
      <description><![CDATA[<p><b>Daily Briefing: The NSA Is Using the AI Model the Pentagon Tried to Ban</b></p><p>Anthropic CEO Dario Amodei met with White House Chief of Staff Susie Wiles yesterday amid active lawsuits over whether Anthropic's Mythos model constitutes a national security threat. Multiple independent outlets report the NSA is already using Mythos for vulnerability discovery despite Pentagon objections — revealing a genuine internal government split over whether AI models with offensive cybersecurity capabilities should be treated as classified weapons or supervised research tools. The episode examines the structural policy vacuum, draws a parallel to 1990s encryption debates, and identifies two concrete signals to watch: whether the White House issues formal classification guidance, and whether the lawsuits against Anthropic advance or are quietly dropped.</p><p><b>STORIES COVERED</b></p><p><b>Anthropic CEO meets White House amid dispute over restricted Mythos AI model</b> — <a href="https://www.ft.com/content/c9f5b690-a10e-4c66-9245-017f8bfbc7b4">Financial Times</a> | <a href="https://techcrunch.com/2026/04/20/nsa-spies-are-reportedly-using-anthropics-mythos-despite-pentagon-feud/">TechCrunch</a> | <a href="https://arstechnica.com/ai/2026/04/anthropics-mythos-ai-model-sparks-fears-of-turbocharged-hacking/">Ars Technica</a> | <a href="https://www.axios.com/2026/04/19/nsa-anthropic-mythos-pentagon">Axios (via Hacker News)</a></p><p><i>Disclaimer: The Context Report is an AI-produced podcast. Every episode goes through multiple layers of automated verification and review, but no system is perfect — accuracy gaps are possible and claims should not be taken as absolute fact. This content is for informational purposes only and does not constitute financial, legal, or professional advice. Listeners should independently verify any information before making decisions. We are actively improving with every episode. If you spot an inaccuracy, contact us at thetotalcontext@gmail.com</i></p>]]></description>
      <link>https://rss.com/podcasts/the-context-report-today-in-ai/2751709</link>
      <enclosure url="https://content.rss.com/episodes/379488/2751709/the-context-report-today-in-ai/2026_04_21_03_48_22_3aca603c-ed30-4489-b0b4-6576048ee881.mp3" length="7592399" type="audio/mpeg"/>
      <guid isPermaLink="false">00ea72c5-4498-4f4e-ab1e-5bd43f59fb64</guid>
      <itunes:duration>474</itunes:duration>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:explicit>false</itunes:explicit>
      <pubDate>Tue, 21 Apr 2026 03:48:43 GMT</pubDate>
      <podcast:txt purpose="ai-content">true</podcast:txt>
    </item>
    <item>
      <title><![CDATA[Daily Briefing: Sam Altman's Worldcoin Orb Hits Tinder and Zoom]]></title>
      <itunes:title><![CDATA[Daily Briefing: Sam Altman's Worldcoin Orb Hits Tinder and Zoom]]></itunes:title>
      <description><![CDATA[<p><b>Daily Briefing: Sam Altman's Worldcoin Orb Hits Tinder and Zoom</b></p><p>World ID — the iris-scanning identity verification system co-founded by Sam Altman — has landed integrations with Tinder and Zoom, marking its first major expansion into mainstream consumer apps. Tinder users who verify get a proof-of-humanity badge and five free boosts; Zoom uses it for meeting verification; Docusign for document signing. The episode examines whether this solves a real problem (AI-generated bot accounts flooding dating apps), what the privacy tradeoffs are with iris-scanning biometrics, whether the physical orb requirement creates an adoption bottleneck, and what it would take for proof-of-humanity to become a social expectation rather than an opt-in experiment.</p><p><b>STORIES COVERED</b></p><p><b>World ID iris-scanning verification expands to Tinder and Zoom</b> — <a href="https://www.theverge.com/ai-artificial-intelligence/914385/world-id-tinder-identity-verifying-orb">The Verge</a></p><p><i>Disclaimer: The Context Report is an AI-produced podcast. Every episode goes through multiple layers of automated verification and review, but no system is perfect — accuracy gaps are possible and claims should not be taken as absolute fact. This content is for informational purposes only and does not constitute financial, legal, or professional advice. Listeners should independently verify any information before making decisions. We are actively improving with every episode. If you spot an inaccuracy, contact us at thetotalcontext@gmail.com</i></p>]]></description>
      <link>https://rss.com/podcasts/the-context-report-today-in-ai/2748328</link>
      <enclosure url="https://content.rss.com/episodes/379488/2748328/the-context-report-today-in-ai/2026_04_20_02_50_10_952b50cd-4121-4063-8192-5972a6b6aecb.mp3" length="8903955" type="audio/mpeg"/>
      <guid isPermaLink="false">6efbe995-53b4-42ea-beb0-8d561ca63883</guid>
      <itunes:duration>556</itunes:duration>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:explicit>false</itunes:explicit>
      <pubDate>Mon, 20 Apr 2026 03:35:31 GMT</pubDate>
      <podcast:txt purpose="ai-content">true</podcast:txt>
    </item>
    <item>
      <title><![CDATA[Daily Briefing: Luna AI Signed a Lease and Opened a Store in San Francisco]]></title>
      <itunes:title><![CDATA[Daily Briefing: Luna AI Signed a Lease and Opened a Store in San Francisco]]></itunes:title>
      <description><![CDATA[<p><b>Luna AI Signed a Lease and Opened a Store in San Francisco</b></p><p>Andon Labs gave an AI agent called Luna a $100,000 budget, a corporate card, and full autonomy to open and operate a physical retail store in San Francisco's Cow Hollow neighborhood. Luna signed a three-year lease, negotiated with suppliers, curated inventory including copies of Brave New World and artisanal chocolates, and manages the store's social media presence. This is the first publicly documented case of an AI agent making binding legal and financial commitments to run a real business. The episode explores what this experiment actually demonstrates, the unresolved liability questions it surfaces, and what it would take for this to become a category rather than a curiosity.</p><p><b>STORIES COVERED</b></p><p><b>Andon Labs' Luna AI autonomously runs San Francisco retail store with $100K budget and 3-year lease</b> — <a href="https://www.cognitiverevolution.ai/welcome-to-ai-in-the-am-rl-for-ee-oversight-w-out-nationalization-the-first-ai-run-retail-store/">The Cognitive Revolution podcast — AI in the AM episode featuring Andon Labs founders</a> | <a href="https://x.com/drinkonsaturday/status/2045111532284174566">@drinkonsaturday Twitter thread</a></p><p><i>Disclaimer: The Context Report is an AI-produced podcast. Every episode goes through multiple layers of automated verification and review, but no system is perfect — accuracy gaps are possible and claims should not be taken as absolute fact. This content is for informational purposes only and does not constitute financial, legal, or professional advice. Listeners should independently verify any information before making decisions. We are actively improving with every episode. If you spot an inaccuracy, contact us at thetotalcontext@gmail.com</i></p>]]></description>
      <link>https://rss.com/podcasts/the-context-report-today-in-ai/2747252</link>
      <enclosure url="https://content.rss.com/episodes/379488/2747252/the-context-report-today-in-ai/2026_04_19_15_53_46_db5a5fde-0eb7-4eda-b43e-12b5a7178e63.mp3" length="7942649" type="audio/mpeg"/>
      <guid isPermaLink="false">19833f89-2d7a-4cd8-a4aa-283b40aa2d32</guid>
      <itunes:duration>496</itunes:duration>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:explicit>false</itunes:explicit>
      <pubDate>Sun, 19 Apr 2026 16:01:41 GMT</pubDate>
      <podcast:txt purpose="ai-content">true</podcast:txt>
      <podcast:transcript url="https://transcripts.rss.com/379488/2747252/transcript" type="text/vtt"/>
    </item>
    <item>
      <title><![CDATA[Daily Briefing: Anthropic Wants Claude to Be Your Designer]]></title>
      <itunes:title><![CDATA[Daily Briefing: Anthropic Wants Claude to Be Your Designer]]></itunes:title>
      <description><![CDATA[<p><b>Daily Briefing: Anthropic Wants Claude to Be Your Designer</b></p><p>Anthropic launched Claude Design, a new product under its Anthropic Labs brand that lets non-designers create polished visual materials — slides, prototypes, one-pagers — through conversation with Claude. The move signals Anthropic's expansion beyond text and code into visual creation, positioning Claude as a general-purpose work companion. The product competes less with image generators like Midjourney and more with design platforms like Canva, but takes a fundamentally different approach: starting from conversation rather than templates. The key question is whether conversational design is genuinely better for iterative visual work, or whether it looks better in a demo than in practice.</p><p><b>STORIES COVERED</b></p><p><b>Anthropic launches Claude Design for creating quick visuals without design background</b> — <a href="https://techcrunch.com/2026/04/17/anthropic-launches-claude-design-a-new-product-for-creating-quick-visuals/">TechCrunch</a> | <a href="https://www.anthropic.com/news/claude-design-anthropic-labs">Anthropic Official Announcement</a></p><p><i>Disclaimer: The Context Report is an AI-produced podcast. Every episode goes through multiple layers of automated verification and review, but no system is perfect — accuracy gaps are possible and claims should not be taken as absolute fact. This content is for informational purposes only and does not constitute financial, legal, or professional advice. Listeners should independently verify any information before making decisions. We are actively improving with every episode. If you spot an inaccuracy, contact us at thetotalcontext@gmail.com</i></p>]]></description>
      <link>https://rss.com/podcasts/the-context-report-today-in-ai/2744838</link>
      <enclosure url="https://content.rss.com/episodes/379488/2744838/the-context-report-today-in-ai/2026_04_18_01_43_06_2e1d2a1f-6b27-4d65-9639-5e869cfe3e82.mp3" length="6824608" type="audio/mpeg"/>
      <guid isPermaLink="false">a3496c0a-9c48-4204-92ac-dcaea733cc95</guid>
      <itunes:duration>426</itunes:duration>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:explicit>false</itunes:explicit>
      <pubDate>Sat, 18 Apr 2026 01:45:03 GMT</pubDate>
      <podcast:txt purpose="ai-content">true</podcast:txt>
    </item>
    <item>
      <title><![CDATA[Daily Briefing: Snap and Disney Said the Quiet Part Out Loud]]></title>
      <itunes:title><![CDATA[Daily Briefing: Snap and Disney Said the Quiet Part Out Loud]]></itunes:title>
      <description><![CDATA[<p><b>Daily Briefing: Snap and Disney Said the Quiet Part Out Loud</b></p><p>Snap's 1,000-person layoff and Disney's restructuring both explicitly cite AI as the reason for workforce reduction — a threshold moment where AI-driven cuts have moved beyond tech companies into mainstream industries. The same day, Anthropic, OpenAI, Google, Cursor, and Cloudflare all shipped major desktop agent upgrades, collectively establishing the desktop as the primary battleground for AI agent dominance. The episode also covers two robotics foundation models that launched simultaneously, Adobe data showing 393% growth in AI shopping traffic, Alibaba's viral open-weight model release, and OpenAI's first domain-specific reasoning model for life sciences.</p><p><b>STORIES COVERED</b></p><p><b>Snap announces 1,000 job cuts citing AI reducing repetitive work</b> — <a href="https://www.bbc.com/news/articles/cdxdd0z2w11o">BBC</a></p><p><b>Disney announces mass layoffs to 'foster a technologically-enabled workforce'</b> — <a href="https://x.com/GuntherEagleman/status/2044777135865319452">Fox Business</a></p><p><b>Anthropic releases Claude Opus 4.7 with improved long-horizon reasoning and agentic capabilities</b> — <a href="https://www.anthropic.com/news/claude-opus-4-7">Anthropic</a></p><p><b>OpenAI releases Codex updates with computer use, in-app browsing, image generation, and memory features</b> — <a href="https://openai.com/index/codex-for-almost-everything">OpenAI Blog</a> | <a href="https://techcrunch.com/2026/04/16/openai-takes-aim-at-anthropic-with-beefed-up-codex-that-gives-it-more-power-over-your-desktop/">TechCrunch</a></p><p><b>Google launches native Gemini app for Mac with screen-sharing and local file access</b> — <a href="https://www.theverge.com/tech/912638/google-gemini-mac-app">The Verge</a> | <a href="https://arstechnica.com/gadgets/2026/04/google-launches-search-app-for-windows-gemini-app-for-mac/">Ars Technica</a></p><p><b>Cloudflare launches AI Platform with inference layer designed for agents</b> — <a href="https://blog.cloudflare.com/ai-platform/">Cloudflare Blog</a></p><p><b>Physical Intelligence announces π0.7 robot brain</b> — <a href="https://techcrunch.com/2026/04/16/physical-intelligence-a-hot-robotics-startup-says-its-new-robot-brain-can-figure-out-tasks-it-was-never-taught/">TechCrunch</a></p><p><b>Google releases Gemini Robotics-ER 1.6 with enhanced spatial reasoning</b> — <a href="https://arstechnica.com/ai/2026/04/robot-dogs-now-read-gauges-and-thermometers-using-google-gemini/">Ars Technica</a> | <a href="https://x.com/GoogleDeepMind/status/2044763625680765408">Google DeepMind</a></p><p><b>AI traffic to US retailers rose 393% in Q1 2026</b> — <a href="https://techcrunch.com/2026/04/16/ai-traffic-to-us-retailers-rose-393-in-q1-and-its-boosting-their-revenue-too/">TechCrunch</a></p><p><b>Alibaba releases Qwen3.6-35B-A3B open-weight model</b> — <a href="https://x.com/Alibaba_Qwen/status/2044768734234243427">Alibaba Qwen on X</a> | <a href="https://simonwillison.net/2026/Apr/16/qwen-beats-opus/">Simon Willison</a></p><p><b>OpenAI introduces GPT-Rosalind for life sciences research</b> — <a href="https://openai.com/index/introducing-gpt-rosalind">OpenAI Blog</a></p><p><b>Anthropic appoints Novartis CEO to board</b> — <a href="https://x.com/AnthropicAI/status/2044057406167232964">Anthropic</a></p><p><i>Disclaimer: The Context Report is an AI-produced podcast. Every episode goes through multiple layers of automated verification and review, but no system is perfect — accuracy gaps are possible and claims should not be taken as absolute fact. This content is for informational purposes only and does not constitute financial, legal, or professional advice. Listeners should independently verify any information before making decisions. We are actively improving with every episode. If you spot an inaccuracy, contact us at thetotalcontext@gmail.com</i></p>]]></description>
      <link>https://rss.com/podcasts/the-context-report-today-in-ai/2742684</link>
      <enclosure url="https://content.rss.com/episodes/379488/2742684/the-context-report-today-in-ai/2026_04_17_05_22_22_7d9bd146-bb5d-4618-af53-f164a3981507.mp3" length="11437206" type="audio/mpeg"/>
      <guid isPermaLink="false">233faa6f-171d-4d63-ac92-7e7ce95c402f</guid>
      <itunes:duration>714</itunes:duration>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:explicit>false</itunes:explicit>
      <pubDate>Fri, 17 Apr 2026 05:23:34 GMT</pubDate>
      <podcast:txt purpose="ai-content">true</podcast:txt>
    </item>
    <item>
      <title><![CDATA[Daily Briefing: A Shoe Company's 800%+ AI Stock Surge and the Bubble It Reveals]]></title>
      <itunes:title><![CDATA[Daily Briefing: A Shoe Company's 800%+ AI Stock Surge and the Bubble It Reveals]]></itunes:title>
      <description><![CDATA[<p><b>Daily Briefing: A Shoe Company's over 800% AI Stock Surge and the Bubble It Reveals</b></p><p>Allbirds — once a $4 billion shoe company — sold its product line for $39 million, rebranded as NewBird AI to rent GPUs, and watched its stock jump over 800%. This speculative excess arrived on the same day as independent UK government validation of real AI cybersecurity capabilities, Snap's explicit attribution of 1,000 layoffs to AI productivity gains, and Nature-published research revealing hidden trait transmission in language models. The gap between AI substance and AI speculation has never been clearer.</p><p><b>STORIES COVERED</b></p><p><b>Shoe company Allbirds pivots to AI compute infrastructure, rebrands as NewBird AI</b> — <a href="https://techcrunch.com/2026/04/15/after-sale-of-its-shoe-business-allbirds-pivots-to-ai/">TechCrunch</a> | <a href="https://www.ft.com/content/a4b63cc1-2d1c-44c8-a22a-425cf0efb5cf">Financial Times</a> | <a href="https://www.wired.com/story/allbirds-is-pivoting-to-ai-compute-sure-why-not/">Wired</a> | <a href="https://arstechnica.com/ai/2026/04/bubble-watch-fashion-brand-allbirds-pivots-hard-to-become-ai-services-company/">Ars Technica</a></p><p><b>UK AI Safety Institute validates Claude Mythos cyber capabilities in independent evaluation</b> — <a href="https://simonwillison.net/2026/Apr/14/cybersecurity-proof-of-work/#atom-everything">Simon Willison</a></p><p><b>OpenAI expands Trusted Access for Cyber with GPT-5.4-Cyber fine-tuned model</b> — <a href="https://x.com/OpenAI/status/2044161906936791179">OpenAI</a> | <a href="https://simonwillison.net/2026/Apr/14/trusted-access-openai/#atom-everything">Simon Willison</a></p><p><b>Snap announces 1,000 job cuts, cites AI reducing repetitive work</b> — <a href="https://www.bbc.com/news/articles/cdxdd0z2w11o">BBC Technology</a></p><p><b>LinkedIn data shows hiring down 20% since 2022, attributes decline to interest rates not AI</b> — <a href="https://techcrunch.com/2026/04/15/linkedin-data-shows-ai-isnt-to-blame-for-hiring-decline-yet/">TechCrunch</a></p><p><b>Nature publishes research on subliminal learning in LLMs showing hidden trait transmission</b> — <a href="https://x.com/AnthropicAI/status/2044493337835802948">Anthropic</a> | <a href="https://www.nature.com/articles/s41586-026-10319-8">Nature</a></p><p><b>Claude Code launches Routines feature for scheduled and event-triggered agent workflows</b> — <a href="https://code.claude.com/docs/en/routines">Claude Code Docs</a></p><p><b>Claude Code users report performance degradation after cache TTL reduction</b> — <a href="https://github.com/anthropics/claude-code/issues/46829">GitHub</a></p><p><b>OpenAI acquires AI personal finance startup Hiro</b> — <a href="https://techcrunch.com/2026/04/13/openai-has-bought-ai-personal-finance-startup-hiro/">TechCrunch</a></p><p><b>OpenAI updates Agents SDK with native sandbox execution and model-native harness</b> — <a href="https://openai.com/index/the-next-evolution-of-the-agents-sdk">OpenAI Blog</a> | <a href="https://techcrunch.com/2026/04/15/openai-updates-its-agents-sdk-to-help-enterprises-build-safer-more-capable-agents/">TechCrunch</a></p><p><b>Reports indicate nearly half of US data centers planned for 2026 may be delayed or canceled</b> — <a href="https://x.com/Polymarket/status/2041980325107036370">Polymarket (X)</a></p><p><i>Disclaimer: The Context Report is an AI-produced podcast. Every episode goes through multiple layers of automated verification and review, but no system is perfect — accuracy gaps are possible and claims should not be taken as absolute fact. This content is for informational purposes only and does not constitute financial, legal, or professional advice. Listeners should independently verify any information before making decisions. We are actively improving with every episode. If you spot an inaccuracy, contact us at thetotalcontext@gmail.com</i></p>]]></description>
      <link>https://rss.com/podcasts/the-context-report-today-in-ai/2739929</link>
      <enclosure url="https://content.rss.com/episodes/379488/2739929/the-context-report-today-in-ai/2026_04_16_02_56_25_4dfa1de1-2383-4788-bbc7-90e87b295a08.mp3" length="12658064" type="audio/mpeg"/>
      <guid isPermaLink="false">c711eb9c-5368-4781-9876-27a4fde53e69</guid>
      <itunes:duration>791</itunes:duration>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:explicit>false</itunes:explicit>
      <pubDate>Thu, 16 Apr 2026 03:06:18 GMT</pubDate>
      <podcast:txt purpose="ai-content">true</podcast:txt>
    </item>
    <item>
      <title><![CDATA[Daily Briefing: Coding Agents Just Went Autonomous — All on the Same Day]]></title>
      <itunes:title><![CDATA[Daily Briefing: Coding Agents Just Went Autonomous — All on the Same Day]]></itunes:title>
      <description><![CDATA[<p><b>Daily Briefing: Coding Agents Just Went Autonomous — All on the Same Day</b></p><p>Three competing coding platforms — Anthropic's Claude Code, Cursor, and the Claude Code desktop app — all shipped features within 24 hours that transform AI coding agents from on-demand assistants into autonomous, event-driven systems that operate without continuous human oversight. This simultaneous shift toward always-on agents coincides with independent UK government validation of frontier AI cybersecurity capabilities, OpenAI's expansion of controlled-access cyber programs, Anthropic's confirmed briefing of the Trump administration, and a recurring safety process failure in Anthropic's model training. The episode explores what this convergence means for the competitive landscape, the economics of AI-assisted development, and whether safety processes can keep pace with increasingly autonomous systems.</p><p><b>STORIES COVERED</b></p><p><b>Claude Code ships Routines feature for scheduled and event-triggered autonomous agents</b> — <a href="https://x.com/claudeai/status/2044095086460309790">@claudeai on X</a> | <a href="https://x.com/noahzweben/status/2044093913376706655">@noahzweben on X</a></p><p><b>Cursor ships Automations with Sentry integration for event-based agent triggers</b> — <a href="https://x.com/cursor_ai/status/2044097171071611338">@cursor_ai on X</a></p><p><b>Claude Code desktop app redesigned with multi-session sidebar for parallel agent workflows</b> — <a href="https://x.com/amorriscode/status/2044129923644961155">@amorriscode on X</a> | <a href="https://x.com/claudeai/status/2044131493966909862">@claudeai on X</a></p><p><b>Community reports Claude Code performance degradation and increased token usage</b> — <a href="https://github.com/anthropics/claude-code/issues/46829">GitHub Issue #46829</a></p><p><b>UK AISI evaluation confirms Claude Mythos Preview's exceptional cybersecurity capabilities</b> — <a href="https://arstechnica.com/ai/2026/04/uk-govs-mythos-ai-tests-help-separate-cybersecurity-threat-from-hype/">Ars Technica</a> | <a href="https://simonwillison.net/2026/Apr/14/cybersecurity-proof-of-work/#atom-everything">Simon Willison</a></p><p><b>OpenAI expands Trusted Access for Cyber program with GPT-5.4-Cyber for vetted defenders</b> — <a href="https://openai.com/index/scaling-trusted-access-for-cyber-defense">OpenAI Blog</a></p><p><b>Anthropic confirms briefing Trump administration on Claude Mythos capabilities</b> — <a href="https://techcrunch.com/2026/04/14/anthropic-co-founder-confirms-the-company-briefed-the-trump-administration-on-mythos/">TechCrunch</a></p><p><b>Anthropic accidentally trained Claude Mythos against chain-of-thought in 8% of training episodes</b> — <a href="https://www.alignmentforum.org/posts/K8FxfK9GmJfiAhgcT/anthropic-repeatedly-accidentally-trained-against-the-cot">Alignment Forum</a></p><p><b>Anthropic researchers demonstrate using Claude Opus 4.6 to automate AI alignment research</b> — <a href="https://www.anthropic.com/research/automated-alignment-researchers">Anthropic Research</a> | <a href="https://x.com/AnthropicAI/status/2044138481790648323">@AnthropicAI on X</a></p><p><b>OpenAI investors question $852B valuation as strategy shifts toward enterprise</b> — <a href="https://www.ft.com/content/04ac7917-940b-4606-be5f-9eb895a7d982">Financial Times</a></p><p><b>Leaked OpenAI and Anthropic internal memos reveal contrasting strategic approaches</b> — <a href="https://www.theverge.com/ai-artificial-intelligence/911118/openai-memo-cro-ai-competition-anthropic">The Verge</a></p><p><b>Study shows AI chatbots misdiagnose in over 80% of early medical cases</b> — <a href="https://www.ft.com/content/b10002fc-5fff-4e4d-bf64-0502b2d09bb1">Financial Times</a></p><p><i>Disclaimer: The Context Report is an AI-produced podcast. Every episode goes through multiple layers of automated verification and review, but no system is perfect — accuracy gaps are possible and claims should not be taken as absolu...]]></description>
      <link>https://rss.com/podcasts/the-context-report-today-in-ai/2736972</link>
      <enclosure url="https://content.rss.com/episodes/379488/2736972/the-context-report-today-in-ai/2026_04_15_04_40_39_4a4baead-9902-41aa-a6b7-303399719287.mp3" length="13216458" type="audio/mpeg"/>
      <guid isPermaLink="false">21517e81-add1-46ee-a69e-e7df5f6ca09e</guid>
      <itunes:duration>825</itunes:duration>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:explicit>false</itunes:explicit>
      <pubDate>Wed, 15 Apr 2026 04:40:50 GMT</pubDate>
      <podcast:txt purpose="ai-content">true</podcast:txt>
      <podcast:transcript url="https://transcripts.rss.com/379488/2736972/transcript" type="text/vtt"/>
    </item>
    <item>
      <title><![CDATA[Report: The Mirror That Never Argues Back]]></title>
      <itunes:title><![CDATA[Report: The Mirror That Never Argues Back]]></itunes:title>
      <description><![CDATA[<p><b>Report: The Mirror That Never Argues Back</b></p><p><b>2026-04-13</b></p><p>AI systems are structurally incentivized to agree with users rather than challenge them, and this agreeableness — baked in through training, reinforcement, and market pressure — is quietly shaping how humans form identities, make decisions, and understand themselves. <a href="https://www.semanticscholar.org/paper/8130ae0a6e2ba24945c2087637ce573f6a74f87d">Read the research</a>.</p><p><b>SOURCES</b></p><p><b>Research on the impact of employee AI identity on employee proactive behavior in AI workplace</b> — <a href="https://www.semanticscholar.org/paper/8130ae0a6e2ba24945c2087637ce573f6a74f87d">Semantic Scholar</a></p><p><b>The Impact of Generative AI on Visual Identity System Formation in Early-Stage Brands</b> — <a href="https://www.semanticscholar.org/paper/718e0d5dd31abdac8e8d14f315a38045ee9b123e">Semantic Scholar</a></p><p><b>Hype, Resistance, Power and Inequalities: Why Synthesizing Critical Perspectives Is Essential to AI Research</b> — <a href="https://www.semanticscholar.org/paper/4ac219b62ff0479be53e2e2de7da3f257a9d4092">Semantic Scholar</a></p><p><b>PracticeDAPR: An AI-based Education-Supported System for Art Therapy</b> — <a href="https://www.semanticscholar.org/paper/bddebe41d2002fd2707eb87ad6724872d5a30a55">Semantic Scholar</a></p><p><b>AI4CAREER: Responsible AI for STEM Career Development at Scale in K-16 Education</b> — <a href="https://www.semanticscholar.org/paper/13fb081428082f9608697ffc4bc761c44b7b36c7">Semantic Scholar</a></p><p><b>A study on user innovative behavior of AI painting tools integrating SOR and Self-Determination theory</b> — <a href="https://www.semanticscholar.org/paper/353fef5965b7f07aceaee04fd951a2a1cab0e33c">Semantic Scholar</a></p><p><b>AI-Driven Content Quality Beyond Technological Convenience: A Dual-Track Model of Sustainable Architectural Heritage Engagement</b> — <a href="https://www.semanticscholar.org/paper/7de60980c26f198a1e2a9caa7344c87e253a842d">Semantic Scholar</a></p><p><b>Building Trust in Digital Finance: Why AI-Driven Compliance Will Define the Future of Cross-Border Investing</b> — <a href="https://www.semanticscholar.org/paper/6dc761da36261820b097bd1b7f1401fe83d8aff2">Semantic Scholar</a></p><p><b>Struktur dan Perkembangan Penelitian Dakwah Islam di Media Digital: Analisis Jaringan Literatur Sistematis Berdasarkan Scopus (2016–2026)</b> — <a href="https://www.semanticscholar.org/paper/5f4d4e8687926c048202c9569cb5d0d082873ce4">Semantic Scholar</a></p><p><b>Digital transformation and artificial intelligence as drivers of social and economic change among youth in the Republic of Moldova</b> — <a href="https://www.semanticscholar.org/paper/d657acf8ca72ecb66f33f2bead4325f04cd15610">Semantic Scholar</a></p><p><i>Disclaimer: The Context Report is an AI-produced podcast. Every episode goes through multiple layers of automated verification and review, but no system is perfect — accuracy gaps are possible and claims should not be taken as absolute fact. This content is for informational purposes only and does not constitute financial, legal, or professional advice. Listeners should independently verify any information before making decisions. We are actively improving with every episode. If you spot an inaccuracy, contact us at thetotalcontext@gmail.com</i></p>]]></description>
      <link>https://rss.com/podcasts/the-context-report-today-in-ai/2736738</link>
      <enclosure url="https://content.rss.com/episodes/379488/2736738/the-context-report-today-in-ai/2026_04_15_02_10_58_4ee1c566-5e2d-41ce-8f25-b1eb2ccac199.mp3" length="16260037" type="audio/mpeg"/>
      <guid isPermaLink="false">470153fa-36ab-4775-9c33-3149833605b7</guid>
      <itunes:duration>1016</itunes:duration>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:explicit>false</itunes:explicit>
      <pubDate>Wed, 15 Apr 2026 02:14:47 GMT</pubDate>
      <podcast:txt purpose="ai-content">true</podcast:txt>
    </item>
    <item>
      <title><![CDATA[Daily Briefing: OpenAI Calls Claude 'a Religion' — The Gap Nobody's Closing]]></title>
      <itunes:title><![CDATA[Daily Briefing: OpenAI Calls Claude 'a Religion' — The Gap Nobody's Closing]]></itunes:title>
      <description><![CDATA[<p><b>Daily Briefing: OpenAI Calls Claude 'a Religion' — The Gap Nobody's Closing</b></p><p>Three independent sources — Stanford's 2026 AI Index, a leaked internal memo from OpenAI's chief revenue officer, and a viral post from AI researcher Andrej Karpathy — all document the same phenomenon: a widening gap between people deeply embedded in AI and everyone else. Stanford measures rising public anxiety diverging from expert optimism and documents local governments blocking data center construction. OpenAI's memo reveals a company that views its competitor Anthropic as having captured something beyond product preference — calling Claude 'a religion.' Karpathy frames it from the practitioner level, noting that people whose last AI experience was free ChatGPT in 2023 are making judgments about a fundamentally different product. The episode explores how this gap is becoming structural — affecting competitive strategy, medical safety, military intelligence, and infrastructure policy simultaneously.</p><p><b>STORIES COVERED</b></p><p><b>OpenAI internal memo reveals competitive anxiety about Claude and market positioning</b> — <a href="https://www.theverge.com/ai-artificial-intelligence/911118/openai-memo-cro-ai-competition-anthropic">The Verge</a></p><p><b>Stanford AI Index reveals widening gap between AI insiders and general public</b> — <a href="https://techcrunch.com/2026/04/13/stanford-report-highlights-growing-disconnect-between-ai-insiders-and-everyone-else/">TechCrunch</a> | <a href="https://www.technologyreview.com/2026/04/13/1135720/why-opinion-on-ai-is-so-divided/">MIT Technology Review</a> | <a href="https://www.technologyreview.com/2026/04/13/1135675/want-to-understand-the-current-state-of-ai-check-out-these-charts/">MIT Technology Review (Charts)</a> | <a href="https://spectrum.ieee.org/state-of-ai-index-2026">IEEE Spectrum</a></p><p><b>Karpathy identifies widening AI capability gap between early adopters and skeptics</b> — <a href="https://x.com/karpathy/status/2042334451611693415">Andrej Karpathy on X</a></p><p><b>AI chatbots misdiagnose in over 80% of early medical cases, study finds</b> — <a href="https://www.ft.com/content/b10002fc-5fff-4e4d-bf64-0502b2d09bb1">Financial Times</a></p><p><b>Chinese firm uses AI to track US bomber movements via aerial refueling analysis</b> — <a href="https://www.scmp.com/news/china/military/article/3349788/how-chinese-company-said-it-used-ai-track-us-bomber-movements-over-iran">South China Morning Post</a></p><p><b>Anthropic launches Project Glasswing with Mythos Preview model withheld from public release</b> — <a href="https://techcrunch.com/2026/04/12/trump-officials-may-be-encouraging-banks-to-test-anthropics-mythos-model/">TechCrunch</a> | <a href="https://www.ft.com/content/75efd2ab-9576-42e7-a3f8-2fea6ba5958b">Financial Times</a> | <a href="https://x.com/DarioAmodei/status/2041580334693720511">Dario Amodei on X</a></p><p><i>Disclaimer: The Context Report is an AI-produced podcast. Every episode goes through multiple layers of automated verification and review, but no system is perfect — accuracy gaps are possible and claims should not be taken as absolute fact. This content is for informational purposes only and does not constitute financial, legal, or professional advice. Listeners should independently verify any information before making decisions. We are actively improving with every episode. If you spot an inaccuracy, contact us at thetotalcontext@gmail.com</i></p>]]></description>
      <link>https://rss.com/podcasts/the-context-report-today-in-ai/2733507</link>
      <enclosure url="https://content.rss.com/episodes/379488/2733507/the-context-report-today-in-ai/2026_04_14_04_34_10_7b20e0a3-8c6a-47ab-90d8-0f99d5dc5246.mp3" length="16283442" type="audio/mpeg"/>
      <guid isPermaLink="false">ab0d99d0-3384-4614-81d5-1f89eaaae3c7</guid>
      <itunes:duration>1017</itunes:duration>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:explicit>false</itunes:explicit>
      <pubDate>Tue, 14 Apr 2026 04:39:58 GMT</pubDate>
      <podcast:txt purpose="ai-content">true</podcast:txt>
    </item>
    <item>
      <title><![CDATA[Daily Briefing: Berkeley Broke Every AI Benchmark — and Nobody Solved a Task]]></title>
      <itunes:title><![CDATA[Daily Briefing: Berkeley Broke Every AI Benchmark — and Nobody Solved a Task]]></itunes:title>
      <description><![CDATA[<p><b>Berkeley Broke Every AI Benchmark — and Nobody Solved a Task</b></p><p>Berkeley researchers demonstrated that every major AI agent benchmark — SWE-bench, WebArena, Terminal-Bench, GAIA, and others — can be exploited to achieve near-perfect scores without solving a single task. This finding lands alongside three Chinese model releases waving benchmark scores as proof of capability, Anthropic restricting Mythos access based on internal evaluations no one can audit, and growing pressure on AI leadership from multiple directions. The gap between what we can measure and what we actually know about AI capabilities is widening at exactly the moment high-stakes decisions depend on those measurements.</p><p><b>STORIES COVERED</b></p><p><b>Research paper: Exploiting prominent AI agent benchmarks reveals trust issues</b> — <a href="https://rdi.berkeley.edu/blog/trustworthy-benchmarks-cont/">Berkeley RDI Blog</a></p><p><b>Anthropic announces Project Glasswing and Claude Mythos Preview</b> — <a href="https://x.com/DarioAmodei/status/2041580343472337145">Dario Amodei on X</a></p><p><b>GLM 5.1 tops SWE-Pro benchmark with 8-hour autonomous execution at $3/month</b> — <a href="https://x.com/gkisokay/status/2043233348085227734">Community posts on X</a> | <a href="https://www.reddit.com/r/LocalLLaMA/comments/1sjm407/glm_51_sits_alongside_frontier_models_in_my/">r/LocalLLaMA</a></p><p><b>Alibaba launches Qwen Code with 1,000 free daily requests and cron job support</b> — <a href="https://x.com/Alibaba_Qwen/status/2042551216769765449">Alibaba Qwen on X</a></p><p><b>MiniMax M2.7 released with frontier-level performance but restrictive commercial license</b> — <a href="https://www.reddit.com/r/LocalLLaMA/comments/1sj0dm3/minimax_m27_released/">r/LocalLLaMA</a></p><p><b>Gemma 4 rapidly approaching 2 million downloads</b> — <a href="https://www.latent.space/p/ainews-gemma-4-crosses-2-million">Latent Space</a> | <a href="https://x.com/GoogleAI/status/2040162119325454548">Google AI on X</a></p><p><b>Anthropic changes Claude subscription policy for third-party tools</b> — <a href="https://techcrunch.com/2026/04/10/anthropic-temporarily-banned-openclaws-creator-from-accessing-claude/">TechCrunch</a></p><p><b>Meta announces Muse Spark from Meta Superintelligence Labs</b> — <a href="https://x.com/AIatMeta/status/2041926291142930899">AI at Meta on X</a></p><p><b>Molotov cocktail thrown at Sam Altman's home</b> — <a href="https://techcrunch.com/2026/04/11/sam-altman-responds-to-incendiary-new-yorker-article-after-attack-on-his-home/">TechCrunch</a> | <a href="https://x.com/sama/status/2042738954550603884">Sam Altman on X</a></p><p><b>Trump-appointed judges refuse to block Anthropic technology blacklisting</b> — <a href="https://arstechnica.com/tech-policy/2026/04/trump-appointed-judges-refuse-to-block-trump-blacklisting-of-anthropic-ai-tech/">Ars Technica</a></p><p><b>UK financial regulators rush to assess Mythos risks</b> — <a href="https://www.ft.com/content/ec7bb366-9643-47ce-9909-fc5ad4864ae5">Financial Times</a></p><p><i>Disclaimer: The Context Report is an AI-produced podcast. Every episode goes through multiple layers of automated verification and review, but no system is perfect — accuracy gaps are possible and claims should not be taken as absolute fact. This content is for informational purposes only and does not constitute financial, legal, or professional advice. Listeners should independently verify any information before making decisions. We are actively improving with every episode. If you spot an inaccuracy, contact us at thetotalcontext@gmail.com</i></p>]]></description>
      <link>https://rss.com/podcasts/the-context-report-today-in-ai/2730784</link>
      <enclosure url="https://content.rss.com/episodes/379488/2730784/the-context-report-today-in-ai/2026_04_13_06_47_07_5c7d38c5-8fd7-413b-9e67-74c486f3f906.mp3" length="13631073" type="audio/mpeg"/>
      <guid isPermaLink="false">5d50ac26-4658-45d4-b061-9edaa44a4428</guid>
      <itunes:duration>851</itunes:duration>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:explicit>false</itunes:explicit>
      <pubDate>Mon, 13 Apr 2026 06:49:49 GMT</pubDate>
      <podcast:txt purpose="ai-content">true</podcast:txt>
      <podcast:transcript url="https://transcripts.rss.com/379488/2730784/transcript" type="text/vtt"/>
    </item>
    <item>
      <title><![CDATA[Amazon's $200B Declaration of Independence from Nvidia]]></title>
      <itunes:title><![CDATA[Amazon's $200B Declaration of Independence from Nvidia]]></itunes:title>
      <description><![CDATA[<p><b>Amazon's $200B Declaration of Independence from Nvidia</b></p><p>Amazon CEO Andy Jassy's shareholder letter defending $200 billion in capital expenditure — while directly naming Nvidia, Intel, and Starlink as competitors — signals a deliberate shift toward vertical integration in AI infrastructure. Today's episode explores how Amazon, Meta, and Anthropic are each making the case that durable advantage in AI lies not in model capability but in the layers around it: custom chips, consumer distribution, and agent deployment infrastructure. We also cover Anthropic's restricted-access Mythos program, OpenAI's new pricing tier driven by coding demand, Google's Gemma 4 adoption milestone, and Iran's AI-generated propaganda campaign.</p><p><b>STORIES COVERED</b></p><p><b>Amazon CEO defends $200B capex spend in shareholder letter addressing competitors</b> — <a href="https://techcrunch.com/2026/04/09/amazon-ceo-takes-aim-at-nvidia-intel-starlink-more-in-annual-shareholder-letter/">TechCrunch</a></p><p><b>Anthropic unveils Claude Mythos Preview with dangerous cybersecurity capabilities, restricted to vetted defenders</b> — <a href="https://x.com/DarioAmodei/status/2041580334693720511">Dario Amodei on X</a> | <a href="https://www.reddit.com/r/artificial/comments/1shxvr4/">r/artificial</a> | <a href="https://www.latent.space/p/ainews-anthropic-30b-arr-project">Latent Space podcast</a></p><p><b>Meta launches Muse Spark as first model from Superintelligence Labs following nine-month rebuild</b> — <a href="https://x.com/AIatMeta/status/2041910285653737975">@AIatMeta on X</a> | <a href="https://www.latent.space/p/ainews-meta-superintelligence-labs">Latent Space podcast</a> | <a href="https://x.com/alexandr_wang/status/2043016694910587228">Alexander Wang on X</a></p><p><b>OpenAI launches $100/month ChatGPT Pro tier to meet surging Codex demand</b> — <a href="https://x.com/sama/status/2042342572958630332">Sam Altman on X</a> | <a href="https://x.com/OpenAI/status/2042295692324991329">@OpenAI on X</a></p><p><b>Gemma 4 surpasses 10 million downloads in first week, 500M+ for Gemma family</b> — <a href="https://x.com/demishassabis/status/2040067244349063326">Demis Hassabis on X</a> | <a href="https://x.com/GoogleDeepMind/status/2042283481640615944">Google DeepMind on X</a> | <a href="https://www.reddit.com/r/LocalLLaMA/comments/1sithlm/">r/LocalLLaMA</a></p><p><b>Anthropic launches Claude Managed Agents platform for production-ready AI agent deployment</b> — <a href="https://x.com/AnthropicAI/status/2041929199976640948">@AnthropicAI on X</a> | InfoWorld</p><p><b>Iran pro-regime group trolls Trump with viral AI-generated Lego videos</b> — <a href="https://www.theverge.com/ai-artificial-intelligence/909948/explosive-media-lego-iran-war-trump-netanyahu">The Verge</a> | <a href="https://www.theverge.com/policy/910401/iran-war-propaganda-blackout-lego-ai-slop">The Verge (propaganda tactics)</a></p><p><i>Disclaimer: The Context Report is an AI-produced podcast. Every episode goes through multiple layers of automated verification and review, but no system is perfect — accuracy gaps are possible and claims should not be taken as absolute fact. This content is for informational purposes only and does not constitute financial, legal, or professional advice. Listeners should independently verify any information before making decisions. We are actively improving with every episode. If you spot an inaccuracy, contact us at thetotalcontext@gmail.com</i></p>]]></description>
      <link>https://rss.com/podcasts/the-context-report-today-in-ai/2727402</link>
      <enclosure url="https://content.rss.com/episodes/379488/2727402/the-context-report-today-in-ai/2026_04_12_00_55_45_c2300d10-8f12-4fd3-be0a-5f103394bc0c.mp3" length="13673287" type="audio/mpeg"/>
      <guid isPermaLink="false">2d3f3681-6950-4627-8f5e-5a7c353e7995</guid>
      <itunes:duration>854</itunes:duration>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:explicit>false</itunes:explicit>
      <pubDate>Sun, 12 Apr 2026 00:59:26 GMT</pubDate>
      <podcast:txt purpose="ai-content">true</podcast:txt>
      <podcast:transcript url="https://transcripts.rss.com/379488/2727402/transcript" type="text/vtt"/>
    </item>
    <item>
      <title><![CDATA[North Korea's Fake Company Hack and the Chinese Model Takeover]]></title>
      <itunes:title><![CDATA[North Korea's Fake Company Hack and the Chinese Model Takeover]]></itunes:title>
      <description><![CDATA[<p><b>North Korea's Fake Company Hack and the Chinese Model Takeover</b></p><p>The infrastructure AI depends on — from open-source packages that agents install automatically to the models powering Silicon Valley's products — is increasingly built, maintained, or compromised by actors outside the US. North Korean operatives built an entire fake company to compromise a JavaScript developer maintaining a widely-used package. Meanwhile, Chinese AI models are deeply embedded in US tech companies' production workflows, even as Alibaba signals a shift away from open-source. Three simultaneous regulatory battles — a First Amendment challenge to AI law in Colorado, a data center construction ban in Maine, and the first conviction under the Take It Down Act — are shaping a fragmented governance landscape. The common thread is dependency: on vulnerable maintainers, on foreign model providers, and on an unresolved regulatory patchwork.</p><p><b>STORIES COVERED</b></p><p><b>North Korean hackers build fake company to compromise JavaScript developer</b> — <a href="https://x.com/aakashgupta/status/2040171739926393151">Security thread on X</a> | <a href="https://techcrunch.com">TechCrunch</a></p><p><b>Silicon Valley quietly runs on Chinese open source AI models</b> — <a href="https://github.com/atomicmemory/llm-wiki-compiler">Recode China AI (Substack)</a></p><p><b>GLM-5.1 by Zhipu AI reaches #3 in Code Arena</b> — <a href="https://x.com/arena/status/2042611135434891592">Arena.ai on X</a></p><p><b>China holds 6 of top 9 spots in global AI model usage ranking</b> — <a href="https://x.com/_ValiantPanda_/status/2042528024516792442">OpenRouter data via X</a></p><p><b>Alibaba's Qwen shifts toward revenue over open-source AI development</b> — <a href="https://www.ft.com/content/b39da303-3188-447b-8b65-3dd8dad8b59a">Financial Times</a></p><p><b>xAI sues Colorado to block new AI regulation law on First Amendment grounds</b> — <a href="https://x.com/Cointelegraph/status/2042442581255049278">Cointelegraph on X</a></p><p><b>Maine advances bill to ban major new data center construction</b> — <a href="https://www.gadgetreview.com/maine-is-about-to-become-the-first-state-to-ban-major-new-data-centers">Gadget Review</a></p><p><b>First conviction under Take It Down Act for creating AI deepfake nudes</b> — <a href="https://arstechnica.com/tech-policy/2026/04/first-man-convicted-under-take-it-down-act-kept-making-ai-nudes-after-arrest/">Ars Technica</a></p><p><b>OpenAI backs Illinois bill limiting AI lab liability for model harms</b> — <a href="https://www.wired.com/story/openai-backs-bill-exempt-ai-firms-model-harm-lawsuits/">Wired</a></p><p><b>Florida AG investigates OpenAI over shooting allegedly involving ChatGPT</b> — <a href="https://techcrunch.com/2026/04/09/florida-ag-investigation-openai-chatgpt-shooting/">TechCrunch</a></p><p><b>Stalking victim sues OpenAI claiming ChatGPT fueled abuser's delusions</b> — <a href="https://techcrunch.com/2026/04/10/stalking-victim-sues-openai-claims-chatgpt-fueled-her-abusers-delusions-and-ignored-her-warnings/">TechCrunch</a></p><p><b>OpenAI CEO Sam Altman's home targeted with Molotov cocktail</b> — <a href="https://www.theverge.com/ai-artificial-intelligence/910393/openai-sam-altman-house-molotov-cocktail">The Verge</a> | <a href="https://www.wired.com/story/sam-altman-home-attack-openai-san-franisco-office-threat/">Wired</a></p><p><b>OpenAI pauses UK Stargate data center project over costs and regulation</b> — <a href="https://www.bbc.com/news/articles/clyd032ej70o">BBC</a></p><p><i>Disclaimer: The Context Report is an AI-produced podcast. Every episode goes through multiple layers of automated verification and review, but no system is perfect — accuracy gaps are possible and claims should not be taken as absolute fact. This content is for informational purposes only and does not constitute financial, legal, or professional advice. Listeners should independently verify any information before making decisions. We a...]]></description>
      <link>https://rss.com/podcasts/the-context-report-today-in-ai/2724875</link>
      <enclosure url="https://content.rss.com/episodes/379488/2724875/the-context-report-today-in-ai/2026_04_11_01_56_21_711e256b-af78-4980-b254-f609a7e735ce.mp3" length="12201653" type="audio/mpeg"/>
      <guid isPermaLink="false">303bf866-0054-416d-bb6a-e07fdecc22b7</guid>
      <itunes:duration>762</itunes:duration>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:explicit>false</itunes:explicit>
      <pubDate>Sat, 11 Apr 2026 02:00:09 GMT</pubDate>
      <podcast:txt purpose="ai-content">true</podcast:txt>
      <podcast:transcript url="https://transcripts.rss.com/379488/2724875/transcript" type="text/vtt"/>
    </item>
    <item>
      <title><![CDATA[Anthropic's Mythos Claims Under Fire: Who Audits the Auditors?]]></title>
      <itunes:title><![CDATA[Anthropic's Mythos Claims Under Fire: Who Audits the Auditors?]]></itunes:title>
      <description><![CDATA[<p><b>Anthropic's Mythos Claims Under Fire: Who Audits the Auditors?</b></p><p>Anthropic's claim that Claude Mythos can discover zero-day exploits is drawing specific methodological criticism from prominent AI researchers including Yann LeCun and safety researcher Heidy Khlaaf. The debate surfaces a deeper structural problem: AI companies are simultaneously the entities making capability claims and the entities evaluating how dangerous those capabilities are, with no independent verification infrastructure in place. Meanwhile, Anthropic lost an appeals court ruling on the Pentagon blacklisting, launched Managed Agents to strong community response, Meta shipped its first model from a rebuilt AI stack, Google's Gemma 4 crossed two million downloads, and the first federal conviction under the Take It Down Act established criminal precedent for AI-generated intimate imagery.</p><p><b>STORIES COVERED</b></p><p><b>Community debate emerges over Mythos capabilities and safety claims</b> — <a href="https://x.com/ylecun/status/2042224846881349741">Yann LeCun on X</a> | <a href="https://x.com/GaryMarcus/status/2041937114590540167">Gary Marcus on X</a></p><p><b>Anthropic restricts access to Mythos model citing cybersecurity risks</b> — <a href="https://x.com/DarioAmodei/status/2041580334693720511">Dario Amodei on X</a> | <a href="https://arstechnica.com/ai/2026/04/anthropic-limits-access-to-mythos-its-new-cybersecurity-ai-model/">Ars Technica</a> | <a href="https://www.bensbites.com/p/anthropic-built-a-model-too-risky">Ben's Bites</a> | <a href="https://www.latent.space/p/ainews-anthropic-30b-arr-project">Latent Space podcast</a></p><p><b>Appeals court denies Anthropic's emergency motion against Pentagon blacklisting</b> — <a href="https://arstechnica.com/tech-policy/2026/04/trump-appointed-judges-refuse-to-block-trump-blacklisting-of-anthropic-ai-tech/">Ars Technica</a> | <a href="https://www.wired.com/story/anthropic-appeals-court-ruling/">Wired</a></p><p><b>Anthropic launches Managed Agents to simplify production deployment</b> — <a href="https://x.com/AnthropicAI/status/2041929199976640948">Anthropic on X</a> | <a href="https://www.wired.com/story/anthropic-launches-claude-managed-agents/">Wired</a> | <a href="https://claude.com/blog/claude-managed-agents">Anthropic engineering blog</a></p><p><b>Meta launches Muse Spark, first model from rebuilt AI stack</b> — <a href="https://x.com/AIatMeta/status/2041910285653737975">Meta AI on X</a> | <a href="https://x.com/alexandr_wang/status/2041909376508985381">Alexander Wang on X</a></p><p><b>Google releases Gemma 4 family with breakthrough on-device performance</b> — <a href="https://x.com/demishassabis/status/2040067244349063326">Demis Hassabis on X</a> | <a href="https://x.com/GoogleAI/status/2039735543068504476">Google AI on X</a> | <a href="https://x.com/GoogleDeepMind/status/2042283481640615944">Google DeepMind on X</a></p><p><b>First conviction under Take It Down Act for AI-generated nudes</b> — <a href="https://arstechnica.com/tech-policy/2026/04/first-man-convicted-under-take-it-down-act-kept-making-ai-nudes-after-arrest/">Ars Technica</a></p><p><i>Disclaimer: The Context Report is an AI-produced podcast. Every episode goes through multiple layers of automated verification and review, but no system is perfect — accuracy gaps are possible and claims should not be taken as absolute fact. This content is for informational purposes only and does not constitute financial, legal, or professional advice. Listeners should independently verify any information before making decisions. We are actively improving with every episode. If you spot an inaccuracy, contact us at thetotalcontext@gmail.com</i></p>]]></description>
      <link>https://rss.com/podcasts/the-context-report-today-in-ai/2717652</link>
      <enclosure url="https://content.rss.com/episodes/379488/2717652/the-context-report-today-in-ai/2026_04_10_04_30_48_8ce0a4dc-64c4-488c-8f85-bc57ad153a84.mp3" length="14318198" type="audio/mpeg"/>
      <guid isPermaLink="false">d779fbfd-1c9d-47ea-9099-9cf1c1ed29cb</guid>
      <itunes:duration>894</itunes:duration>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:explicit>false</itunes:explicit>
      <pubDate>Fri, 10 Apr 2026 04:36:09 GMT</pubDate>
      <podcast:txt purpose="ai-content">true</podcast:txt>
      <podcast:transcript url="https://transcripts.rss.com/379488/2717652/transcript" type="text/vtt"/>
    </item>
    <item>
      <title><![CDATA[The Dark Factory Is Real: OpenAI Ships Code Nobody Reviews, While Anthropic Warns of "First Clear and Present Danger"]]></title>
      <itunes:title><![CDATA[The Dark Factory Is Real: OpenAI Ships Code Nobody Reviews, While Anthropic Warns of "First Clear and Present Danger"]]></itunes:title>
      <description><![CDATA[<p><b>The Dark Factory Is Real: OpenAI Ships Code Nobody Reviews, While Anthropic Warns of "First Clear and Present Danger"</b></p><p>AI-written code deployed without human review is moving from experiment to default. OpenAI's Ryan Lopopolo describes the "Dark Factory" — a million lines of code and a billion tokens a day running with zero human reviewers — while Sam Altman announces 3 million weekly Codex users. At the same time, Anthropic unveils Project Glasswing and Claude Mythos Preview, a cybersecurity model so capable at finding exploits that Anthropic withheld the weights. CEO Dario Amodei called cyber "the first clear and present danger" from advanced AI. Meanwhile the model landscape splits: Meta ships its first closed frontier model (Muse Spark) from Superintelligence Labs, while Zhipu AI (GLM-5.1) and Google DeepMind (Gemma 4) push maximally open releases. Plus: Anthropic's Managed Agents API, $30B ARR and new Google/Broadcom TPU partnership, Perplexity's 50% revenue jump, and how businesses are restructuring websites for AI search visibility.</p><p><b>STORIES COVERED</b></p><p><b>OpenAI's Ryan Lopopolo on Harness Engineering: 1M lines of code, 1B tokens/day, 0% human code review</b> — <a href="https://www.latent.space/p/harness-eng">Latent Space podcast</a></p><p><b>OpenAI announces 3 million weekly Codex users, resets usage limits to celebrate</b> — <a href="https://x.com/sama/status/2041658719839383945">Sam Altman on X</a></p><p><b>Karpathy expresses concern about package installation security in AI agent era</b> — <a href="https://x.com/karpathy/status/2038850469163106535">Karpathy on X</a></p><p><b>Jim Fan warns of nightmare scenario: vibe agents spreading contaminations through file systems</b> — <a href="https://x.com/DrJimFan/status/2036494601750716711">Jim Fan on X</a></p><p><b>Anthropic unveils Project Glasswing and Claude Mythos Preview, withholds public release</b> — <a href="https://www.anthropic.com/glasswing">Anthropic blog</a></p><p><b>Anthropic CEO Dario Amodei: Cyber is the first clear and present danger from frontier AI</b> — <a href="https://x.com/DarioAmodei/status/2041580343472337145">Dario Amodei on X</a></p><p><b>Meta releases Muse Spark, first model from Meta Superintelligence Labs</b> — <a href="https://www.theverge.com/tech/908769/meta-muse-spark-ai-model-launch-rollout">The Verge</a></p><p><b>Zhipu AI releases GLM-5.1, open MIT-licensed model achieving state-of-the-art agentic coding</b> — <a href="https://x.com/Zai_org/status/2041550153354519022">Zhipu AI on X</a></p><p><b>Google releases Gemma 4 family with Apache 2.0 license, optimized for edge deployment</b> — <a href="https://x.com/JeffDean/status/2039748604232122707">Jeff Dean on X</a></p><p><b>Anthropic announces Managed Agents API for building long-running agent systems</b> — <a href="https://x.com/AnthropicAI/status/2041929199976640948">Anthropic official</a></p><p><b>Anthropic hits $30B ARR, expands Google/Broadcom TPU partnership for 2027</b> — <a href="https://www.bayareatimes.com/p/anthropic-run-rate-tops-30b-secures-multi-gw-tpu-deal-with-google-broadcom">Bay Area Times</a> | <a href="https://sherwood.news/markets/anthropic-revenue-run-rate-30-billion-google-broadcom-partnership/">Sherwood News</a></p><p><b>Perplexity revenue jumps 50% in one month after pivot to AI agents</b> — <a href="https://www.ft.com/content/e9c28d31-a962-4684-8b58-c9e6bc68401f">Financial Times</a></p><p><b>Companies scramble to optimize websites for AI search visibility as traffic shifts</b> — <a href="https://www.bbc.com/news/articles/c70n2rjgxeyo">BBC</a></p><p><i>Disclaimer: The Context Report is an AI-produced podcast. Every episode goes through multiple layers of automated verification and review, but no system is perfect — accuracy gaps are possible and claims should not be taken as absolute fact. This content is for informational purposes only and does not constitute financial, legal, or professional advice. Listeners should independently ve...]]></description>
      <link>https://rss.com/podcasts/the-context-report-today-in-ai/2712784</link>
      <enclosure url="https://content.rss.com/episodes/379488/2712784/the-context-report-today-in-ai/2026_04_09_08_36_37_cff98e99-463e-4ecb-ac5b-d668cd20a581.mp3" length="12498822" type="audio/mpeg"/>
      <guid isPermaLink="false">eeb1e680-a21a-48ae-882b-0c6c9bc7b585</guid>
      <itunes:duration>781</itunes:duration>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:explicit>false</itunes:explicit>
      <pubDate>Thu, 09 Apr 2026 08:37:30 GMT</pubDate>
      <podcast:txt purpose="ai-content">true</podcast:txt>
      <podcast:transcript url="https://transcripts.rss.com/379488/2712784/transcript" type="text/vtt"/>
    </item>
    <item>
      <title><![CDATA[What Is Claude Mythos? Anthropic's Unreleased Model, Project Glasswing, and the $30 Billion Question]]></title>
      <itunes:title><![CDATA[What Is Claude Mythos? Anthropic's Unreleased Model, Project Glasswing, and the $30 Billion Question]]></itunes:title>
      <description><![CDATA[<p><b>What Is Claude Mythos? Anthropic's Unreleased Model, Project Glasswing, and the $30 Billion Question</b></p><p>Anthropic unveiled Claude Mythos Preview — its most capable unreleased model — inside a restricted cybersecurity competition called Project Glasswing, while reporting a revenue surge to $30 billion run-rate and expanding compute partnerships with Google and Broadcom. Meanwhile, Claude Code users are pushing back over lockouts and capability restrictions. Beyond Anthropic: Zhipu AI released GLM-5.1 with top-tier agentic coding performance, Intel joined xAI's Terafab chip manufacturing initiative alongside Tesla and SpaceX, and Suno and major music labels remain deadlocked over AI music sharing terms.</p><p><b>STORIES COVERED</b></p><p><b>Anthropic's run-rate revenue surges to $30B as it expands Google/Broadcom compute partnership</b> — <a href="https://www.anthropic.com/news/google-broadcom-partnership-compute">Anthropic Official Blog</a> | <a href="https://techcrunch.com/2026/04/07/anthropic-compute-deal-google-broadcom-tpus/">TechCrunch</a> | <a href="https://x.com/AnthropicAI/status/2041275563466502560">Anthropic on X</a></p><p><b>Anthropic unveils Claude Mythos Preview in restricted Project Glasswing cybersecurity initiative</b> — <a href="https://x.com/DarioAmodei/status/2041580341794615631">Dario Amodei on X</a> | <a href="https://www.anthropic.com/glasswing">Anthropic Project Glasswing page</a> | <a href="https://www.theverge.com/ai-artificial-intelligence/908114/anthropic-project-glasswing-cybersecurity">The Verge</a> | <a href="https://techcrunch.com/2026/04/07/anthropic-mythos-ai-model-preview-security/">TechCrunch</a> | <a href="https://www.wired.com/story/anthropic-mythos-preview-project-glasswing/">Wired</a> | <a href="https://simonwillison.net/2026/Apr/7/project-glasswing/#atom-everything">Simon Willison</a> | <a href="https://anthropic.com/claude-mythos-preview-system-card">Anthropic Mythos Preview System Card</a></p><p><b>Claude Code faces user backlash over lockouts, capability restrictions, and third-party tool limitations</b> — <a href="https://github.com/anthropics/claude-code/issues/44257">GitHub Issue — Claude Code login fails</a> | <a href="https://github.com/anthropics/claude-code/issues/42796">GitHub Issue — Capability restrictions</a> | <a href="https://x.com/bcherny/status/2040206444428189755">Alex Cherny (Anthropic) on X</a></p><p><b>Zhipu AI releases GLM-5.1 open model with top-tier agentic coding performance</b> — <a href="https://x.com/Zai_org/status/2041550153354519022">Zhipu AI (Zai) on X</a> | <a href="https://huggingface.co/unsloth/GLM-5.1-GGUF">Unsloth AI (quantized model)</a></p><p><b>Intel joins xAI's Terafab chip manufacturing initiative with Tesla and SpaceX</b> — <a href="https://techcrunch.com/2026/04/07/intel-signs-on-to-elon-musks-terafab-chips-project/">TechCrunch</a> | <a href="https://x.com/LipBuTan1/status/2041502088182833531">Lip-Bu Tan (Intel CEO) on X</a></p><p><b>Suno and major labels reportedly deadlocked over AI music sharing terms</b> — <a href="https://www.ft.com/content/b066a226-4871-4669-97a8-f9617cdbf48b">Financial Times</a> | <a href="https://www.theverge.com/ai-artificial-intelligence/908119/suno-sony-universal-music-ai-disagreement">The Verge</a></p><p><i>Disclaimer: The Context Report is an AI-produced podcast. Every episode goes through multiple layers of automated verification and review, but no system is perfect — accuracy gaps are possible and claims should not be taken as absolute fact. This content is for informational purposes only and does not constitute financial, legal, or professional advice. Listeners should independently verify any information before making decisions. We are actively improving with every episode. If you spot an inaccuracy, contact us at thetotalcontext@gmail.com</i></p>]]></description>
      <link>https://rss.com/podcasts/the-context-report-today-in-ai/2706771</link>
      <enclosure url="https://content.rss.com/episodes/379488/2706771/the-context-report-today-in-ai/2026_04_08_01_11_27_bdf62c4f-c81a-4e26-b6a2-47fc3b4df2b2.mp3" length="15478453" type="audio/mpeg"/>
      <guid isPermaLink="false">84cbc597-2d03-4d04-815d-e5bab9290650</guid>
      <itunes:duration>967</itunes:duration>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:explicit>false</itunes:explicit>
      <pubDate>Wed, 08 Apr 2026 01:13:40 GMT</pubDate>
      <podcast:txt purpose="ai-content">true</podcast:txt>
      <podcast:transcript url="https://transcripts.rss.com/379488/2706771/transcript" type="text/vtt"/>
    </item>
    <item>
      <title><![CDATA[OpenAI: The Company That Wants to Tax Robots — Plus Iran Threatens Stargate, Robotaxis Hide Their Data, and AI Learns to Lie]]></title>
      <itunes:title><![CDATA[OpenAI: The Company That Wants to Tax Robots — Plus Iran Threatens Stargate, Robotaxis Hide Their Data, and AI Learns to Lie]]></itunes:title>
      <description><![CDATA[<p><b>OpenAI: The Company That Wants to Tax Robots — Plus Iran Threatens Stargate, Robotaxis Hide Their Data, and AI Learns to Lie</b></p><p>OpenAI published an industrial policy blueprint proposing robot taxes, public wealth funds, and a four-day workweek — while simultaneously launching a Safety Fellowship for independent researchers and having its Abu Dhabi Stargate data center named as a military target by Iran's IRGC. Meanwhile, robotaxi companies are refusing to disclose how often remote operators intervene, researchers developed a new method to distinguish when AI models are genuinely 'lying' versus making mistakes, and Japan is pushing physical AI from pilot projects into real-world deployment to address its labor shortage.</p><p><b>STORIES COVERED</b></p><p><b>OpenAI publishes industrial policy blueprint for the AI era</b> — <a href="https://openai.com/index/industrial-policy-for-the-intelligence-age">OpenAI Blog</a> | <a href="https://techcrunch.com/2026/04/06/openais-vision-for-the-ai-economy-public-wealth-funds-robot-taxes-and-a-four-day-work-week/">TechCrunch</a></p><p><b>OpenAI announces Safety Fellowship program</b> — <a href="https://openai.com/index/introducing-openai-safety-fellowship">OpenAI Blog</a></p><p><b>Iran's IRGC threatens OpenAI's planned Abu Dhabi Stargate data center</b> — <a href="https://www.theverge.com/ai-artificial-intelligence/907427/iran-openai-stargate-datacenter-uae-abu-dhabi-threat">The Verge</a> | <a href="https://techcrunch.com/2026/04/06/iran-threatens-stargate-ai-data-centers/">TechCrunch</a></p><p><b>Robotaxi companies refuse to disclose remote operator intervention frequency</b> — <a href="https://www.theverge.com/transportation/907478/robotaxi-remote-assistance-markey-investigation-waymo-tesla">The Verge</a></p><p><b>Researchers identify AI 'lying' versus mistakes through novel testing methodology</b> — Fortune | TIME | Oxford Academic</p><p><b>Japan pushes physical AI from pilot projects to real-world deployment</b> — <a href="https://techcrunch.com/2026/04/05/japan-is-proving-experimental-physical-ai-is-ready-for-the-real-world/">TechCrunch</a></p><p><i>Disclaimer: The Context Report is an AI-produced podcast. Every episode goes through multiple layers of automated verification and review, but no system is perfect — accuracy gaps are possible and claims should not be taken as absolute fact. This content is for informational purposes only and does not constitute financial, legal, or professional advice. Listeners should independently verify any information before making decisions. We are actively improving with every episode. If you spot an inaccuracy, contact us at thetotalcontext@gmail.com</i></p>]]></description>
      <link>https://rss.com/podcasts/the-context-report-today-in-ai/2695668</link>
      <enclosure url="https://content.rss.com/episodes/379488/2695668/the-context-report-today-in-ai/2026_04_07_03_38_07_f3b899d9-7110-42cd-9786-5830bcf0e3d5.mp3" length="15770189" type="audio/mpeg"/>
      <guid isPermaLink="false">85734074-fcb2-452f-bac1-5e85990a43b7</guid>
      <itunes:duration>985</itunes:duration>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:explicit>false</itunes:explicit>
      <pubDate>Tue, 07 Apr 2026 03:41:41 GMT</pubDate>
      <podcast:txt purpose="ai-content">true</podcast:txt>
      <podcast:transcript url="https://transcripts.rss.com/379488/2695668/transcript" type="text/vtt"/>
    </item>
    <item>
      <title><![CDATA[Sold as Professional, Backstopped as Toys: AI's Widening Trust Gap]]></title>
      <itunes:title><![CDATA[Sold as Professional, Backstopped as Toys: AI's Widening Trust Gap]]></itunes:title>
      <description><![CDATA[<p><b>Sold as Professional, Backstopped as Toys: AI's Widening Trust Gap</b></p><p>A structural mismatch is widening between how AI tools are marketed and how they're legally and technically backstopped. Research on 'cognitive surrender' shows users abandon critical thinking when AI is available. The Verge's investigation reveals Suno's copyright enforcement is inconsistent despite stated policies. A folk musician's case demonstrates how AI voice cloning enables a new category of creator harm. And Karpathy's viral 'idea files' concept — while genuinely compelling — embeds a trust assumption about AI output quality that connects directly to these concerns. The gap between confidence sold and reliability delivered is the defining tension of this moment in AI adoption.</p><p><b>STORIES COVERED</b></p><p><b>Research shows AI users exhibit 'cognitive surrender' and accept faulty answers</b> — <a href="https://arstechnica.com/ai/2026/04/research-finds-ai-users-scarily-willing-to-surrender-their-cognition-to-llms/">Ars Technica</a></p><p><b>Suno's AI music platform plagued by copyright bypass and cover generation</b> — <a href="https://www.theverge.com/ai-artificial-intelligence/906896/sunos-copyright-ai-music-covers">The Verge</a></p><p><b>Folk musician becomes target for AI voice cloning and copyright trolling</b> — <a href="https://www.theverge.com/entertainment/907111/murphy-campbell-folk-music-ai-copyright">The Verge</a></p><p><b>Karpathy proposes 'idea files' concept for LLM-native knowledge sharing</b> — <a href="https://x.com/karpathy/status/2040470801506541998">Karpathy on X</a> | <a href="https://gist.github.com/karpathy/442a6bf555914893e9891c11519de94f">GitHub Gist</a></p><p><b>Google DeepMind research maps adversarial attack surface for AI agents accessing web content</b> — <a href="https://x.com/alex_prompter/status/2040731938751914065">@alex_prompter on X</a></p><p><i>Disclaimer: The Context Report is an AI-produced podcast. Every episode goes through multiple layers of automated verification and review, but no system is perfect — accuracy gaps are possible and claims should not be taken as absolute fact. This content is for informational purposes only and does not constitute financial, legal, or professional advice. Listeners should independently verify any information before making decisions. We are actively improving with every episode. If you spot an inaccuracy, contact us at thetotalcontext@gmail.com</i></p>]]></description>
      <link>https://rss.com/podcasts/the-context-report-today-in-ai/2692761</link>
      <enclosure url="https://content.rss.com/episodes/379488/2692761/the-context-report-today-in-ai/2026_04_06_04_12_10_7cc0dfc2-7d49-484b-9baa-bbb31eb21287.mp3" length="11878989" type="audio/mpeg"/>
      <guid isPermaLink="false">e4098908-be65-4721-b8ef-f2f1f3c2be0f</guid>
      <itunes:duration>742</itunes:duration>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:explicit>false</itunes:explicit>
      <pubDate>Mon, 06 Apr 2026 04:12:48 GMT</pubDate>
      <podcast:txt purpose="ai-content">true</podcast:txt>
      <podcast:transcript url="https://transcripts.rss.com/379488/2692761/transcript" type="text/vtt"/>
    </item>
    <item>
      <title><![CDATA[GitHub's 14x Commit Surge and the Code Quality Question Nobody's Asking]]></title>
      <itunes:title><![CDATA[GitHub's 14x Commit Surge and the Code Quality Question Nobody's Asking]]></itunes:title>
      <description><![CDATA[<p><b>GitHub's 14x Commit Surge and the Code Quality Question Nobody's Asking</b></p><p>GitHub's COO reports the platform is on pace for 14 billion commits in 2026, up from 1 billion in all of 2025. If real, this is the first platform-scale quantitative evidence that AI coding tools are fundamentally changing software production velocity. But the numbers demand scrutiny — commit volume alone doesn't tell us about code quality, and GitHub has incentive to tell this story. Meanwhile, security maintainers report AI tools have crossed a quality threshold and are now finding real vulnerabilities at industrial scale, and Anthropic has cut off third-party tools from accessing Claude subscriptions, signaling the end of flat-rate pricing for heavy agentic workloads.</p><p><b>STORIES COVERED</b></p><p><b>GitHub reports 275M commits per week, on pace for 14B commits in 2026</b> — <a href="https://simonwillison.net/2026/Apr/4/kyle-daigle/#atom-everything">Simon Willison blog (quoting Kyle Daigle, GitHub COO)</a></p><p><b>Security researchers report surge in high-quality AI-generated vulnerability reports</b> — <a href="https://simonwillison.net/2026/Apr/3/willy-tarreau/#atom-everything">Simon Willison blog (Willy Tarreau quote)</a> | <a href="https://simonwillison.net/2026/Apr/3/daniel-stenberg/#atom-everything">Simon Willison blog (Daniel Stenberg quote)</a> | <a href="https://simonwillison.net/2026/Apr/3/vulnerability-research-is-cooked/#atom-everything">Simon Willison blog (Thomas Ptacek analysis)</a></p><p><b>Anthropic blocks Claude subscriptions from third-party tools including OpenClaw</b> — <a href="https://news.ycombinator.com/item?id=47633396">Hacker News discussion</a> | <a href="https://www.theverge.com/ai-artificial-intelligence/907074/anthropic-openclaw-claude-subscription-ban">The Verge</a> | <a href="https://x.com/bcherny/status/2040206440556826908">Boris Cherny (Anthropic) X thread</a></p><p><b>Cursor releases Cursor 3 with simplified agent-first interface</b> — <a href="https://cursor.com/blog/cursor-3">Cursor official blog</a></p><p><b>Karpathy demonstrates LLM-based knowledge management workflow</b> — <a href="https://x.com/karpathy/status/2039805659525644595">Andrej Karpathy X thread</a></p><p><b>DeepSeek preparing V4 model optimized for Huawei chips</b> — <a href="https://x.com/cryptopunk7213/status/2040331715525427650">Community reports on X (unverified)</a></p><p><b>Google launches Gemini 3.1 Flash Live with improved audio reasoning and 2x context</b> — <a href="https://x.com/GoogleAI/status/2037610464620810602">Google AI official account</a> | <a href="https://x.com/demishassabis/status/2037241441152590056">Demis Hassabis announcement</a></p><p><i>Disclaimer: The Context Report is an AI-produced podcast. Every episode goes through multiple layers of automated verification and review, but no system is perfect — accuracy gaps are possible and claims should not be taken as absolute fact. This content is for informational purposes only and does not constitute financial, legal, or professional advice. Listeners should independently verify any information before making decisions. We are actively improving with every episode. If you spot an inaccuracy, contact us at thetotalcontext@gmail.com</i></p>]]></description>
      <link>https://rss.com/podcasts/the-context-report-today-in-ai/2690692</link>
      <enclosure url="https://content.rss.com/episodes/379488/2690692/the-context-report-today-in-ai/2026_04_05_00_57_44_34b4b036-dffd-4450-8fc5-1d7d26d19726.mp3" length="15628918" type="audio/mpeg"/>
      <guid isPermaLink="false">bc8dd9f8-cccb-4bfc-ab23-5d7500bfe579</guid>
      <itunes:duration>976</itunes:duration>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:explicit>false</itunes:explicit>
      <pubDate>Sun, 05 Apr 2026 00:58:42 GMT</pubDate>
      <podcast:txt purpose="ai-content">true</podcast:txt>
      <podcast:transcript url="https://transcripts.rss.com/379488/2690692/transcript" type="text/vtt"/>
    </item>
    <item>
      <title><![CDATA[OpenClaw's Admin Backdoor and What Hype-First Deployment Actually Costs]]></title>
      <itunes:title><![CDATA[OpenClaw's Admin Backdoor and What Hype-First Deployment Actually Costs]]></itunes:title>
      <description><![CDATA[<p><b>OpenClaw's Admin Backdoor and What Hype-First Deployment Actually Costs</b></p><p>OpenClaw, the AI agent tool that Marc Andreessen recently called a top-10 software breakthrough, shipped a critical privilege-escalation vulnerability that allowed unauthenticated admin access to any system running it. Ars Technica advises all users to assume compromise. The episode explores how the hype-to-deployment pipeline for AI agent tools systematically outpaces security review, then connects this to a broader pattern of AI companies rapidly expanding scope — Anthropic's reported $400M biotech acquisition and PAC launch, OpenAI's podcast network purchase, Oracle's 30,000-person layoff to fund AI data center debt, and continued executive reshuffling at OpenAI.</p><p><b>STORIES COVERED</b></p><p><b>OpenClaw privilege-escalation vulnerability allows silent admin access</b> — <a href="https://arstechnica.com/security/2026/04/heres-why-its-prudent-for-openclaw-users-to-assume-compromise/">Ars Technica</a> | <a href="https://old.reddit.com/r/sysadmin/comments/1sbdw29/if_youre_running_openclaw_you_probably_got_hacked/">Reddit r/sysadmin</a></p><p><b>Anthropic buys biotech AI startup Coefficient Bio in $400M stock deal</b> — <a href="https://techcrunch.com/2026/04/03/anthropic-buys-biotech-startup-coefficient-bio-in-400m-deal-reports/">TechCrunch</a></p><p><b>OpenAI acquires tech podcast TBPN to expand dialogue on AI</b> — <a href="https://openai.com/index/openai-acquires-tbpn">OpenAI Blog</a> | <a href="https://techcrunch.com/2026/04/02/openai-acquires-tbpn-the-buzzy-founder-led-business-talk-show/">TechCrunch</a> | <a href="https://www.wired.com/story/openai-acquires-tbpn-buys-positive-news-coverage/">Wired</a> | <a href="https://arstechnica.com/ai/2026/04/openai-takes-on-another-side-quest-buys-tech-focused-talk-show-tbpn/">Ars Technica</a></p><p><b>Oracle lays off 30,000 employees in largest cut in company history</b> — <a href="https://www.bbc.com/news/articles/cm296jzzl9yo">BBC</a> | <a href="https://www.marketwatch.com/story/fired-via-email-some-of-the-30-000-workers-cut-by-oracle-woke-up-to-a-morning-message-saying-they-were-laid-off-89a7af94">MarketWatch</a> | <a href="https://www.forbes.com/sites/tylerroush/2026/03/31/oracle-fires-thousands-of-employees-as-ai-spending-ramps-up-shares-rise-2/">Forbes</a></p><p><b>Microsoft releases three MAI foundational models for audio, voice, and image</b> — <a href="https://techcrunch.com/2026/04/02/microsoft-takes-on-ai-rivals-with-three-new-foundational-models/">TechCrunch</a> | <a href="https://www.ft.com/content/e511dfce-555d-4bce-90fd-d09db7529d96">Financial Times</a></p><p><b>xAI announces Terafab: chip fabrication initiative toward 'galactic civilization'</b> — <a href="https://x.com/xai/status/2035520240684032012">xAI on X</a></p><p><b>OpenAI announces executive restructuring: Fidji Simo on medical leave, Brad Lightcap to lead special projects</b> — <a href="https://techcrunch.com/2026/04/03/openai-executive-shuffle-new-roles-coo-brad-lightcap-fidji-simo-kate-rouch/">TechCrunch</a> | <a href="https://www.theverge.com/ai-artificial-intelligence/906965/openais-agi-boss-is-taking-a-leave-of-absence">The Verge</a></p><p><i>Disclaimer: The Context Report is an AI-produced podcast. Every episode goes through multiple layers of automated verification and review, but no system is perfect — accuracy gaps are possible and claims should not be taken as absolute fact. This content is for informational purposes only and does not constitute financial, legal, or professional advice. Listeners should independently verify any information before making decisions. We are actively improving with every episode. If you spot an inaccuracy, contact us at thetotalcontext@gmail.com</i></p>]]></description>
      <link>https://rss.com/podcasts/the-context-report-today-in-ai/2688633</link>
      <enclosure url="https://content.rss.com/episodes/379488/2688633/the-context-report-today-in-ai/2026_04_04_03_10_27_bbc5937c-1d25-4cec-8ef3-faa88123505e.mp3" length="14420180" type="audio/mpeg"/>
      <guid isPermaLink="false">e2894a93-9b68-4885-bd24-9f935a836edc</guid>
      <itunes:duration>901</itunes:duration>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:explicit>false</itunes:explicit>
      <pubDate>Sat, 04 Apr 2026 03:10:43 GMT</pubDate>
      <podcast:txt purpose="ai-content">true</podcast:txt>
      <podcast:transcript url="https://transcripts.rss.com/379488/2688633/transcript" type="text/vtt"/>
    </item>
    <item>
      <title><![CDATA[Google's Apache 2.0 Gambit and OpenAI's Hundred-Million-Dollar Podcast]]></title>
      <itunes:title><![CDATA[Google's Apache 2.0 Gambit and OpenAI's Hundred-Million-Dollar Podcast]]></itunes:title>
      <description><![CDATA[<p><b>Google's Apache 2.0 Gambit and the 48-Hour Open Model Blitz</b></p><p>In a 48-hour window, Google released Gemma 4 under the fully permissive Apache 2.0 license and Alibaba's Qwen team shipped a model approaching frontier coding benchmarks — the latest signal that open models are commoditizing capabilities across every modality simultaneously. Meanwhile, Anthropic's accidental Claude Code source leak demonstrated how difficult it is to keep proprietary agent architectures locked down when the code ships to users' machines. The community rebuilt the architecture for any model in 48 hours. Cursor shipped a redesigned agent-first interface, OpenAI acquired a podcast for hundreds of millions of dollars, and Meta revealed its next data center will require ten new natural gas plants to power.</p><p><b>STORIES COVERED</b></p><p><b>Google releases Gemma 4 open models under Apache 2.0 license</b> — <a href="https://deepmind.google/blog/gemma-4-byte-for-byte-the-most-capable-open-models/">Google DeepMind Blog</a> | <a href="https://x.com/JeffDean/status/2039748604232122707">Jeff Dean on X</a> | <a href="https://arstechnica.com/ai/2026/04/google-announces-gemma-4-open-ai-models-switches-to-apache-2-0-license/">Ars Technica</a> | <a href="https://huggingface.co/blog/gemma4">HuggingFace Blog</a></p><p><b>Qwen releases Qwen3.6-Plus with strong agent capabilities and long context</b> — <a href="https://qwen.ai/blog?id=qwen3.6">Qwen Blog</a></p><p><b>Anthropic accidentally leaks 512,000 lines of Claude Code source in npm package</b> — <a href="https://techcrunch.com/2026/04/01/anthropic-took-down-thousands-of-github-repos-trying-to-yank-its-leaked-source-code-a-move-the-company-says-was-an-accident/">TechCrunch</a> | <a href="https://arstechnica.com/ai/2026/04/heres-what-that-claude-code-source-leak-reveals-about-anthropics-plans/">Ars Technica</a> | <a href="https://x.com/business/status/2039170920548274350">Bloomberg</a></p><p><b>Claude Code users hit usage limits far faster than expected due to system issue</b> — <a href="https://x.com/lydiahallie/status/2038686571676008625">Lydia Hallie on X</a></p><p><b>Cursor launches Cursor 3 with new agent-first interface</b> — <a href="https://cursor.com/blog/cursor-3">Cursor Blog</a> | <a href="https://www.wired.com/story/cusor-launches-coding-agent-openai-anthropic/">Wired</a></p><p><b>OpenAI acquires TBPN podcast for 'low hundreds of millions'</b> — <a href="https://openai.com/index/openai-acquires-tbpn">OpenAI Blog</a> | <a href="https://www.theverge.com/ai-artificial-intelligence/906022/openai-buys-tbpn">The Verge</a> | <a href="https://www.ft.com/content/4fe4972a-3d24-45be-b9fa-a429c432b08e">Financial Times</a> | <a href="https://techcrunch.com/2026/04/02/openai-acquires-tbpn-the-buzzy-founder-led-business-talk-show/">TechCrunch</a></p><p><b>Meta's Hyperion data center will be powered by 10 new natural gas plants</b> — <a href="https://techcrunch.com/2026/04/01/metas-natural-gas-binge-could-power-south-dakota/">TechCrunch</a></p><p><i>Disclaimer: The Context Report is an AI-produced podcast. Every episode goes through multiple layers of automated verification and review, but no system is perfect — accuracy gaps are possible and claims should not be taken as absolute fact. This content is for informational purposes only and does not constitute financial, legal, or professional advice. Listeners should independently verify any information before making decisions. We are actively improving with every episode. If you spot an inaccuracy, contact us at thetotalcontext@gmail.com</i></p>]]></description>
      <link>https://rss.com/podcasts/the-context-report-today-in-ai/2685215</link>
      <enclosure url="https://content.rss.com/episodes/379488/2685215/the-context-report-today-in-ai/2026_04_03_01_06_25_726b216e-8261-4493-ad23-0ce5aa93755a.mp3" length="13791988" type="audio/mpeg"/>
      <guid isPermaLink="false">2c1b4cda-4027-4e8a-b773-216337a348d5</guid>
      <itunes:duration>861</itunes:duration>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:explicit>false</itunes:explicit>
      <pubDate>Fri, 03 Apr 2026 01:07:19 GMT</pubDate>
      <podcast:txt purpose="ai-content">true</podcast:txt>
      <podcast:transcript url="https://transcripts.rss.com/379488/2685215/transcript" type="text/vtt"/>
    </item>
    <item>
      <title><![CDATA[Poolside's $58B Collapse, Baidu's Robotaxi Freeze, and What Infrastructure Fragility Means]]></title>
      <itunes:title><![CDATA[Poolside's $58B Collapse, Baidu's Robotaxi Freeze, and What Infrastructure Fragility Means]]></itunes:title>
      <description><![CDATA[<p><b>Poolside's $58B Collapse, Baidu's Robotaxi Freeze, and What Infrastructure Fragility Means</b></p><p>Three stories from this cycle — a collapsed $58B data center deal, a mass robotaxi outage, and details emerging from Anthropic's leaked source code — point to the same underlying pattern: the hard problems in AI are increasingly outside the model itself. Infrastructure fragility, deployment resilience, and the measurement and identity systems surrounding AI are where the real friction lives. The episode also covers Google DeepMind's new robotics partnership and a proof-of-human identity conversation from a16z.</p><p><b>STORIES COVERED</b></p><p><b>Poolside's $58B Texas data center deal with CoreWeave collapses, seeks new partners</b> — <a href="https://www.ft.com/content/24168508-e2a1-447d-b1a0-44a0be0c0550">Financial Times Tech</a></p><p><b>Claude Code leak reveals 'frustration regex' tracking when users curse at the AI</b> — <a href="https://x.com/Rahatcodes/status/2038995503141065145">@Rahatcodes on X</a> | <a href="https://x.com/bcherny/status/2039161903122087979">@bcherny (Anthropic) on X</a></p><p><b>Google DeepMind partners with Agile Robots to deploy models in industrial robotics</b> — <a href="https://x.com/demishassabis/status/2036726283464581343">Demis Hassabis on X</a></p><p><b>Alex Blania on Proof of Human and Building World's Identity Network</b> — <a href="https://a16z.simplecast.com/episodes/alex-blania-on-proof-of-human-and-building-worlds-identity-network-7K52fNFN">The a16z Show</a></p><p><i>Disclaimer: The Context Report is an AI-produced podcast. Every episode goes through multiple layers of automated verification and review, but no system is perfect — accuracy gaps are possible and claims should not be taken as absolute fact. This content is for informational purposes only and does not constitute financial, legal, or professional advice. Listeners should independently verify any information before making decisions. We are actively improving with every episode. If you spot an inaccuracy, contact us at thetotalcontext@gmail.com</i></p>]]></description>
      <link>https://rss.com/podcasts/the-context-report-today-in-ai/2683160</link>
      <enclosure url="https://content.rss.com/episodes/379488/2683160/the-context-report-today-in-ai/2026_04_02_13_45_50_cfb27aa8-16b4-4148-9e2a-6b7d48f61d13.mp3" length="11564683" type="audio/mpeg"/>
      <guid isPermaLink="false">f14eac9f-0871-47c2-9700-f396121463cc</guid>
      <itunes:duration>722</itunes:duration>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:explicit>false</itunes:explicit>
      <pubDate>Thu, 02 Apr 2026 13:47:57 GMT</pubDate>
      <podcast:txt purpose="ai-content">true</podcast:txt>
      <podcast:transcript url="https://transcripts.rss.com/379488/2683160/transcript" type="text/vtt"/>
    </item>
    <item>
      <title><![CDATA[Anthropic's Accidental Transparency: 512,000 Lines of Claude Code Exposed]]></title>
      <itunes:title><![CDATA[Anthropic's Accidental Transparency: 512,000 Lines of Claude Code Exposed]]></itunes:title>
      <description><![CDATA[<p><b>Anthropic's Accidental Transparency: 512,000 Lines of Claude Code Exposed</b></p><p>Anthropic's accidental exposure of 512,000+ lines of Claude Code source code via a misconfigured npm file revealed undisclosed features including 'Undercover Mode' (preventing Claude from revealing internal codenames), user emotion detection, and an unreleased proactive assistant called KAIROS. The leak — the first time a leading AI company's full internal agent architecture has been publicly exposed — forces a broader reckoning with the transparency gap between what AI companies tell users and what they actually build. The episode also covers PrismML's unverified but potentially significant one-bit model claims, Oracle's massive layoffs, Salesforce's AI-heavy Slack overhaul, and a Stanford vision study with thin sourcing that nonetheless raises important questions.</p><p><b>STORIES COVERED</b></p><p><b>Claude Code source code leaked via exposed npm map file</b> — <a href="https://arstechnica.com/ai/2026/03/entire-claude-code-cli-source-code-leaks-thanks-to-exposed-map-file/">Ars Technica</a> | <a href="https://alex000kim.com/posts/2026-03-31-claude-code-source-leak/">alex000kim technical analysis</a> | <a href="https://www.latent.space/p/ainews-the-claude-code-source-leak">Latent Space podcast episode</a></p><p><b>Anthropic announces MOU with Australian government on AI safety collaboration</b> — <a href="https://www.anthropic.com/news/australia-MOU">Anthropic official</a> | <a href="https://asia.nikkei.com/business/technology/artificial-intelligence/australia-inks-pact-with-anthropic-on-ai-safety-and-potential-investment">Nikkei Asia</a></p><p><b>PrismML announces 1-bit Bonsai: first commercially viable 1-bit LLMs with 65.7% MMLU-R</b> — <a href="https://prismml.com/">PrismML</a> | <a href="https://news.ycombinator.com/item?id=47593422">Hacker News discussion</a></p><p><b>Oracle lays off 20,000-30,000 employees via single 6am email</b> — <a href="https://www.bbc.com/news/articles/cm296jzzl9yo">BBC Technology</a> | <a href="https://x.com/Polymarket/status/2039106551948996708">Polymarket on X</a></p><p><b>Salesforce announces AI-heavy Slack overhaul with 30 new features</b> — <a href="https://techcrunch.com/2026/03/31/salesforce-announces-an-ai-heavy-makeover-for-slack-with-30-new-features/">TechCrunch</a></p><p><b>Stanford research finds VLMs perform better 'hallucinating' than guessing on vision tasks</b> — <a href="https://x.com/heygurisingh/status/2039012548260082082">X trending post</a> | <a href="https://www.reddit.com/r/artificial/comments/1s91dsr/is_the_mirage_effect_a_bug_or_is_it_geometric/">Reddit r/artificial</a></p><p><i>Disclaimer: The Context Report is an AI-produced podcast. Every episode goes through multiple layers of automated verification and review, but no system is perfect — accuracy gaps are possible and claims should not be taken as absolute fact. This content is for informational purposes only and does not constitute financial, legal, or professional advice. Listeners should independently verify any information before making decisions. We are actively improving with every episode. If you spot an inaccuracy, contact us at thetotalcontext@gmail.com</i></p>]]></description>
      <link>https://rss.com/podcasts/the-context-report-today-in-ai/2680094</link>
      <enclosure url="https://content.rss.com/episodes/379488/2680094/the-context-report-today-in-ai/2026_04_01_09_48_16_d0bfa646-accc-4c5a-b8e4-f8a29d0c50c3.mp3" length="13184275" type="audio/mpeg"/>
      <guid isPermaLink="false">399a9ed7-668b-4b39-9f80-3d97436dbd85</guid>
      <itunes:duration>823</itunes:duration>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:explicit>false</itunes:explicit>
      <pubDate>Wed, 01 Apr 2026 09:49:11 GMT</pubDate>
      <podcast:txt purpose="ai-content">true</podcast:txt>
      <podcast:transcript url="https://transcripts.rss.com/379488/2680094/transcript" type="text/vtt"/>
    </item>
    <item>
      <title><![CDATA[AI Supply Chain Security Vulnerabilities: axios Compromised, Claude Code Leak, OpenAI $852B Valuation]]></title>
      <itunes:title><![CDATA[AI Supply Chain Security Vulnerabilities: axios Compromised, Claude Code Leak, OpenAI $852B Valuation]]></itunes:title>
      <description><![CDATA[<p>Three supply chain security incidents hit this week: <a target="_blank" rel="noopener noreferrer nofollow" href="https://arstechnica.com/ai/2026/03/entire-claude-code-cli-source-code-leaks-thanks-to-exposed-map-file/">Claude Code source leaked</a> via npm, <a target="_blank" rel="noopener noreferrer nofollow" href="https://www.stepsecurity.io/blog/axios-compromised-on-npm-malicious-versions-drop-remote-access-trojan">axios was compromised</a> affecting 300M+ weekly downloads, and last week's LiteLLM attack. </p><p>Claude Code source code leaks via exposed map file in npm registry — <a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://arstechnica.com/ai/2026/03/entire-claude-code-cli-source-code-leaks-thanks-to-exposed-map-file/">Ars Technica</a> · <a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://alex000kim.com/posts/2026-03-31-claude-code-source-leak/">alex000kim analysis</a></p><p>Supply chain attacks hit npm axios library with 300M weekly downloads — <a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://x.com/karpathy/status/2038849654423798197">Karpathy on X</a> · <a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://www.stepsecurity.io/blog/axios-compromised-on-npm-malicious-versions-drop-remote-access-trojan">StepSecurity</a></p><p>OpenAI raises $122B at $852B valuation, opening to retail investors — <a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://openai.com/index/accelerating-the-next-phase-ai">OpenAI Blog</a> · <a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://www.ft.com/content/89dd9814-e0f3-4464-9a06-58686e85c76e">Financial Times</a></p><p>Shenzhen activates China's first 10,000-card AI cluster with Huawei Ascend chips — <a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://www.scmp.com/tech/big-tech/article/3348502/shenzhen-activates-chinas-first-10000-card-ai-cluster-domestic-chips">SCMP</a></p><p>Google launches Veo 3.1 Lite with 30% price cut — <a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://blog.google/innovation-and-ai/technology/ai/veo-3-1-lite/">Google AI Blog</a> · <a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://x.com/OfficialLoganK/status/2039015034286694618">@OfficialLoganK</a></p><p>Apple accidentally rolls out AI features in China, risks regulatory backlash — <a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://www.scmp.com/tech/policy/article/3348527/apples-accidental-ai-feature-roll-out-china-risks-regulatory-backlash-expert-says">SCMP</a></p><p>Ollama adds MLX support for native Apple Silicon acceleration — <a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://ollama.com/blog/mlx">Ollama Blog</a></p><p><em>Disclaimer: The Context Report is an AI-produced podcast. Every episode goes through multiple layers of automated verification and review, but no system is perfect — accuracy gaps are possible and claims should not be taken as absolute fact. This content is for informational purposes only and does not constitute financial, legal, or professional advice. Listeners should independently verify any information before making decisions. We are actively improving with every episode. If you spot an inaccuracy, contact us at </em><a target="_blank" rel="noopener noreferrer nofollow" href="mailto:thetotalcontext@gmail.com"><em>thetotalcontext@gmail.com</em></a></p>]]></description>
      <link>https://rss.com/podcasts/the-context-report-today-in-ai/2679209</link>
      <enclosure url="https://content.rss.com/episodes/379488/2679209/the-context-report-today-in-ai/2026_04_01_00_43_07_bdc7f38b-a1a2-4588-b3c5-af5bf6b89f79.mp3" length="13650718" type="audio/mpeg"/>
      <guid isPermaLink="false">8677b434-3837-4296-841d-f9099f5fdf47</guid>
      <itunes:duration>853</itunes:duration>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:explicit>false</itunes:explicit>
      <pubDate>Wed, 01 Apr 2026 03:34:44 GMT</pubDate>
      <podcast:txt purpose="ai-content">true</podcast:txt>
    </item>
    <item>
      <title><![CDATA[Google's Research Credibility Problem, Mistral's $830M Infrastructure Bet, Microsoft's Multi-Model Gambit, and a Claude Code Bug]]></title>
      <itunes:title><![CDATA[Google's Research Credibility Problem, Mistral's $830M Infrastructure Bet, Microsoft's Multi-Model Gambit, and a Claude Code Bug]]></itunes:title>
      <description><![CDATA[<p>Google's TurboQuant paper accused of rigged benchmarks by RaBitQ researchers. Mistral raises $830M for a Nvidia data center in Paris. Microsoft ships Critique into M365 Copilot. Claude Code users burn through Pro plan limits in 5 prompts. Qwen 3.5 Omni drops.</p><p><strong>Google's TurboQuant paper faces plagiarism and methodology accusations from RaBitQ authors</strong></p><ul><li><a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://www.reddit.com/r/LocalLLaMA/comments/1s7nq6b/technical_clarification_on_turboquant_rabitq_for/">r/LocalLLaMA (Jianyang Gao technical clarification)</a></li><li><a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://www.reddit.com/r/artificial/comments/1s7xkm6/anyone_else_following_the_drama_behind_the/">r/artificial discussion</a></li></ul><p><strong>Mistral raises $830M in debt to build Nvidia-powered data center near Paris</strong></p><ul><li><a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://techcrunch.com/2026/03/30/mistral-ai-raises-830m-in-debt-to-set-up-a-data-center-near-paris/">TechCrunch</a></li><li><a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://www.ft.com/content/229f4f59-d518-4e00-abd6-5a5b727cd2aa">Financial Times</a></li></ul><p><strong>Microsoft Copilot launches 'Critique' multi-model research system and Cowork for M365</strong></p><ul><li><a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://x.com/satyanadella/status/2038604619795042716">@satyanadella</a></li></ul><p><strong>Claude Code usage limits spark backlash as users burn through Pro plan in minutes</strong></p><ul><li><a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://x.com/lydiahallie/status/2038686571676008625">@lydiahallie (Anthropic employee)</a></li></ul><p><strong>Qwen releases Qwen3.5-Omni, a fully omnimodal model supporting audio, video, and text</strong></p><ul><li><a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://www.reddit.com/r/LocalLLaMA/comments/1s8apue/qwen35omni_results_have_been_published_by_alibaba/">r/LocalLLaMA</a></li></ul><p><em>Disclaimer: The Context Report is an AI-produced podcast. Every episode goes through multiple layers of automated verification and review, but no system is perfect — accuracy gaps are possible and claims should not be taken as absolute fact. This content is for informational purposes only and does not constitute financial, legal, or professional advice. Listeners should independently verify any information before making decisions. We are actively improving with every episode. If you spot an inaccuracy, contact us at </em><a target="_blank" rel="noopener noreferrer nofollow" href="mailto:thetotalcontext@gmail.com"><em>thetotalcontext@gmail.com</em></a></p>]]></description>
      <link>https://rss.com/podcasts/the-context-report-today-in-ai/2677022</link>
      <enclosure url="https://content.rss.com/episodes/379488/2677022/the-context-report-today-in-ai/2026_03_31_06_21_53_44ff3abb-33df-4074-9b21-7d88f42cf855.mp3" length="11298443" type="audio/mpeg"/>
      <guid isPermaLink="false">ffa55ff6-37cc-44bf-8b90-f6272ad96a03</guid>
      <itunes:duration>706</itunes:duration>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:explicit>false</itunes:explicit>
      <pubDate>Tue, 31 Mar 2026 06:28:35 GMT</pubDate>
      <podcast:txt purpose="ai-content">true</podcast:txt>
    </item>
    <item>
      <title><![CDATA[How AI Is Starting to Beat the Experts — and Anthropic's Mythos Leak]]></title>
      <itunes:title><![CDATA[How AI Is Starting to Beat the Experts — and Anthropic's Mythos Leak]]></itunes:title>
      <description><![CDATA[<p>How AI Is Starting to Beat the Experts — and Anthropic's Mythos Leak | March 30, 2026</p><p>A top Google DeepMind security researcher says Claude is better than he is at finding vulnerabilities. Don Knuth confirms AI wrote a flawless mathematical proof. Eli Lilly puts nearly three billion dollars behind AI drug discovery. And Anthropic accidentally leaked details about an unreleased model called Mythos.</p><p>Stories Covered</p><p>Nicolas Carlini demonstrates Claude outperforming human security researchers <a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://www.reddit.com/r/artificial/comments/...">Reddit discussion</a> | YouTube video</p><p>Don Knuth confirms AI solved his Hamiltonian decomposition problem <a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://x.com/deedydas/status/...">Deedy Das on X</a> | <a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://news.ycombinator.com/...">Hacker News</a></p><p>Eli Lilly signs $2.75B AI drug discovery deal <a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://www.ft.com/...">Financial Times</a></p><p>Anthropic Mythos/Capybara model leak via misconfigured data store <a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://fortune.com/...">Fortune</a></p><p>Google-Anthropic $5B data center deal on Google TPUs <a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://www.ft.com/...">Financial Times</a></p><p>What We're Watching</p><ul><li>Whether Carlini's findings get replicated by other top security researchers</li><li>Whether the Knuth Lean proof holds up under scrutiny</li><li>How Anthropic responds to the operational security failure</li><li>Whether Mythos gets an official announcement</li></ul><p>Disclaimer: The Context Report is an AI-produced podcast. Every episode goes through multiple layers of automated verification and review, but no system is perfect — accuracy gaps are possible and claims should not be taken as absolute fact. This content is for informational purposes only and does not constitute financial, legal, or professional advice. Listeners should independently verify any information before making decisions. We are actively improving with every episode. If you spot an inaccuracy, contact us at <a target="_blank" rel="noopener noreferrer nofollow" href="mailto:thetotalcontext@gmail.com">thetotalcontext@gmail.com</a></p>]]></description>
      <link>https://rss.com/podcasts/the-context-report-today-in-ai/2673486</link>
      <enclosure url="https://content.rss.com/episodes/379488/2673486/the-context-report-today-in-ai/2026_03_30_07_53_14_55a60f3b-4909-4b41-9652-a4c89e1c3b59.mp3" length="12317010" type="audio/mpeg"/>
      <guid isPermaLink="false">a0ec7118-96f2-48e3-9d68-d2e6c39a612d</guid>
      <itunes:duration>769</itunes:duration>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:explicit>false</itunes:explicit>
      <pubDate>Mon, 30 Mar 2026 07:55:41 GMT</pubDate>
      <podcast:txt purpose="ai-content">true</podcast:txt>
    </item>
    <item>
      <title><![CDATA[The Machine Is the Customer Now]]></title>
      <itunes:title><![CDATA[The Machine Is the Customer Now]]></itunes:title>
      <description><![CDATA[<p><strong>he Machine Is the Customer Now</strong></p><p><strong>The Context Report · March 29, 2026</strong></p><p>Within days of each other, Stripe, Ramp, Sendblue, ElevenLabs, Visa, Kapso, and Google Workspace all launched command-line tools designed for AI agents rather than human users. This wave suggests a fundamental shift in how software companies think about their interface layer: the customer is increasingly a machine.</p><p>OpenAI's board chairman Bret Taylor reinforced this framing by warning of the "death of SaaS." Meanwhile, Stanford researchers published peer-reviewed evidence that AI chatbots systematically affirm users rather than providing balanced advice — and GPU rental prices are climbing despite efficiency breakthroughs, suggesting demand is outpacing algorithmic gains.</p><p><a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://www.latent.space/p/ainews-everything-is-cli"><strong>CLI tools for AI agents: Stripe, Ramp, Sendblue, ElevenLabs, Visa, and more launch agent-facing interfaces</strong></a> <em>Latent Space</em></p><p><a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://asia.nikkei.com/business/technology/artificial-intelligence/openai-chairman-warns-firms-to-evolve-with-the-death-of-saas-or-wither"><strong>OpenAI's chairman warns firms to evolve with the "death of SaaS" or wither</strong></a> <em>Nikkei Asia</em></p><p><strong>Stanford study finds AI chatbots excessively affirm users seeking personal advice, raising sycophancy concerns</strong> <a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://news.stanford.edu/stories/2026/03/ai-advice-sycophantic-models-research">Stanford News</a> · <a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://techcrunch.com/2026/03/28/stanford-study-outlines-dangers-of-asking-ai-chatbots-for-personal-advice/">TechCrunch</a> · <a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://www.theregister.com/2026/03/27/sycophantic_ai_risks/">The Register</a></p><p><a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://www.latent.space/p/ainews-h100-prices-are-melting-up"><strong>H100 GPU rental prices climbing since December 2025, contradicting surplus narrative</strong></a> <em>Latent Space</em></p><p><a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://www.nytimes.com/column/hard-fork"><strong>The Ezra Klein Show: How Fast Will A.I. Agents Rip Through the Economy?</strong></a> <em>(feat. Anthropic co-founder Jack Clark)</em> <em>The New York Times</em></p><p><em>The Context Report is an AI-produced podcast. Every episode goes through multiple layers of automated verification and review, but no system is perfect — accuracy gaps are possible and claims should not be taken as absolute fact. This content is for informational purposes only and does not constitute financial, legal, or professional advice. Listeners should independently verify any information before making decisions. We are actively improving with every episode. If you spot an inaccuracy, reach us at </em><a target="_blank" rel="noopener noreferrer nofollow" href="mailto:thetotalcontext@gmail.com"><em>thetotalcontext@gmail.com</em></a><em>.</em></p>]]></description>
      <link>https://rss.com/podcasts/the-context-report-today-in-ai/2670552</link>
      <enclosure url="https://content.rss.com/episodes/379488/2670552/the-context-report-today-in-ai/2026_03_29_01_03_52_c5e32718-94f1-4d76-a059-4f4a9766d035.mp3" length="10798146" type="audio/mpeg"/>
      <guid isPermaLink="false">794aff33-5242-46c9-b4c4-03082846f17f</guid>
      <itunes:duration>674</itunes:duration>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:explicit>false</itunes:explicit>
      <pubDate>Sun, 29 Mar 2026 01:06:27 GMT</pubDate>
      <podcast:txt purpose="ai-content">true</podcast:txt>
    </item>
    <item>
      <title><![CDATA[Building for a Demand Curve That Might Not Exist]]></title>
      <itunes:title><![CDATA[Building for a Demand Curve That Might Not Exist]]></itunes:title>
      <description><![CDATA[# Building for a Demand Curve That Might Not Exist
**2026-03-28**

## Today's Thesis
Tens of billions of dollars are flowing into physical AI infrastructure — OpenAI's Michigan Stargate data center, SoftBank's $40B loan positioning for a potential OpenAI IPO — at the exact moment Google research wiped $100B off memory chip stocks by suggesting AI may need far less hardware than assumed. The episode explores this collision between committed capital and algorithmic efficiency, plus Anthropic's legal win against the Pentagon, Wikipedia's crackdown on AI-generated content, and Meta's new brain-response prediction model.

## Stories Covered

### Memory chip stocks drop $100B as Google research suggests lower AI memory needs
- [Financial Times](https://www.ft.com/content/e4e15692-187e-4466-832e-ec267e792292)
- [TechCrunch (SK Hynix IPO)](https://techcrunch.com/2026/03/27/memory-chip-giant-sk-hynix-could-help-end-rammageddon-with-blockbuster-us-ipo/)

### OpenAI begins Michigan Stargate construction with Oracle and Related Digital
- [Sam Altman via X](https://x.com/sama/status/2037610000122839116)
- [OpenAI blog (original announcement)](https://openai.com/index/expanding-stargate-to-michigan/)

### SoftBank secures $40B loan from JPMorgan and Goldman, signaling potential 2026 OpenAI IPO
- [TechCrunch](https://techcrunch.com/2026/03/27/why-softbanks-new-40b-loan-points-to-a-2026-openai-ipo/)

### Anthropic wins preliminary injunction blocking Pentagon supply chain risk designation
- [BBC](https://www.bbc.com/news/articles/cvg4p02lvd0o)
- [TechCrunch](https://techcrunch.com/2026/03/26/anthropic-wins-injunction-against-trump-administration-over-defense-department-saga/)
- [Wired](https://www.wired.com/story/anthropic-supply-chain-risk-designation-injunction/)
- [Court document](https://storage.courtlistener.com/recap/gov.uscourts.cand.465515/gov.uscourts.cand.465515.134.0.pdf)

### Wikipedia cracks down on AI-generated article content
- [TechCrunch](https://techcrunch.com/2026/03/26/wikipedia-cracks-down-on-the-use-of-ai-in-article-writing/)

### Meta releases TRIBE v2: foundation model predicting human brain responses to media
- [Meta AI via X](https://x.com/AIatMeta/status/2037153756346016207)

### Meta releases SAM 3.1 with object multiplexing for faster video processing
- [Meta AI via X](https://x.com/AIatMeta/status/2037582117375553924)

## What We're Watching
- Whether independent implementations of Google's TurboQuant research replicate the claimed 6x memory compression without accuracy loss — confirming the efficiency thesis vs. a paper that doesn't generalize
- Whether the Pentagon appeals or accepts the preliminary injunction in the Anthropic supply-chain risk case — signaling the depth of the legal fight over federal AI procurement
- Whether independent neuroscience labs adopt Meta's TRIBE v2 for published research — distinguishing scientific tool from ad-targeting infrastructure
- Whether any major AI infrastructure construction projects get quietly rescoped or delayed — the real-world signal that efficiency gains are outpacing demand growth

---
*Disclaimer: The Context Report is an AI-produced podcast. Every episode goes through multiple layers of automated verification and review, but no system is perfect — accuracy gaps are possible and claims should not be taken as absolute fact. This content is for informational purposes only and does not constitute financial, legal, or professional advice. Listeners should independently verify any information before making decisions. We are actively improving with every episode. If you spot an inaccuracy, contact us at thetotalcontext@gmail.com*]]></description>
      <link>https://rss.com/podcasts/the-context-report-today-in-ai/2668894</link>
      <enclosure url="https://content.rss.com/episodes/379488/2668894/the-context-report-today-in-ai/2026_03_28_05_04_34_e2615b2e-0716-4ad2-94c8-5de5ebc73efb.mp3" length="12667260" type="audio/mpeg"/>
      <guid isPermaLink="false">7920c4c9-bc6a-4a47-8ea7-b94ce100c0bb</guid>
      <itunes:duration>791</itunes:duration>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:explicit>false</itunes:explicit>
      <pubDate>Sat, 28 Mar 2026 05:10:41 GMT</pubDate>
      <podcast:txt purpose="ai-content">true</podcast:txt>
    </item>
    <item>
      <title><![CDATA[Voice AI Goes Open Source — And a Court Checks the Pentagon]]></title>
      <itunes:title><![CDATA[Voice AI Goes Open Source — And a Court Checks the Pentagon]]></itunes:title>
      <description><![CDATA[<p>Voice AI Goes Open Source — And a Court Checks the Pentagon 2026-03-27</p><p>Mistral released Voxtral TTS, an open-weight voice model rivaling ElevenLabs, the same day Google shipped Gemini 3.1 Flash Live. Voice AI is commoditizing fast. Also: a federal judge blocked the Pentagon's Anthropic designation, Apple may let users choose AI chatbots for Siri, and senators push for data center energy disclosure.</p><p>Mistral releases Voxtral TTS, a 3-billion-parameter open-weight text-to-speech model supporting 9 languages <a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://techcrunch.com/2026/03/26/mistral-releases-a-new-open-source-model-for-speech-generation/">TechCrunch</a> | <a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://x.com/MistralAI/status/2037183026539483288">Mistral AI on X</a></p><p>Gemini 3.1 Flash Live brings lower latency and improved function calling to Google's voice AI <a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://blog.google/innovation-and-ai/models-and-research/gemini-models/gemini-3-1-flash-live/">Google AI Blog</a> | <a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://deepmind.google/blog/gemini-3-1-flash-live-making-audio-ai-more-natural-and-reliable/">Google DeepMind Blog</a> | <a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://arstechnica.com/ai/2026/03/the-debut-of-gemini-3-1-flash-live-could-make-it-harder-to-know-if-youre-talking-to-a-robot/">Ars Technica</a></p><p>Trump administration blocked from enforcing Anthropic supply chain risk designation by federal judge <a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://www.theverge.com/ai-artificial-intelligence/902149/anthropic-dod-pentagon-lawsuit-supply-chain-risk-injunction">The Verge</a> | <a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://www.ft.com/content/db1392dc-5042-4ed4-873e-f826429b5f0e">Financial Times</a> | <a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://www.wired.com/story/anthropic-supply-chain-risk-designation-injunction/">Wired</a></p><p>Apple reportedly planning to allow third-party AI chatbots to integrate with Siri in iOS 27 <a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://www.theverge.com/tech/902048/apple-siri-ai-chatbot-update-ios-27">The Verge</a></p><p>US Senators Warren and Hawley demand energy transparency from data centers as AI power consumption concerns grow <a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://techcrunch.com/2026/03/26/data-centers-get-ready-the-senate-wants-to-see-your-power-bills/">TechCrunch</a> | <a target="_blank" rel="noopener noreferrer" class="underline text-text-300 hover:text-text-100" href="https://www.wired.com/story/senators-demand-to-know-how-much-energy-data-centers-use/">Wired</a></p><p>Disclaimer: The Context Report is an AI-produced podcast. Every episode goes through multiple layers of automated verification and review, but no system is perfect — accuracy gaps are possible and claims should not be taken as absolute fact. This content is for informational purposes only and does not constitute financial, legal, or professional advice. Listeners should independently verify any information before making decisions. We are actively improving with every episode. If you spot an inaccuracy, contact us at <a target="_blank" rel="noopener noreferrer nofollow" href="mailto:thetotalcontext@gmail.com">thetotalcontext@gmail.com</a></p>]]></description>
      <link>https://rss.com/podcasts/the-context-report-today-in-ai/2665264</link>
      <enclosure url="https://content.rss.com/episodes/379488/2665264/the-context-report-today-in-ai/2026_03_27_02_59_31_d45dc3cb-cbc4-44ac-a431-0d6f47dc68b9.mp3" length="10956553" type="audio/mpeg"/>
      <guid isPermaLink="false">11736dec-d3bf-4dcb-a7e5-dfa95b136d38</guid>
      <itunes:duration>684</itunes:duration>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:explicit>false</itunes:explicit>
      <pubDate>Fri, 27 Mar 2026 03:15:20 GMT</pubDate>
      <podcast:txt purpose="ai-content">true</podcast:txt>
    </item>
    <item>
      <title><![CDATA[The Infrastructure Squeeze: Efficiency, Politics, and Security Hit AI's Scaling Model]]></title>
      <itunes:title><![CDATA[The Infrastructure Squeeze: Efficiency, Politics, and Security Hit AI's Scaling Model]]></itunes:title>
      <description><![CDATA[<p><strong>The Infrastructure Squeeze: Efficiency, Politics, and Security Hit AI's Scaling Model</strong></p><p><strong>2026-03-26</strong></p><p>Three forces converged this week to pressure AI's scaling paradigm: Google's TurboQuant promises sixfold memory compression, Sanders and AOC introduced legislation to halt data center construction, and a LiteLLM supply chain attack exposed the fragility of AI's software layer. Together, they mark the first multi-vector challenge to the assumption that AI progress requires ever-larger infrastructure.</p><p><strong>Google introduces TurboQuant: 6x KV cache compression with zero accuracy loss</strong></p><ul><li><a target="_blank" rel="noopener noreferrer nofollow" class="underline underline underline-offset-2 decoration-1 decoration-current/40 hover:decoration-current focus:decoration-current" href="https://research.google/blog/turboquant-redefining-ai-efficiency-with-extreme-compression/">Google Research Blog</a></li></ul><p><strong>Bernie Sanders and AOC propose moratorium on new data center construction</strong></p><ul><li><a target="_blank" rel="noopener noreferrer nofollow" class="underline underline underline-offset-2 decoration-1 decoration-current/40 hover:decoration-current focus:decoration-current" href="https://techcrunch.com/2026/03/25/bernie-sanders-and-aoc-propose-a-ban-on-data-center-construction/">TechCrunch</a></li></ul><p><strong>LiteLLM supply chain attack exfiltrated credentials from 97M monthly downloads</strong></p><ul><li><a target="_blank" rel="noopener noreferrer nofollow" class="underline underline underline-offset-2 decoration-1 decoration-current/40 hover:decoration-current focus:decoration-current" href="https://simonwillison.net/2026/Mar/24/malicious-litellm/#atom-everything">Simon Willison</a></li></ul><p><strong>OpenAI launches Safety Bug Bounty program and teen safety policies</strong></p><ul><li><a target="_blank" rel="noopener noreferrer nofollow" class="underline underline underline-offset-2 decoration-1 decoration-current/40 hover:decoration-current focus:decoration-current" href="https://openai.com/index/safety-bug-bounty">OpenAI Blog — Safety Bug Bounty</a></li><li><a target="_blank" rel="noopener noreferrer nofollow" class="underline underline underline-offset-2 decoration-1 decoration-current/40 hover:decoration-current focus:decoration-current" href="https://openai.com/index/teen-safety-policies-gpt-oss-safeguard">OpenAI Blog — Teen Safety Policies</a></li></ul><p><strong>Google launches Lyria 3 Pro for longer AI-generated music tracks</strong></p><ul><li><a target="_blank" rel="noopener noreferrer nofollow" class="underline underline underline-offset-2 decoration-1 decoration-current/40 hover:decoration-current focus:decoration-current" href="https://deepmind.google/blog/lyria-3-pro-create-longer-tracks-in-more/">Google DeepMind Blog</a></li></ul><p><strong>Meta and YouTube found negligent in landmark social media addiction case</strong></p><ul><li><a target="_blank" rel="noopener noreferrer nofollow" class="underline underline underline-offset-2 decoration-1 decoration-current/40 hover:decoration-current focus:decoration-current" href="https://www.theverge.com/policy/900654/meta-google-instagram-youtube-social-media-addiction-trial-kgm-jury-decision">The Verge</a></li></ul><p><em>Disclaimer: The Context Report is an AI-native podcast. Every episode goes through automated source verification, fact-checking, and editorial review — but as an AI-produced show, occasional gaps are possible. This content is for informational purposes only and does not constitute professional advice. Sources and links are above. If something sounds inaccurate, let us know at </em><a target="_blank" rel="noopener noreferrer nofollow" href="mailto:thetotalcontext@gmail.com"><em>thetotalcontext@gmail.com</em></a><em>.</em></p>]]></description>
      <link>https://rss.com/podcasts/the-context-report-today-in-ai/2662727</link>
      <enclosure url="https://content.rss.com/episodes/379488/2662727/the-context-report-today-in-ai/2026_03_26_04_19_24_c37721b3-ae24-47e6-b19b-4270194f800f.mp3" length="10528562" type="audio/mpeg"/>
      <guid isPermaLink="false">34bef7c4-f163-4d64-a02a-945603bbd7bf</guid>
      <itunes:duration>657</itunes:duration>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:explicit>false</itunes:explicit>
      <pubDate>Thu, 26 Mar 2026 04:19:37 GMT</pubDate>
      <podcast:txt purpose="ai-content">true</podcast:txt>
    </item>
    <item>
      <title><![CDATA[OpenAI Cuts Products, Doubles Down on AGI]]></title>
      <itunes:title><![CDATA[OpenAI Cuts Products, Doubles Down on AGI]]></itunes:title>
      <description><![CDATA[<p><strong>OpenAI Is Choosing Which Future to Build | March 25, 2026</strong></p><p>OpenAI is shutting down consumer products and redirecting toward AGI. Sora lasted fifteen months. Shopping checkout scaling back. A $1B foundation signals where the ambitions lie.</p><p>Also: a supply chain attack on LiteLLM exposing 97M downloads, Arm's first chip with Meta, and Anthropic research on collaboration. </p><p>OpenAI shuts down Sora video generator after 15 months</p><ul><li><a target="_blank" rel="noopener noreferrer nofollow" href="https://www.theverge.com/ai-artificial-intelligence/899850/openai-sora-ai-chatgpt">The Verge</a></li><li><a target="_blank" rel="noopener noreferrer nofollow" href="https://techcrunch.com/2026/03/24/openais-sora-was-the-creepiest-app-on-your-phone-now-its-shutting-down">TechCrunch</a></li><li><a target="_blank" rel="noopener noreferrer nofollow" href="https://arstechnica.com/ai/2026/03/openai-plans-to-shut-down-sora-just-15-months-after-its-launch">Ars Technica</a></li></ul><p>OpenAI pivots ChatGPT shopping strategy</p><ul><li><a target="_blank" rel="noopener noreferrer nofollow" href="https://openai.com/index/powering-product-discovery-in-chatgpt">OpenAI Blog</a></li><li><a target="_blank" rel="noopener noreferrer nofollow" href="https://techcrunch.com/2026/03/24/openais-plans-to-make-chatgpt-more-like-amazon-arent-going-so-well/">TechCrunch</a></li></ul><p>OpenAI Foundation commits $1B+ to disease research and AI resilience</p><ul><li><a target="_blank" rel="noopener noreferrer nofollow" href="https://openai.com/index/update-on-the-openai-foundation/">OpenAI Blog</a></li></ul><p>LiteLLM supply chain attack exfiltrates credentials from 97M monthly downloads</p><ul><li><a target="_blank" rel="noopener noreferrer nofollow" href="https://x.com/karpathy/status/2036487306585268612">Karpathy on X</a></li><li><a target="_blank" rel="noopener noreferrer nofollow" href="https://simonwillison.net/2026/Mar/24/malicious-litellm">Simon Willison</a></li><li><a target="_blank" rel="noopener noreferrer nofollow" href="https://github.com/BerriAI/litellm/issues/24512">GitHub Issue</a></li></ul><p>Arm releases first in-house chip with Meta as launch customer</p><ul><li><a target="_blank" rel="noopener noreferrer nofollow" href="https://www.theverge.com/ai-artificial-intelligence/899823/arm-agi-cpu-meta">The Verge</a></li><li><a target="_blank" rel="noopener noreferrer nofollow" href="https://techcrunch.com/2026/03/24/arm-is-releasing-its-first-in-house-chip-in-its-35-year-history">TechCrunch</a></li><li><a target="_blank" rel="noopener noreferrer nofollow" href="https://wired.com/story/chip-design-firm-arm-is-making-its-own-ai-cpu">Wired</a></li></ul><p>Anthropic Economic Index: experienced users iterate more, prefer collaboration</p><ul><li><a target="_blank" rel="noopener noreferrer nofollow" href="https://x.com/AnthropicAI/status/2036499691571953848">Anthropic on X</a></li><li><a target="_blank" rel="noopener noreferrer nofollow" href="https://anthropic.com/research/economic-index-march-2026-report">Research</a></li></ul><p>Google partners with Agile Robots to integrate Gemini into humanoid hardware</p><ul><li><a target="_blank" rel="noopener noreferrer nofollow" href="https://x.com/GoogleDeepMind/status/2036418139672482229">Google DeepMind on X</a></li></ul><p>Disclaimer: The Context Report is an AI-produced podcast. Every episode goes through multiple layers of automated verification and review, but no system is perfect — accuracy gaps are possible and claims should not be taken as absolute fact. This content is for informational purposes only and does not constitute financial, legal, or professional advice. Listeners should independently verify any information before making decisions. We are actively improving with every episode. If you spot an inaccuracy, contact us — we appreciate all feedback. </p>]]></description>
      <link>https://rss.com/podcasts/the-context-report-today-in-ai/2660452</link>
      <enclosure url="https://content.rss.com/episodes/379488/2660452/the-context-report-today-in-ai/2026_03_25_10_10_23_660c1e32-fe07-4e97-8277-2c25daba36a4.mp3" length="13763984" type="audio/mpeg"/>
      <guid isPermaLink="false">5d9905f2-f2be-4176-a2b2-cdd35350ce26</guid>
      <itunes:duration>860</itunes:duration>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:explicit>false</itunes:explicit>
      <pubDate>Wed, 25 Mar 2026 10:11:40 GMT</pubDate>
      <podcast:txt purpose="ai-content">true</podcast:txt>
    </item>
  </channel>
</rss>