LWiAI Podcast #238 - GPT 5.4 mini, OpenAI Pivot, Mamba 3, Attention Residuals

Last Week in AI / 4/1/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The podcast episode discusses recent developments in the AI model landscape, including “GPT 5.4 mini” and what it implies for smaller, more efficient model deployments.
  • It covers strategic commentary on an “OpenAI pivot,” framing how shifting directions could affect product roadmaps and the broader competitive landscape.
  • The episode examines “Mamba 3,” highlighting attention- and sequence-modeling alternatives and their potential impact on training/inference tradeoffs.
  • It also explores the concept of “attention residuals,” which focuses on architectural or training techniques that may improve performance or stability in transformer-like systems.
  • Overall, the content is positioned as an AI news-and-research roundup intended to connect emerging research ideas with near-term engineering and product implications.
Last Week in AI
Podcast
LWiAI Podcast #238 - GPT 5.4 mini, OpenAI Pivot, Mamba 3, Attention Residuals
0:00
Current time: 0:00 / Total time: -2:00:49
-2:00:49
Audio playback is not supported on your browser. Please upgrade.

LWiAI Podcast #238 - GPT 5.4 mini, OpenAI Pivot, Mamba 3, Attention Residuals

OpenAI ships GPT-5.4 mini and nano, faster and more capable but up to 4x pricier, DLSS 5 looks like a real-time generative AI filter for video games | The Verge, and more!
Last Week in AI's avatar
Apr 01, 2026
Share

Note from Andrey: this ep came out a week ago on RSS, but I was delayed posting it to youtube and therefore also Substack. My bad!

Our 238th episode with a summary and discussion of last week’s big AI news!

Recorded on 03/18/2026

Hosted by Andrey Kurenkov and Jeremie Harris

Feel free to email us your questions and feedback at andreyvkurenkov@gmail.com and/or hello@gladstone.ai

In this episode:

* OpenAI released GPT-5.4 mini and nano with 400k-token context windows, higher per-token prices but claimed token-efficiency gains in Codex; nano is API-only and pitched for high-volume classification/data extraction despite a major price increase.

* Mistral open-sourced the Small 4 model family (MoE, 119B total/6B active) combining reasoning, multimodal, and coding-agent capabilities, and announced Forge to help businesses train or post-train custom models.

* Agent “operating system” competition intensified with Meta’s acquired Manus launching a local Mac agent, Nvidia announcing NeMo/“Open Shell” sandboxed agent runtime, and Nvidia also unveiling DLSS 5 plus major hardware forecasts including Groq LPU integration.

* Business and safety updates included OpenAI shifting focus toward productivity/enterprise amid competition, Microsoft reorganizing Copilot and frontier-model efforts, Meta delaying its next model, China-linked ByteDance deploying large Nvidia clusters abroad, and new safety work on steganography, chain-of-thought faithfulness, fine-tuning defenses, cyber-attack evals, and constitution/spec compliance.

A thank you to our current sponsors:

  • Box - visit Box.com/AI to learn more

  • ODSC AI - go to odsc.ai/east and use promo code LWAI for an additional 15% off your pass to ODSC AI East 2026.

  • Factor - head to factormeals.com/lwai50off and use code lwai50off to get 50 percent off and free breakfast for a year

Timestamps:

Discussion about this episode

CommentsRestacks
User's avatar