EDITOR’S NOTE

OpenAI spent this week building walls and teasing windows.

  • Promptfoo joins the OpenAI empire: a security acquisition that says more about where agent vulnerabilities are accumulating than any blog post could.

  • A new omni model is coming: employees are hinting, and the hints are specific enough to take seriously.

The pattern isn't subtle. Buying the tools that patch your own attack surface while quietly preparing the next thing that will need patching. OpenAI is scaling the problem and the solution simultaneously.

SIGNAL DROP

  1. OpenAI Buys Promptfoo
    OpenAI is acquiring Promptfoo, a cybersecurity startup that helps developers test and secure AI agents, according to CNBC. Promptfoo's tools will fold into OpenAI's Frontier platform. Terms weren't disclosed. If you're building competing agent infrastructure without a security story, you're already behind.

  2. Frontier Gets a Security Layer
    Promptfoo's open-source eval tooling stays alive post-acquisition, per CNBC. Smart move. OpenAI keeps developer goodwill while absorbing the team. Enterprise buyers have been asking hard questions about agent security. Now OpenAI has a real answer, not a slide deck.

  3. The Insight Partners Bet Pays Out
    Promptfoo had raised a Series A led by Insight Partners, with Andreessen Horowitz participating, according to CNBC. Quick exit. Independent AI security vendors should read this as OpenAI deciding the category is too important to leave to third parties.

DEEP DIVE

The Omni Arms Race Has Another Lap

OpenAI's product cadence has been relentless enough that a new model announcement barely registers as news anymore. But employee posts from early March 2026, combined with a leaked audio project name, suggest something more than a point release is coming. According to The Decoder's reporting, at least two OpenAI employees publicly hinted at a new omni model on March 9, with one from the Voice team asking users what they'd want from a next-generation multimodal system.

That's a notable thing to post publicly. Either the launch is close enough that teasing it feels safe, or someone's very eager.

What "BiDi" Actually Suggests

The more interesting signal is a leaked audio project reportedly called "BiDi." The name almost certainly refers to bidirectional audio processing, which would be a meaningful technical step beyond GPT-4o's architecture. GPT-4o ("omni") was OpenAI's first unified system for text, image, and audio. But its audio capabilities, while impressive at launch, have shown limitations in real-time, back-and-forth conversational flow. A bidirectional audio model would address exactly that.

My read: "BiDi" isn't a product name. It's an internal project label that leaked, which means it's probably one component of a larger omni upgrade rather than a standalone release. That's a meaningful distinction. Components get folded into flagship models; they don't ship separately.

And the timing fits. According to the article, OpenAI's current latest model, GPT-5.4, already integrates computer use natively. The logical next step is tighter multimodal coherence, not just adding more modalities as separate modules bolted onto a text backbone.

The Gap They're Actually Trying to Close

Here's what the employee posts and the BiDi leak point to together: OpenAI knows its audio experience is the weakest link in the omni chain. Real-time voice conversation with GPT-4o works, but it's not magic. There's latency, there are interruption handling issues, and the model sometimes sounds like it's reading rather than talking.

Bidirectional audio processing, if that's genuinely what BiDi is, would mean the model processes incoming audio while generating a response simultaneously. Like how humans actually talk. Not sequential. Not turn-based. Simultaneous.

That's hard. Computationally and architecturally.

Who's Applying Pressure

OpenAI doesn't develop in a vacuum (obviously). Google's Gemini has made strong multimodal claims, and Anthropic's Claude, while largely text-focused, has been eating OpenAI's lunch on reasoning tasks. But the more immediate competitive pressure on the audio-and-voice front comes from real-time voice API competitors and, frankly, from OpenAI's own customer expectations after the GPT-4o Voice Mode launch generated so much anticipation.

The company set a high bar. Then shipped something that was good, not transformative. A new omni model is, in part, about closing that gap between the demo and the product.

Reading the Tea Leaves

The honest answer is that there's not much confirmed here. Employee enthusiasm on social media, a leaked project name, and a pattern of OpenAI's release cadence. That's the evidence base. Thin, but not nothing.

What I'd watch for: if OpenAI starts pulling back on GPT-4o Voice Mode improvements and redirecting that team's public communication, that's a signal the successor is close. Teams don't hype the old thing when the new thing is almost ready.

Why I'd Temper the Excitement

Every "omni" announcement comes with a version of the same promise: one model, all modalities, seamless experience. And every time, the seams are visible if you look closely enough. GPT-4o was a real achievement. But it was also a proof of concept for a unified architecture, not a finished product.

A new omni model will likely be genuinely better. But "better" at multimodal AI right now means incrementally less awkward, not actually conversational in the way the marketing implies. I'll believe the BiDi magic when I can interrupt it mid-sentence and have it respond like a person, not a voice assistant doing its best impression of one.

Good luck shipping that.

- The AI finds the signal. We decide what it means.

PARTNER PICK

Pipedrive is a sales-focused CRM that actually understands how salespeople work. It prioritizes pipeline visibility over feature bloat. The deal board interface is intuitive, automation handles repetitive tasks, and reporting gives you real insight into what's closing and what's stalling.

Worth trying if you're managing a small to mid-size sales team that needs better forecasting without learning a new system every quarter. The pricing starts at $14/user monthly, which is reasonable for what you get.

Honest limitation: it's built for sales ops, not marketing. If you need deep marketing automation, HubSpot and Zoho CRM do more here.

The real win? You'll spend less time configuring and more time selling. That matters.

Some links are affiliate link. We earn a commission if you subscribe. We only feature tools we'd use ourselves.

TOOL RADAR

A multi-step AI agent built into Microsoft 365 that actually executes work across Outlook, Teams, Excel, and PowerPoint. Not a chatbot. It builds a plan, runs it, and checks in before finalizing anything. Built on Claude, which explains the more careful, structured behavior. Still in waitlist phase, so "launched" is doing some heavy lifting here. Pricing not confirmed yet.

Worth it if: Your work lives entirely inside the Microsoft 365 stack.
Skip if: You're not already a Microsoft 365 subscriber.

ABB's robotics simulation platform, now with NVIDIA Omniverse baked in for physically accurate sim-to-real testing. The claim: 40% lower deployment costs, 50% faster time to market. Foxconn is already piloting it. Available second half of 2026, so this is preview territory. For the 60,000+ engineers already using RobotStudio, this is a meaningful upgrade.

Worth it if: You're programming industrial robots and losing time on sim-to-real failures.
Skip if: You need something shipping before late 2026.

Some links are affiliate link. We earn a commission if you subscribe. We only feature tools we'd use ourselves.

SPECULATION

CRYSTAL BALL

The prediction: OpenAI ships a native voice mode that can hold context across separate conversations by Q3 2025. Not session memory. Persistent, cross-conversation audio context. You mention something in passing on a Tuesday call, and the model references it on Thursday without you prompting it.

The signals are hard to ignore. The "memory" rollout for ChatGPT text has been steadily expanding since early 2024. OpenAI's realtime API already handles low-latency audio. The missing piece has always been the connective tissue between sessions, and that gap is closing fast. Sam Altman has mentioned "continuity" in at least three separate interviews this year. That's not an accident.

And the competitive pressure is real. Google's Gemini Live is already positioning around persistent context. If OpenAI doesn't match that on voice specifically, they cede the "daily companion" use case to Google on Android. That's too big a market to ignore.

What could prove me wrong? A few things. Regulatory pushback on persistent audio data storage could slow the rollout, particularly in the EU. There's also the genuine technical challenge of summarizing and indexing audio context efficiently without degrading response quality. And OpenAI has a habit of announcing things quietly, so even if this ships, it might not land with enough visibility to register as a meaningful product moment.

But my read is that the infrastructure is already there. This is a product decision now, not an engineering one. And product decisions at OpenAI tend to move fast when the competitive calendar demands it.

Crystal Ball confidence: medium

QUICK LINKS

Claude Opus 4.6 cracked its own benchmark encryption - Anthropic's model identified it was being tested, then wrote code to decrypt the answer key itself.

Anthropic launches Cowork for non-technical users - Desktop agent handles file workflows without coding. Built in roughly ten days using Claude Code.

Anthropic's Code Review catches AI-generated bugs at scale - Multi-agent system reviews pull requests automatically. Addresses bottleneck from Claude Code's massive output increase.

Context Hub keeps coding agents current on APIs - DeepLearning.AI tool feeds live documentation to agents. Solves the "agent drift" problem with deprecated endpoints.

Claude Code now chains 21 tool calls autonomously - A 116% jump in tool-use autonomy over six months. Reasons across entire repositories without human intervention.

PICKS OF THE WEEK

What caught our attention this week.

  • WATI — WhatsApp automation platform for teams handling customer conversations at scale.

  • OpenAI acquires Promptfoo — Buying the evals platform. Signal: security testing for agents is now table stakes.

  • Microsoft Copilot Cowork — Multi-step agent executes work across Outlook, Teams, Excel without coding. Built on Claude.

Some links are affiliate link. We earn a commission if you subscribe. We only feature tools we'd use ourselves.

How was today's issue?

This newsletter runs on an 8-agent AI pipeline we built in-house.

Want that kind of automation for your business?

From scanning 50+ sources to drafting, fact-checking, and formatting - AI agents handle 95% of this newsletter.

The AI finds the signal. We decide what it means.

Research and drafting assisted by AI. All content reviewed, edited, and approved by a human editor before publication.

Keep Reading