Elena' s AI Blog

AI's New Bottleneck

Power, Policy, and Persistent Agents

27 Mar 2026 (updated: 27 Mar 2026) / 15 minutes to read

Elena Daehnhardt


Generated by Gemini 3 Flash / Nano Banana 2. Prompt: Minimalist infrastructure visualization showing constraints in power grids, regulatory policy, and persistent software agents.


TL;DR: AI development hit a physical and regulatory ceiling this week. Google shipped Gemini 3.1 Flash Live (March 26) and Lyria 3 Pro (March 25), but a proposed U.S. datacenter moratorium — backed by over 100 local community actions across 12 states — highlighted growing infrastructure constraints on future scaling. At the same time, Mythos/Capybara leak reporting suggested frontier releases are being gated by cyber-risk concerns and staged-access strategy, while GitHub's Copilot interaction-data policy shift (effective April 24, 2026) widened default data collection for model improvement.

Introduction

This week felt like two very different AI stories happening at the same time.

On one track, we got concrete, practical model releases — real-time voice and AI-generated music from Google. On the other, the constraints became more visible: energy and infrastructure pressure, data-privacy defaults, and a high-capability model leak that showed just how carefully labs are thinking about staged rollouts.

I find this fascinating. For a long time, the only question that seemed to matter was: how capable is the model? Now, three equally important questions run alongside it: Can we power it? Are we allowed to deploy it? And who gets access first?

This week illustrated all three constraints at once. Let me walk you through what happened.

What happened this week

  • U.S. lawmakers proposed a federal pause on new AI datacenter construction.
  • GitHub changed how it uses Copilot interaction data for Free, Pro, and Pro+ users.
  • AWS made Amazon Bedrock available in New Zealand for the first time.
  • Google launched Gemini 3.1 Flash Live, a low-latency real-time multimodal model.
  • Google launched Lyria 3 Pro, an extended music generation model, in public preview.
  • Details about Anthropic’s unreleased Mythos/Capybara model leaked, and Anthropic confirmed it exists.

Infrastructure and Governance

1. U.S. lawmakers proposed a federal pause on new AI datacenters

Lawmakers introduce bill to pause building of new datacenters - The Guardian

On March 25, 2026, a group of U.S. lawmakers introduced the Artificial Intelligence Data Center Moratorium Act. If passed, it would immediately halt all new AI datacenter construction in the U.S. — and block upgrades to existing facilities — until Congress passes comprehensive federal AI legislation covering worker protections, civil rights, environmental safeguards, and mandatory pre-release government review of AI products.

The bill faces steep odds in the current Congress, and critics on both sides of the aisle have pushed back on it. What makes it worth paying attention to is not whether it passes — it probably will not. What matters is the pressure it reflects.

More than 100 communities across 12 states have already enacted local datacenter moratoriums. That is a real grassroots movement that predates this bill, and it is the more durable signal regardless of what happens federally. The bill also proposes banning U.S. exports of AI computing hardware to countries without equivalent safeguards — a provision with significant implications for global AI supply chain strategy.

Takeaway: Grid-scale AI expansion is now a live federal policy debate, not just a local permitting headache.

Why this matters to you

Even a bill that never passes creates uncertainty. Uncertainty raises financing risk for new infrastructure builds. Utilities, planners, and infrastructure investors are already factoring this in. If you are building or evaluating multi-region AI deployments, it is worth watching how this pressure evolves — it is already influencing where companies choose to place compute.


2. GitHub changed Copilot data training defaults for Free, Pro, and Pro+ users

Updates to GitHub Copilot interaction data usage policy - GitHub Blog

On March 25, GitHub announced that starting April 24, 2026, interaction data from Copilot Free, Pro, and Pro+ plans may be used to train AI models — unless you opt out. This data includes everything you type into Copilot: prompts, suggestions you accept or reject, code snippets, and surrounding context.

GitHub Business and Enterprise plans are not affected. Students and teachers are exempt. If you had previously opted out, your preference carries over automatically.

Action item: If you are on Copilot Free, Pro, or Pro+, go to github.com/settings/copilot/features → Privacy → and disable “Allow GitHub to use my data for AI model training” before April 24.

Takeaway: The default data-training posture for a very large group of developers is changing. Opt-in is becoming opt-out.

Why this matters to you

There is an important asymmetry here. Business and Enterprise customers are protected by contract. Individual developers on free and lower-tier plans are protected only by a setting they must discover and manually disable. This kind of default shift — where data collection is “on” unless you know to turn it off — is becoming a common pattern across AI products. It is worth reviewing your settings regularly, not just for GitHub, but across any AI coding tool you use.


3. Amazon Bedrock is now available in New Zealand

Run Generative AI inference with Amazon Bedrock in Asia Pacific (New Zealand) - AWS

On March 26, AWS launched Amazon Bedrock in its new Auckland region (ap-southeast-6). Requests can be routed across Auckland, Sydney, and Melbourne, or globally where permitted.

Takeaway: Stable, smaller jurisdictions with clear data-residency laws are becoming genuine strategic options for production AI inference — not just latency optimisations.

Why this matters to you

New Zealand is an interesting case. It has a predictable data-protection law, strong geopolitical alignment with Five Eyes partners (Australia, Canada, the UK, and the U.S.), no domestic AI-specific regulatory risk, and strong enterprise buying comfort in regulated industries like finance and healthcare.

With U.S. datacenter politics becoming noisier, teams designing multi-region AI architectures are increasingly treating regional placement as risk management, not just latency tuning. New Zealand — with cross-border routing to Sydney and Melbourne already built in — fits that frame well.


Model Releases and Safety Pressure

4. Google launched Gemini 3.1 Flash Live for real-time multimodal conversations

Gemini 3.1 Flash Live: Making audio AI more natural and reliable - Google

Gemini 3.1 Flash Live Preview - Gemini API docs

Google published Gemini 3.1 Flash Live on March 26, 2026. It is available in preview via the Gemini Live API and Google AI Studio.

This model handles real-time, bidirectional conversations, processing audio and video input together with low latency. What makes this technically interesting is the architecture: instead of the traditional pipeline of transcribe speech → reason → synthesise reply, the model does all three in a single native pass. This dramatically reduces latency and makes the conversation feel much more natural.

Key details:

  • Benchmarks: 90.8% on ComplexFuncBench Audio (multi-step function calling in noisy environments) and 36.1% on Scale AI’s Audio MultiChallenge with “thinking” enabled — both leading scores at launch.
  • Conversation continuity: Google says Gemini Live (the product experience) can follow a conversation thread for 2× longer than before. This is a product-level behavior claim, not a published raw API context-window specification for the model.
  • Languages: 90+ languages supported for real-time multimodal conversations.
  • Safety: All audio output is watermarked with SynthID — an imperceptible watermark embedded at generation time to help identify AI-generated audio.
  • Enterprise deployments: Verizon and The Home Depot were cited as early production customers, both using it for contact centre applications.
  • Also powers the global rollout of Search Live, now active in 200+ countries.

Takeaway: Native multimodal real-time voice, deployed in large enterprise products at launch, suggests voice interaction is consolidating into a primary layer for persistent AI agents.

Why this matters to you

The benchmark number worth watching is 90.8% on ComplexFuncBench Audio. This measures how well the model handles multi-step function calling — for example, a voice assistant that looks up your calendar, checks your email, and books a restaurant, all in a single spoken exchange in a noisy room. That capability makes voice useful for real agentic workflows, not just simple Q&A. If you are evaluating this for production use, that is the number to test against your own use cases.


5. Google launched Lyria 3 Pro for developers

Build with Lyria 3, our newest music generation model - Google

Lyria 3 Pro: Create longer tracks in more Google products - Google

Lyria 3 (the base model) launched in February 2026. On March 25, Google added Lyria 3 Pro — the extended-capability tier. Both are now in public preview for developers worldwide via the Gemini API and Google AI Studio.

The two tiers are designed for different use cases:

  • Lyria 3 (base model): Fast and efficient, generates 30-second audio clips. Best for rapid prototyping, background loops, and social content.
  • Lyria 3 Pro: Generates full tracks up to 3 minutes. The key difference is structural awareness — you can specify an intro, verse, chorus, bridge, and outro directly in your prompt, and the model understands song structure rather than producing one undifferentiated audio block.

Both tiers support tempo conditioning, time-aligned lyrics (the model can synchronize generated vocals to a beat), multimodal image-to-music input (describe or show an image and generate matching music), and realistic vocal generation across many languages and genres. All output is SynthID-watermarked.

Lyria 3 Pro is also available on Vertex AI for enterprise-scale audio generation, in Google Vids (rolling out to Workspace users the week of March 25), and through ProducerAI — a collaborative music production tool Google recently acquired.

Takeaway: Music generation is moving from standalone novelty to platform feature, subject to the same infrastructure and policy constraints as every other AI capability.

Why this matters to you

The pace of iteration here is striking: Lyria 2 launched in April 2025, Lyria 3 in February 2026, and Lyria 3 Pro just one month later in March 2026. This acceleration happens because the model quality bottleneck has been cleared — what remains is market readiness, which moves faster when distribution is already on a cloud platform. If you are building products that involve music, audio branding, or creative media, these APIs are now production-ready infrastructure rather than experimental tools.


6. Anthropic’s Mythos/Capybara model leaked — and Anthropic confirmed it is real

Details leak on Anthropic's "step-change" Mythos model - Techzine

Anthropic acknowledges testing new model after leak - Fortune (paywalled)

On March 26–27, Fortune and Techzine published details from internal Anthropic draft materials that became publicly accessible due to a content management system misconfiguration. The CMS defaulted all uploaded assets to public URLs unless manually changed — approximately 3,000 unpublished assets were exposed as a result. The leak was discovered by security researchers Roy Paz (LayerX Security) and Alexandre Pauwels (University of Cambridge), who notified Fortune. After being informed, Anthropic restricted access.

The leaked materials refer to two overlapping names for the same model: Mythos is the model name; Capybara is the tier name — a new fourth tier positioned above Opus in Anthropic’s existing Opus/Sonnet/Haiku hierarchy. The draft describes Capybara as “larger and more intelligent than our Opus models — which were, until now, our most powerful,” with scores “dramatically higher” than Claude Opus 4.6 on coding, academic reasoning, and cybersecurity benchmarks.

The cybersecurity dimension is the sharpest detail in the leak. The draft states the model is “currently far ahead of any other AI model in cyber capabilities” and warns that it “presages an upcoming wave of models that can exploit vulnerabilities in ways that far outpace the efforts of defenders.” Because of this, Anthropic’s rollout plan gives cyber defenders priority access first — the reasoning being that the organizations most likely to be targeted by adversaries using this capability should have a window to harden their systems before wider availability.

The leaked materials also disclosed a previously unreported incident: a Chinese state-sponsored group conducted a coordinated campaign using Claude Code to infiltrate approximately 30 organisations — including tech companies, financial institutions, and government agencies — before Anthropic detected and disrupted it.

In a statement to Fortune, an Anthropic spokesperson confirmed: “We’re developing a general-purpose model with meaningful advances in reasoning, coding, and cybersecurity. Given the strength of its capabilities, we’re being deliberate about how we release it. As is standard practice across the industry, we’re working with a small group of early access customers to test the model. We consider this model a step change and the most capable we’ve built to date.”

No release date has been announced.

Takeaway: The leak itself is less important than what it confirms: frontier labs are now making deliberate, sequenced deployment decisions based on documented threat models, not marketing windows.

Why this matters to you

Defender-first access is not a PR move — it is a practical safety strategy. It creates a window in which the organisations most likely to face AI-assisted attacks can prepare before adversaries have equivalent capabilities. Whether or not Mythos/Capybara ships broadly, the release logic it represents is likely to become standard practice for any model that scores highly on cybersecurity benchmarks. If you work in security, or if your organisation is a likely target, this kind of staged release timeline is something to build your threat modelling around — not just react to.


Closing Thoughts

What struck me most about this week is how much the conversation has shifted.

We used to talk almost exclusively about model capability — which model is smarter, which benchmark leads. That still matters. But the more durable questions are becoming: Can we power it? Where can we legally deploy it? And who gets access first, and why?

The answer to all three is becoming more complicated. Energy demand for AI is scaling faster than linearly, and it is now running directly into hard limits — grid capacity, permitting politics, and deliberate deployment gating shaped by safety concerns. Deployment capacity is becoming the real constraint, not model quality alone.

Teams that plan for both tracks — capability and the constraints on where and how it can ship — will move faster than those who focus on capability alone.

Did you find this useful? I would love to hear your thoughts. Let me know if you have comments or suggestions!

desktop bg dark

About Elena

Elena, a PhD in Computer Science, simplifies AI concepts and helps you use machine learning.

Citation
Elena Daehnhardt. (2026) 'AI's New Bottleneck', daehnhardt.com, 27 March 2026. Available at: https://daehnhardt.com/blog/2026/03/27/ai-s-new-bottleneck-power-policy-and-persistent-agents/
All Posts