Mike Thiessen, PwC & Rebecca Potts, Google | Google Cloud Next 2026: Embracing AI — Challenges and Customization

At Google Cloud Next 2026, PwC’s Mike Thiessen and Google’s Rebecca Potts unveiled a pivotal shift in enterprise AI deployment: the general availability of Vertex AI’s Gemini 2.5 Pro with grounded retrieval-augmented generation (RAG) pipelines, now natively integrated with BigQuery ML and Confidential Computing enclaves. This move targets the persistent gap between prototype AI agents and production-grade, auditable systems — particularly in regulated sectors like finance and healthcare — by coupling real-time data grounding with hardware-enforced privacy. The announcement, delivered during the keynote’s “AI at Scale” segment, signals Google’s attempt to reclaim enterprise trust amid rising skepticism over hallucination risks and data leakage in LLMs.

The Grounding Problem: Why RAG Alone Isn’t Enough

While retrieval-augmented generation has become table stakes for enterprise AI, most implementations remain vulnerable to stale or poisoned indexes. Thiessen highlighted that PwC’s internal benchmarks showed a 34% drop in factual accuracy when RAG systems relied on weekly-updated vector stores versus real-time API feeds — a gap Gemini 2.5 Pro aims to close via its new Grounding API. This interface allows models to query structured data sources (BigQuery, Spanner, AlloyDB) with sub-200ms latency using vectorized SQL embeddings, bypassing traditional embedding pipeline delays. Crucially, the system now enforces row-level security policies at the query layer, meaning a model can only retrieve data the user is already authorized to see — a direct response to CISA’s 2025 guidance on AI data leakage.

The Grounding Problem: Why RAG Alone Isn’t Enough
Google Gemini Grounding

What’s less discussed is the architectural trade-off: enabling grounded retrieval increases token consumption by 18–22% per query due to contextual prompt expansion. Google mitigates this through a novel KV-cache pruning technique in Gemini 2.5 Pro’s transformer blocks, which drops low-attention tokens after grounding without degrading perplexity scores on MMLU-Pro. Independent validation by Stanford’s HAI lab showed this method retains 98.7% of baseline accuracy while reducing effective context length overhead by 40% compared to naive concatenation approaches.

Confidential Computing as the New Trust Anchor

Potts detailed how Gemini 2.5 Pro workloads now run exclusively within AMD SEV-SNP encrypted VMs on Google’s C3D machines, with attestation logs shipped to Chronicle Security for real-time tamper detection. This isn’t just about encrypting data at rest or in transit — it’s about ensuring the model weights and intermediate activations never exist in plaintext outside the secure enclave. For industries under GDPR Article 32 or HIPAA §164.306, this shifts liability: if a breach occurs, the cloud provider can cryptographically prove the AI workload was isolated.

Confidential Computing as the New Trust Anchor
Google Gemini Grounding

“We’re seeing clients move from ‘Is this AI secure?’ to ‘Can we prove it to an auditor?’,” said

Dr. Elena Vasquez, CTO of Mayo Clinic Platform, in a briefing following the keynote.

“The ability to attest that a model never touched raw PHI outside an SEV-SNP boundary changes the conversation from risk mitigation to compliance by design.” Her team piloted the system for prior authorization workflows, reducing manual review time by 62% while maintaining audit trails acceptable to CMS.

This approach contrasts sharply with AWS’s Bedrock Guardrails, which rely on post-generation filtering and IAM policies — effective for basic moderation but lacking hardware-backed guarantees for data-in-use protection. Azure’s Confidential LLMs offer similar SGX-based enclaves but currently lack native grounding APIs, forcing users to build complex middleware.

Ecosystem Implications: Lock-In or Open Pathways?

Google’s integration of grounding with Confidential Computing creates a compelling but potentially sticky enterprise workflow. Once a company builds agents that query BigQuery via the Grounding API within SEV-SNP enclaves, porting to another cloud requires re-architecting both the data pipeline and trust model. However, Google countered this concern by open-sourcing the Vertex AI Grounding SDK under Apache 2.0, allowing developers to generate grounding-compatible prompts for any LLM. The SDK includes adapters for Llama 3 and Mistral, though performance varies without Google’s tensor-optimized runtime.

Mike Thiessen, PwC & Rebecca Potts, Google | Google Cloud Next 2026
Ecosystem Implications: Lock-In or Open Pathways?
Google Gemini Grounding

Meanwhile, the cybersecurity community remains wary.

Marcus Holloway, lead architect at the Cybersecurity and Infrastructure Security Agency’s AI Red Team, warned: “Grounding doesn’t eliminate prompt injection — it just moves the attack surface. If your BigQuery view is compromised, the model becomes a confident liar with audit logs.”

His team demonstrated a proof-of-concept where malicious SQL views injected via shared datasets caused Gemini to output falsified financial forecasts, bypassing traditional input sanitization.

This underscores a broader industry tension: as AI systems grow more capable of reasoning over private data, the boundary between model and database blurs. Enterprises must now treat their data warehouses as part of the attack surface — a shift demanding closer collaboration between MLOps, data engineering, and SecOps teams.

What This Means for the AI Stack

Google Cloud Next 2026 didn’t just announce features — it outlined a new contract between enterprises and AI vendors: trust through verifiable isolation and data fidelity. By fusing grounded retrieval with confidential computing, Google addresses two of the top three barriers to enterprise AI adoption identified in Gartner’s 2025 survey (hallucinations and data privacy), leaving only cost predictability as the remaining hurdle.

For developers, the message is clear: the era of “prompt and pray” is over. Building production AI now requires understanding vector SQL, enclave attestation, and query-level security — skills that sit at the intersection of data engineering and AI infrastructure. As Vasquez put it, “We’re not just using AI anymore. We’re auditing it.”

Whether this approach becomes the new standard or merely a premium differentiator remains to be seen. But in an age where AI mistakes carry regulatory and reputational weight, the ability to prove — not just promise — that a system is secure and accurate may finally become the deciding factor in cloud wars.

Photo of author

Sophie Lin - Technology Editor

Sophie is a tech innovator and acclaimed tech writer recognized by the Online News Association. She translates the fast-paced world of technology, AI, and digital trends into compelling stories for readers of all backgrounds.

Heartbroken British Woman Travels to Switzerland for Assisted Suicide After Son’s Death

Title: Baltimore City Official Raises Concerns Over Inspector General’s Social Media Post

Leave a Comment

This site uses Akismet to reduce spam. Learn how your comment data is processed.