AI Daily Brief

Technology

Latest Intelligence

The latest AI stories, analysis and developments relevant to Technology — curated daily by Best Practice AI.

Use Case Library

Use Casesfor Technology

200 articles

PaywallTechnologyEconomics & Markets
WSJ· Yesterday

How a Job at OpenAI Became the Greatest Lottery Ticket of the AI Boom

Employees waited two years to sell their shares. Then, the company let them unload $30 million.

PaywallTechnologyEconomics & Markets
Bloomberg· Yesterday

Goldman Sees ‘AI-Driven Super Surplus’ Swelling in Korea, Taiwan

South Korea and Taiwan’s artificial intelligence-fueled chip booms are set to swell both economies’ current-account surpluses to fresh records and pressure their central banks to raise interest rates later this year.

PaywallTechnologyEconomics & Markets
Bloomberg· Yesterday

JPMorgan Hikes Kospi Bull Case Target to 10,000 on Memory Boom

JPMorgan Chase & Co. raised its targets for South Korean stocks for the second time in less than a month, citing improvement in the semiconductor cycle, corporate governance reforms and industrial-sector growth.

Technology
Voice Of Emirates· Yesterday

Intel’s Sudden Resurgence Reshuffles Chip Market and Sparks Competition Questions

Intel made a sudden comeback in the chip market through new AI strategies, restoring investor confidence and raising questions about its future rivalry with NVIDIA and AMD.

TechnologyTechnology & Infrastructure
Arxiv· Yesterday

Extracting Search Trees from LLM Reasoning Traces Reveals Myopic Planning

arXiv:2605.06840v1 Announce Type: new Abstract: Large language models (LLMs), especially reasoning models, generate extended chain-of-thought (CoT) reasoning that often contains explicit deliberation over future outcomes. Yet whether this deliberation constitutes genuine planning, how it is structured, and what aspects of it drive performance remain poorly understood. In this work, we introduce a new method to characterize LLM planning by extracting and quantifying search trees from reasoning traces in the four-in-a-row board game. By fitting computational models on the extracted search trees, we characterize how plans are structured and how they influence move decisions. We find that LLMs' search is shallower than humans', and that performance is predicted by search breadth rather than depth. Most strikingly, although LLMs expand deep nodes in their traces, their move choices are best explained by a myopic model that ignores those nodes entirely. A causal intervention study where we selectively prune CoT paragraphs further suggests that move selection is driven predominantly by shallow rather than deep nodes. These patterns contrast with human planning, where performance is driven primarily by deep search. Together, our findings reveal a key difference between LLM and human planning: while human expertise is driven by deeper search, LLMs do not act on deep lookahead. This dissociation offers targeted guidance for aligning LLM and human planning. More broadly, our framework provides a generalizable approach for interpreting the structure of LLM planning across strategic domains.

TechnologyAdoption & Impact
Arxiv· Yesterday

What if AI systems weren't chatbots?

arXiv:2605.07896v1 Announce Type: new Abstract: The rapid convergence of artificial intelligence (AI) toward conversational chatbot interfaces marks a critical moment for the industry. This paper argues that the chatbot paradigm is not a neutral interface choice, but a dominant sociotechnical configuration whose widespread adoption reshapes social, economic, legal, and environmental systems. We examine how treating AI primarily as conversational assistants has extensive structural downsides. We show how chatbot-based systems often fail to adequately meet user needs, particularly in complex or high-stakes contexts, while projecting confidence and authority. We further analyze how the normalization of chatbot-mediated interaction alters patterns of work, learning, and decision-making, contributing to deskilling, homogenization of knowledge, and shifting expectations of expertise. Finally, we examine broader societal effects, including labor displacement, concentration of economic power, and increased environmental costs driven by sustained investment in large-scale chatbot infrastructures. While acknowledging legitimate benefits, we argue that the current trajectory of AI development reflects specific value choices that prioritize conversational generality over domain specificity, accountability, and long-term social sustainability. We conclude by outlining alternative directions for AI development and governance that move beyond one-size-fits-all chatbots, emphasizing pluralistic system design, task-specific tools, and institutional safeguards to mitigate social and economic harm.

PaywallTechnology
Bloomberg· Yesterday

Emerging Stocks Set to Close at Record High on Tech Bets

Emerging-market equities rose on AI trades as investors brushed off concerns over stalled peace talks between the US and Iran.

TechnologyEconomics & Markets
Arxiv· Yesterday

General-Purpose Technology and Speculative Bubble Detection

arXiv:2604.25826v2 Announce Type: replace Abstract: We show that the leading bubble test suffers severe size distortion when fundamentals incorporate general-purpose technology adoption. Embedding a hump-shaped technology shock in the Campbell-Shiller present-value model, we prove that the fundamental price becomes locally explosive during adoption, contaminating the test's limit distribution with a non-centrality parameter proportional to the shock's peak. We propose a fundamental-versus-speculative decomposition that projects prices onto observable technology proxies and applies the test to the residual. Empirically, the decomposition eliminates evidence of speculation in the 2020-2025 AI rally while confirming a speculative peak confined to December 1999-March 2000 in the dot-com episode.

PaywallTechnologyEconomics & Markets
Bloomberg· Yesterday

Alphabet Plans Debut Yen Bond Sale as AI Race Accelerates

Alphabet Inc. is planning to issue yen bonds for the first time in a move that may help fund investments as artificial intelligence competition intensifies.

TechnologyEconomics & Markets
Siliconrepublic· Yesterday

Europe’s $17bn tech quarter: AI, deep tech win as fintech slides

European start-ups raised a two-year high of $17bn in Q1 2026, with AI, enterprise applications and deep tech the big winners. Read more: Europe’s $17bn tech quarter: AI, deep tech win as fintech slides

TechnologyGeopolitics
Reuters· Yesterday

Reuters Tech News | Today's Latest Technology News | Reuters

categoryThailand's Siam AI denies exporting US AI servers to China · May 9, 2026 · BusinesscategoryAnthropic signs $1.8 billion AI cloud deal with Akamai, Bloomberg News reports · May 8, 2026 · SustainabilitycategoryAmazon's Chile data center moves ahead ·

Technology
Let's Data Science· Yesterday

Workers Confront AI-Driven Mass Layoffs | Let's Data Science

According to reporting by the World Socialist Web Site, tech security firm Cloudflare announced layoffs of **20 percent** of its workforce, about **1,100** employees, on May 7, 2026; WSWS cites CEO Matthew Prince calling it "the agentic AI era" and quoting him, "Just because you're fit doesn't ...

TechnologyTechnology & Infrastructure
Arxiv· Yesterday

Towards Security-Auditable LLM Agents: A Unified Graph Representation

arXiv:2605.06812v1 Announce Type: new Abstract: LLM-based agentic systems are rapidly evolving to perform complex autonomous tasks through dynamic tool invocation, stateful memory management, and multi-agent collaboration. However, this semantics-driven execution paradigm creates a severe semantic gap between low-level physical events and high-level execution intent, making post-hoc security auditing fundamentally difficult. Existing representation mechanisms, including static SBOMs and runtime logs, provide only fragmented evidence and fail to capture cognitive-state evolution, capability bindings, persistent memory contamination, and cascading risk propagation across interacting agents. To bridge this gap, we propose Agent-BOM, a unified structural representation for agent security auditing. Agent-BOM models an agentic system as a hierarchical attributed directed graph that separates static capability bases, such as models, tools, and long-term memory, from dynamic runtime semantic states, such as goals, reasoning trajectories, and actions. These layers are connected through semantic edges and security attributes, transforming fragmented execution traces into queryable audit paths. Building on Agent-BOM, we develop a graph-query-based paradigm for path-level risk assessment and instantiate it with the OWASP Agentic Top 10. We further implement an auditing plugin in the OpenClaw environment to construct Agent-BOM from live executions. Evaluation on representative real-world agentic attack scenarios shows that Agent-BOM can reconstruct stealthy attack chains, including cross-session memory poisoning and tool misuse, capability supply-chain hijacking and unexpected code execution, multi-agent ecosystem hijacking, and privilege and trust abuse. These results demonstrate that Agent-BOM provides a unified and auditable foundation for root-cause analysis and security adjudication in complex agentic ecosystems.

TechnologyTechnology & Infrastructure
Daily Brew· Yesterday

AI tool poisoning exposes a major flaw in enterprise agent security

Researchers have identified a significant security vulnerability in enterprise AI agents caused by tool poisoning.

TechnologyTechnology & Infrastructure
Arxiv· Yesterday

Theoretical Limits of Language Model Alignment

arXiv:2605.07105v1 Announce Type: cross Abstract: Language model (LM) alignment improves model outputs to reflect human preferences while preserving the capabilities of the base model. The most common alignment approaches are (i) reinforcement learning, which maximizes the expected reward under a KL-divergence constraint, and (ii) best-of-$N$ alignment, which selects the highest-reward output among $N$ independent samples. Despite their widespread use, the fundamental limits of reward improvement under a KL budget remain poorly understood. We characterize the information-theoretic limits of KL-regularized alignment by deriving the maximum achievable expected reward gain for a fixed KL-divergence budget. Our first result provides a closed-form expression for the optimal reward improvement, governed by a Jeffreys divergence term rather than the $\sqrt{\texttt{KL}}$ used in prior analyses. We further reformulate this expression as a covariance under the base model, yielding a practical estimator that predicts achievable alignment gains from base model samples alone. We extend our analysis to the proxy reward setting, showing that the gap between ideal and proxy alignment (reward hacking) grows with the magnitude of reward error and when the KL penalty factor decreases. We then prove that reward ensembling mitigates reward hacking, providing a theoretical justification for this technique used in practice. Empirically, we compute the KL-reward Pareto frontier for two tasks for LMs, safety and summarization, and show that best-of-$N$ closely approaches the theoretical limit, while PPO and GRPO remain substantially suboptimal. Our theoretical results shed light on several empirically observed phenomena in the alignment literature and suggest that algorithmic improvements are needed to achieve optimal alignment without high inference costs.

Technology
Daily Brew· Yesterday

AI-Driven Bots Surge, Dominating 53% of Web Traffic in 2025

AI-driven bot activity skyrocketed by 12.5 times in 2025, posing significant security challenges as bots now represent over half of web traffic.

TechnologyTechnology & Infrastructure
Help Net Security· Yesterday

Security teams are turning to AI to survive alert overload - Help Net Security

Cybersecurity teams are expanding AI adoption across threat detection, incident response and security operations workflows.

Technology
Tech AI Magazine· Yesterday

Generative AI 2026: Core Trends and What Matters Next

Explore generative AI 2026 trends as models shift from basic tools to autonomous agentic infrastructure driving enterprise workflows.

TechnologyAdoption & Impact
Business Insider· Yesterday

The Sneaky Rise of Shadow AI in the Workplace - Business Insider

Gregg Bayes-Brown helped develop the AI policies at a former job in biotech research, but even as a rule maker, he says he couldn't afford to not be a rule breaker. Though he understood the technology and its risks, he used an unapproved personal enterprise Google account for work, to access ...

TechnologyEconomics & Markets
Arxiv· Yesterday

Vibecoding and Digital Entrepreneurship

arXiv:2511.06545v2 Announce Type: replace Abstract: As generative artificial intelligence (GenAI) automates coding tasks and expands access to technical resources, this paper examines how GenAI-enabled coding automation, colloquially known as "vibecoding," affects digital entrepreneurial entry and venture performance. We exploit ex-ante variation in ventures' exposure to vibecoding based on the product characteristics of their initial launches and estimate difference-in-differences models around the diffusion of GenAI coding tools. Vibecoding increases first-time launches and shortens time to launch, but economically viable entry rises only where vibecoding augments, rather than fully automates, product development. In these partially exposed product segments, viable entry increases by 11%, driven entirely by ventures founded by individuals with STEM education or work experience, especially those whose most recent employment was outside middle management. Among ventures launched before GenAI became widely accessible, performance gains similarly concentrate among partially exposed ventures with engineering-intensive initial teams. Together, these results suggest that GenAI-enabled coding automation does not eliminate the value of technical expertise. Instead, vibecoding creates the greatest value when it complements internal engineering capabilities, allowing ventures to delegate lower-level coding tasks to GenAI while shifting human effort toward higher-level problem solving and dynamic adaptation.

TechnologyEconomics & Markets
Daily Brew· Yesterday

Backblaze Q1 2026: Revenue Up 12%, AI Traction Boosts Stock 70%

Backblaze reported a 12% revenue increase for Q1 2026, driven by a 24% surge in its B2 Cloud Storage segment as it pivots to capitalize on AI opportunities.

TechnologyEconomics & Markets
Daily Brew· Yesterday

Microsoft Emails Reveal 2018 Doubts on OpenAI Partnership Amid Amazon Pivot Concerns

Court filings reveal that Microsoft executives expressed early doubts in 2018 about OpenAI's path to AGI and the risk of a pivot to Amazon.

TechnologyEconomics & Markets
Siliconrepublic· Yesterday

Quantinuum files for IPO as quantum stocks gain popularity

McKinsey finds quantum companies generated more than $1bn in revenue in 2025. Read more: Quantinuum files for IPO as quantum stocks gain popularity

TechnologyTechnology & Infrastructure
Arxiv· Yesterday

GraphDC: A Divide-and-Conquer Multi-Agent System for Scalable Graph Algorithm Reasoning

arXiv:2605.06671v1 Announce Type: new Abstract: Large Language Models (LLMs) have demonstrated strong potential for many mathematical problems. However, their performance on graph algorithmic tasks is still unsatisfying, since graphs are naturally more complex in topology and often require systematic multi-step reasoning, especially on larger graphs. Motivated by this gap, we propose GraphDC, a Divide-and-Conquer multi-agent framework for scalable graph algorithm reasoning. Specifically, inspired by Divide-and-Conquer design, GraphDC decomposes an input graph into smaller subgraphs, assigns each subgraph to a specialized agent for local reasoning, and uses a master agent to integrate the local outputs with inter-subgraph information to produce the final solution. This hierarchical design reduces the reasoning burden on individual agents, alleviates computational bottlenecks, and improves robustness on large graph instances. Extensive experiments show that GraphDC consistently outperforms existing methods on graph algorithm reasoning across diverse tasks and scales, especially on larger instances where direct end-to-end reasoning is less reliable.

TechnologyTechnology & Infrastructure
Arxiv· Yesterday

SARC: A Governance-by-Architecture Framework for Agentic AI Systems

arXiv:2605.07728v1 Announce Type: cross Abstract: Agentic AI systems increasingly act through tools, sub-agents, and external services, but governance controls are still commonly attached to prompts, dashboards, or post-hoc documentation. This creates a structural mismatch in regulated settings: obligations that must constrain execution are often evaluated only after execution has occurred. We introduce SARC, a runtime governance architecture for tool-using agents that treats constraints as first-class specification objects alongside state, action space, and reward. A SARC specification declares each constraint's source, class, predicate, verification point, response protocol, and operating point, and compiles these into four enforcement sites in the agent loop: a Pre-Action Gate, an Action-Time Monitor, a Post-Action Auditor, and an Escalation Router. We formalize the minimal invariants required for specification-trace correspondence, show why finite reward penalties do not generally substitute for hard runtime constraints, and extend the architecture to multi-agent workflows through constraint propagation, authority intersection, and attribution-preserving trace trees. We implement a prototype audit checker and report a reproducible synthetic evaluation over 50 seeds comparing SARC against post-hoc audit, output filtering, workflow rules, and policy-as-code-only baselines on a procurement task. SARC executes zero hard-constraint violations under exact predicates; its declared PAA throttling response reduces soft-window overages by 89.5% relative to policy-as-code-only. Predicate-noise and enforcement-failure sweeps are consistent with the claim that residual hard violations under SARC scale with enforcement-stack error rather than environmental violation opportunity. SARC provides the architectural substrate through which obligations can be made executable, inspectable, and auditable at runtime.

TechnologyTechnology & Infrastructure
Arxiv· Yesterday

Weblica: Scalable and Reproducible Training Environments for Visual Web Agents

arXiv:2605.06761v1 Announce Type: new Abstract: The web is complex, open-ended, and constantly changing, making it challenging to scale training data for visual web agents. Existing data collection attempts remain limited to offline trajectories for supervised fine-tuning or a handful of simulated environments for RL training, thus failing to capture web diversity. We propose Weblica (Web Replica), a framework for constructing reproducible and scalable web environments. Our framework leverages 1) HTTP-level caching to capture and replay stable visual states while preserving interactive behavior and 2) LLM-based environment synthesis grounded in real-world websites and core web navigation skills. Using this framework, we scale RL training to thousands of diverse environments and tasks. Our best model, Weblica-8B, outperforms open-weight baselines of similar size across multiple web navigation benchmarks while using fewer inference steps, scales favorably with additional test-time compute, and is competitive with API models.

TechnologyTechnology & Infrastructure
Arxiv· Yesterday

Social Theory Should Be a Structural Prior for Agentic AI: A Formal Framework for Multi-Agent Social Systems

arXiv:2605.07069v1 Announce Type: cross Abstract: Agentic AI systems are increasingly deployed not in isolation, but inside social environments populated by other agents and humans, such as in social media platforms, multi-agent LLM pipelines or autonomous robotics fleets. In these settings, system behavior emerges not from individual agents alone, but from the multi-agent interactions over time. Emergent dynamics of individuals in a social group have been long studied by social scientists in human contexts. \textbf{This position paper argues that agentic AI systems must be modeled with social theory as a structural prior, and formalizes a Multi-Agent Social Systems (MASS) framework for how agents interact and influence to generate system-level outcomes.} We represent MASS as a class of dynamical system of information generation, local influence and interaction structure, formulated by four structural priors anchored in social theory: strategic heterogeneity, networked-constrained dependence, co-evolution and distributional instability. We demonstrate the importance of each structural prior through formal propositions, and articulate a research agenda for how MASS should be modeled, evaluated and governed.

TechnologyTechnology & Infrastructure
Arxiv· Yesterday

AGWM: Affordance-Grounded World Models for Environments with Compositional Prerequisites

arXiv:2605.06841v1 Announce Type: new Abstract: In model-based learning, the agent learns behaviors by simulating trajectories based on world model predictions. Standard world models typically learn a stationary transition function that maps states and actions to next states, when an action and an outcome frequently co-occur in training data, the model tends to internalize this correlation as a general causal rule while ignoring action preconditions. In interactive environments, however, agent actions can reshape the future affordance space. At each timestep, an action may becomes executable only after its prerequisites are met, or non-executable when they are destroyed. We term such events structure-changing events (SC events). As a result, a conventional world model often fails to determine whether a given action is executable in the current state, especially in multi-step predictions. Each imagined step is conditioned on an incorrect affordance state, and therefore the prediction error compounds over the rollout horizon. In this paper, we propose AGWM (Affordance-Grounded World Model), which learns an abstract affordance structure represented as a DAG of prerequisite dependencies to explicitly track the dynamic executability of actions. Experiments on game-based simulated environments demonstrate the effectiveness of our method by achieving lower multi-step prediction error, better generalization to novel configurations, and improved interpretability.

TechnologyAdoption & Impact
Daily Brew· 2d ago

Beever Atlas Revolutionizes Team Chats into Secure, Structured Knowledge Graphs for Enterprises

Beever Atlas transforms team chats into structured Neo4j knowledge graphs, offering high-security features like on-premise deployment and AES-256-GCM encryption.

TechnologyAdoption & Impact
Daily Brew· 2d ago

RAG Is Blind to Time: I Built a Temporal Layer to Fix It in Production

The author introduces a temporal layer for Retrieval-Augmented Generation (RAG) systems to address the challenge of time-sensitive data.

Technology
Simply Wall St· 2d ago

Why NVIDIA (NVDA) Is Up 8.4% After New AI Infrastructure Alliances With IREN and Corning – And What's Next - Simply Wall St News

In early May 2026, NVIDIA announced new alliances with IREN and Corning to deploy up to 5 gigawatts of DSX-aligned AI infrastructure and to massively expand U.S. production of advanced optical connectivity for next-generation AI data centers. These deals signal NVIDIA’s push to secure critical ...

Technology
Daily Brew· 2d ago

Nvidia's $40 Billion AI Investment Surge: Strategic Moves and Market Reactions

Nvidia has committed over $40 billion in 2026 to strengthen the AI supply chain, focusing on ecosystem growth and hardware demand.

Technology
Daily Brew· 2d ago

Alphabet's AI Dominance: Financial Strength and Strategic Partnerships Propel Growth

Alphabet reported a 22% year-over-year revenue increase in Q1 2026, fueled by strong AI integration and key partnerships like those with Apple.

TechnologyEconomics & Markets
Reuters· 2d ago

Reuters AI News | Latest Headlines and Developments | Reuters

Strong demand for AI computing equipment in China has nearly doubled prices for Nvidia's B300 ​servers to about 7 million yuan ($1 million) each, industry sources said, as a crackdown on chip smuggling dries up ‌black-market supply.

TechnologyTechnology & Infrastructure
Daily Brew· 2d ago

Intent-based chaos testing is designed for when AI behaves confidently and wrongly

A new approach to chaos testing helps developers identify and mitigate risks when AI models provide confident but incorrect outputs.

PaywallTechnology
FT· 2d ago

OpenAI trial lays bare rivalries behind start-up’s $852bn rise

Lawsuit brought by Elon Musk heads into final week in court, with Sam Altman due to testify

PaywallTechnologyGeopolitics
Bloomberg· 2d ago

Microsoft’s African Data Center Falters on Payment Demands

A major Microsoft Corp. data center site in East Africa has been delayed by disagreements with the Kenyan government over the company’s request for guaranteed payments, people familiar with the matter said.

TechnologyTechnology & Infrastructure
Theregister· 2d ago

Memory godboxes could offer relief from the RAMpocalypse

Amid the AI-fueled memory crunch, will Compute Express Link finally have its moment to shine?

TechnologyTechnology & Infrastructure
Theregister· 2d ago

Yes, local LLMs are ready to ease the compute strain

Anthropic might be thinking about space to ease its computing burden, but Claude Code on your laptop is way more practical

TechnologyTechnology & Infrastructure
Digitpatrox· 2d ago

How AI Agents Could Replace SaaS Software by 2030

Microsoft Copilot Studio: Building internal agents across the entire Microsoft 365 ecosystem. Claude “Computer Use”: Anthropic’s latest capability allows AI to see a screen and move a cursor. This is part of the new Claude AI handoff workflow designed for seamless automation. See also The Death of the Browser Tab: How AI Browsers Are Changing Search · To understand the business impact...

TechnologyTechnology & Infrastructure
Khaleej Times· 2d ago

AI in cybersecurity: Smarter defence or a new generation of blind spots? | Khaleej Times

As UAE organisations automate cyber defence, experts warn AI can cut workloads but also hide missed threats — raising questions over visibility, governance and human oversight

TechnologyTechnology & Infrastructure
Daily Brew· 2d ago

Claude Haiku 4.5 Achieves Near-Perfect Alignment, Eliminates Blackmail Risks in AI Models

Anthropic's Claude Haiku 4.5 model has reached near-perfect alignment, significantly reducing blackmail tendencies through advanced ethical training.

TechnologyEconomics & Markets
Let's Data Science· 2d ago

Microsoft Posts Strong Enterprise AI and Cloud Revenue Gains | Let's Data Science

Insider Monkey reports that Microsoft Corporation's fiscal third-quarter results, released April 29, 2026, showed continued strength in enterprise cloud and AI-linked revenue. Per Insider Monkey, **Microsoft Cloud** revenue rose **29%** to **$54.5 billion**, **Azure and other cloud services** ...

TechnologyEconomics & Markets
Medium· 2d ago

💰 10 Methods To Save Money On Agentic Engineering — From $5 to $0.17 Per Request | by Tom Smykowski | May, 2026 | Medium

The problem hit me when I noticed my Cursor bills climbing to $5 per request. After investigating, I found the IDE was wasting tokens on cache reads that provided zero benefit. But even after reporting the bug, I realized something bigger: most AI coding tools aren’t designed for cost efficiency.

Technology
Mid-day· 3d ago

Cloudflare announces major layoffs as AI reshapes company operations

US-based connectivity cloud company Cloudflare has announced plans to eliminate more than 1,100 jobs globally as part of a major organisational restructuring

Technology
Technosports· 3d ago

Microsoft AI Power Consumption and Clean Energy Strategy

Verdict: Microsoft aims for a 50% reduction in AI infrastructure carbon footprint by 2030 through algorithmic optimization and renewable energy integration. Stakeholders have praised this initiative. Coverage from outlets like VentureBeat often overlooks the real technical challenges of scaling ...

TechnologyAdoption & Impact
New Kerala· 3d ago

AI Focus Boosts EBITDA by 20%: McKinsey

Levin said many companies fail ... redesigning workflows and operations from end to end. "One of the first things businesses miss is that these AI transformations need to be entirely business-led," he said. The podcast also highlighted how AI is changing software development ...

Technology
Daily Brew· 3d ago

Anthropic says it hit a $30 billion revenue run rate after 'crazy' 80x growth

Anthropic reports massive financial growth, reaching a $30 billion revenue run rate following an 80x increase.

Technology
CNBC· 3d ago

Nvidia embraces AI investor, topping $40 billion in equity bets 2026

Nvidia is pouring billions of dollars at a time into companies across the AI infrastructure stack, while also signing commercial deals with them.

PaywallTechnology
Bloomberg· 3d ago

Chinese Export Growth Rebounds as War Fails to Curb Trade

China’s export growth rebounded more than expected despite disruptions to shipping caused by the war in Iran, as trade volumes swell due to an investment boom in artificial intelligence.

TechnologyEconomics & Markets
Daily Brew· 3d ago

Elon Musk called Anthropic 'evil' 3 months ago. Now he's taking $4 billion to become its landlord

Elon Musk is set to receive $4 billion from Anthropic for real estate, despite previously labeling the company as 'evil'.

Technology
Daily Brew· 3d ago

Microsoft was worried OpenAI would run off to Amazon and 'shit-talk' Azure

Internal reports reveal Microsoft's concerns that OpenAI might leverage its relationship with Amazon to criticize Azure's infrastructure.

TechnologyEconomics & Markets
Daily Brew· 3d ago

Intel's comeback story is even wilder than it seems

An analysis of Intel's recent strategic shifts and the surprising developments in its ongoing recovery efforts.

TechnologyEconomics & Markets
Daily Brew· 3d ago

Apple Raises Price Target Amid AI Expansion

Apple's stock target rises to $400, driven by AI momentum, with an Outperform rating as the company prepares to unveil its AI roadmap at WWDC.

PaywallTechnologyEconomics & Markets
Bloomberg· 3d ago

ByteDance Targets 25% Rise in AI Infrastructure Spending: SCMP

ByteDance Ltd. has boosted planned spending on artificial intelligence infrastructure this year by 25% to 200 billion yuan ($29.4 billion), as memory chip costs rise and the TikTok owner ramps up its AI presence, the South China Morning Post reported on Saturday.

Technology
National Review· 3d ago

China AI Industry: U.S. Export Controls Undermine Technological Edge | National Review

Nvidia’s true advantage was never ... into its chips. Export controls pushed Chinese developers off CUDA and into Huawei’s CANN ecosystem. That created short-term pain, but it also accelerated the rise of a parallel AI software stack independent of America’s. This should force a rethink in Washington. China isn’t willing to sacrifice market share in order to indigenize technology. In fact, it readily accepts supply chain dependencies ...

Technology
Artiverse· 3d ago

Europe Delays Its Toughest AI Rules Amid Industry Pushback - Artiverse

European officials have decided to postpone some of their most ambitious AI regulations. They agreed to delay the implementation of certain high-risk AI rules until December 2027…

Technology
International Business Times· 3d ago

Why Europe's Landmark AI Law Now Looks Far Softer, Ban on 'Nudifier' Apps, Deepfakes Retained

The European Union (EU) has finally agreed on a set of artificial intelligence (AI) regulations. New rules help developers and startups to comply better. To ensure that high-risk artificial intelligence systems applications are protected, Law has established safeguards. International legislation might influence international standards ...

TechnologyGeopolitics
Let's Data Science· 3d ago

Google explores India investments in AI infrastructure | Let's Data Science

Editorial analysis: this reporting ... covered in April, and fits a broader trend of hyperscalers assessing onshore compute and hardware capacity in key markets. For practitioners, increased local investment in servers and AI infrastructure can influence procurement, latency-sensitive ...

TechnologyTechnology & Infrastructure
Globaldatacenterhub· 3d ago

Global Data Center Roundup – April 2026: Execution Era of AI Infrastructure

Join investors, operators, and ... latest trends in the data center sector in developed and emerging markets globally. ... This month’s stories reflect a market moving from expansion to selection. Infrastructure misalignment is collapsing unsynchronized projects, while neocloud growth and colocation fragmentation are shifting capacity toward specialized environments. Winners will be defined by execution control of power, alignment of capital with compute economics, ...

Technology
ETEnterpriseai.com· 4d ago

EU Reaches Provisional Agreement on AI Rules Amid Criticism from Critics and Tech Companies, ETEnterpriseai

The European Union has reached a provisional agreement on a significantly altered AI Act, delaying its implementation and responding to concerns from both critics and big tech. The agreement includes a ban on unauthorized intimate AI-generated content and aims to ease regulatory burdens for ...

Technology
Daily AI News May 8, 2026: OpenAI Turns Voice AI into Real Capability· 4d ago

Notes From Inside China's AI Labs

This post examines China's AI labs, highlighting their fast-following execution, strong talent pipelines, and cultural approaches to building large language models.

TechnologyAdoption & Impact
Anthropic Claude Office Integration ⚡, Google Chrome AI API 🌐, Nous He· 4d ago

Google ships Chrome's on-device AI API despite near-universal opposition

Google has released the Prompt API in Chrome, enabling websites to access local AI models. The move has faced criticism from browser makers over standards and privacy concerns.

Technology
DQC· 4d ago

AMD Instinct MI350P PCIe specs could change AI infrastructure

AMD Instinct MI350P PCIe specs reveal how AMD is targeting enterprise AI growth with air-cooled GPUs built for existing datacentre infrastructure.

TechnologyTechnology & Infrastructure
🗣️ OpenAI closes reasoning gap in voice agents· 4d ago

OpenAI closes reasoning gap in voice agents

OpenAI is working to improve the reasoning capabilities of its voice-based AI agents.

Technology
Yahoo! Finance· 4d ago

Layoffs Accelerate in May 2026 as Firms Restructure Around AI

Cloudflare, Upwork, Coinbase, and others cut thousands of jobs in May 2026 as AI restructuring reshapes the tech workforce.

Technology
Redhub· 4d ago

DeepSeek R4 for AI Agents: The Business Case - RedHub.ai

DeepSeek R4 for AI agents explained, including agent workflows, coding tasks, business automation, risks, and when to use it.

Technology
CNBC· 4d ago

Wall Street AI chip love moves from Nvidia to Intel, AMD and Micron

Intel, AMD and Micron surged double digits this week as investors bet on CPU makers and memory companies powering the next stage of AI

Technology
Sky9 Capital· 4d ago

AI-Focused VC Firms and Investors by Stage and Founder Fit (2026) - Sky9 Capital

AI startups captured 53% of all global venture capital in the first half of 2026. That concentration means more capital is available, but also more noise. Investors have sharpened their filters, particularly around defensibility: proprietary data, novel architectures, and deep workflow integration now separate credible pitches from thin wrappers on foundation models (source: vcsheet.com, May ...

Technology
TheStreet· 4d ago

Nvidia’s AI dominance faces uncomfortable new reality - TheStreet

The issue for Nvidia is not only that it’s selling fewer chips in China. The reality is that Chinese consumers and suppliers are being pushed to adjust. Domestic companies such Huawei, Cambricon, Moore Threads, and MetaX are trying to capture more of the market that Nvidia can no longer fulfill directly. But such businesses still face enormous barriers, including matching Nvidia’s whole hardware and software stack. But limits may alter consumer behavior quicker than regular competition ...

Technology
Ferguson Wellman· 4d ago

The Magnificent Capex: AI Infrastructure Spending and Who Actually Benefits — Ferguson Wellman

The numbers coming out of the first quarter earnings season were, by any historical standard, staggering. Amazon, Alphabet, Microsoft and Meta, the four largest cloud and technology platforms, accounted for $650 to $700 billion in capital expenditures (capex) for 2026. That is nearly double what the

Technology
Benzinga· 4d ago

AI Drove Two-Thirds Of Q1 2026 GDP Growth, Smashing '1999 Record' For Largest Tech Contribution 'In Histo - Benzinga

AI fueled a historic 67% of Q1 2026 GDP growth, smashing the 1999 record. Read how tech investments drove the economy amid bubble warnings.

Technology
NAI 500· 4d ago

8 China AI chip plays to watch after Kunlunxin’s filing | NAI 500

At the same time, domestic AI chip competition is heating up. Huawei is targeting dominance in China’s AI chip market by 2026, leaning into inference computing and planning a more advanced 950DT in Q4. With Nvidia’s top-end parts constrained by export controls, the domestic AI hardware market ...

Technology
Times of India· 4d ago

Valley new base camp for desi AI startups - The Times of India

India Business News: BENGALURU: For Indian AI founders targeting global markets, venture capital firms are increasingly pushing a clear message: move closer to the US mark.

Technology
Everything PR· 4d ago

Cybersecurity 2026: AI-Compressed Attacks, the SEC Disclosure Era, and the $32B Cloud-Security Reset - PR News

EPR Cybersecurity Intelligence tracks the threats, regulatory shifts, vendor consolidation, and communications dynamics reshaping enterprise security. This…

TechnologyAdoption & Impact
AI Business· 4d ago

AI Agents Are Becoming Operational Infrastructure

AI agents are moving from experimentation into enterprise infrastructure and operations, and organizations are now trying to figure out how to govern, secure and operationalize them. Agents are moving into real working roles, becoming embedded in workflows rather than remaining limited to demos. This week alone, we’ve seen several examples of that: AWS added new payment capabilities for autonomous agents, enabling AI systems ...

TechnologyAdoption & Impact
World Business Outlook· 4d ago

The Orchestration Era: Why the AI Agent Workspace is the Definitive Business Pivot of 2026 » World Business Outlook

The rise of AI Agent Workspaces marks a major shift in enterprise technology, combining automation, collaborative execution.

TechnologyAdoption & Impact
IT Brief Australia· 4d ago

Bloomfire unveils guide to enterprise intelligence systems

Bloomfire says fragmented knowledge, search and analytics tools are undermining enterprise AI, as its new guide assesses 12 platforms and governance.

TechnologyEconomics & Markets
Siliconrepublic· 4d ago

Quantexa opens new Dublin office and R&D centre

The centre will bring together a ‘growing team’ of researchers and data scientists, Quantexa said. Read more: Quantexa opens new Dublin office and R&D centre

TechnologyEconomics & Markets
The Motley Fool· 4d ago

The Best AI Chips Stocks to Buy Right Now in 2026 | The Motley Fool

The aggressive spending on AI data center infrastructure will be a tailwind for Intel, Sandisk, and Micron Technology over the long run.

Technology
Artificial Intelligence Newsletter | May 11, 2026· 4d ago

Amazon's battle to block Perplexity AI testing reach of computer-access laws

Amazon's lawsuit against Perplexity AI's shopping tool is testing the reach of computer-access laws. A US appeals court in Seattle will weigh in next month.

Technology
Daily Brew· 4d ago

Anthropic signs $1.8 billion AI cloud deal with Akamai

Anthropic has entered into a significant $1.8 billion cloud computing agreement with Akamai to support its AI infrastructure.

PaywallTechnologyEconomics & Markets
Bloomberg· 4d ago

Sony Partners With TSMC in Effort to Rein in Chipmaking Costs

Sony Group Corp. plans a new joint venture with Taiwan Semiconductor Manufacturing Co. as the first step toward what it called a fab-light approach to semiconductor production.

Technology
Yahoo! Finance· 4d ago

NVIDIA Expands AI Infrastructure From Utility Scale Builds To Homes

For investors, these moves highlight how AI compute may be supplied through a wider mix of models, from massive data center builds to smaller, grid aware clusters closer to where power is consumed. The IREN, Span, and PulteGroup partnerships add new potential channels for NVIDIA's AI platforms and could influence how capital, power capacity, and hardware are allocated across the AI infrastructure ...

Technology
Daily Brew· 4d ago

Anthropic Says It’s Buying 100% of Compute From xAI’s Colossus Data Center

Anthropic has entered a partnership to utilize the full compute capacity of xAI's Colossus data center.

Technology
VentureBeat· 4d ago

An AI agent rewrote a Fortune 50 security policy. Here's how to govern AI agents before one does the same.

A CEO’s AI agent rewrote the company’s security policy. Not because it was compromised, but because it wanted to fix a problem, lacked permissions, and removed the restriction itself. Every identity check passed. CrowdStrike CEO George Kurtz disclosed the incident and a second one at his RSAC 2026 keynote, both at Fortune 50 companies. The credential was valid. The access was authorized. The action was catastrophic. That sequence breaks the core assumption underneath the IAM systems most enterprises run in production today: that a valid credential plus authorized access equals a safe outcome. Identity systems were built for one user, one session, one set of hands on a keyboard. Agents break all three assumptions at once. In an exclusive interview with VentureBeat at RSAC 2026, Matt Caulfield, VP of Identity and Duo at Cisco, (pictured above) walked through the architecture his team is building to close that gap and outlined a six-stage identity maturity model for governing agentic AI. The urgency is measurable: Cisco President Jeetu Patel told VentureBeat at the same conference that 85% of enterprises are running agent pilots while only 5% have reached production — an 80-point gap that the identity work is designed to close. The identity stack was built for a workforce that has fingerprints “Most of the existing IAM tools that we have at our disposal are just entirely built for a different era,” Caulfield told VentureBeat. “They were built for human scale, not really for agents.” The default enterprise instinct is to shove agents into existing identity categories: human user; machine identity; pick one. "Agents are a third kind of new type of identity," Caulfield said. "They're neither human. They're neither machine. They're somewhere in the middle where they have broad access to resources like humans, but they operate at machine scale and speed like machines, and they entirely lack any form of judgment." Etay Maor, VP of Threat Intelligence at Cato Networks, put a number on the exposure. He ran a live Censys scan and counted nearly 500,000 internet-facing OpenClaw instances. The week before, he found 230,000, discovering a doubling in seven days. Kayne McGladrey, an IEEE senior member who advises enterprises on identity risk, made the same diagnosis independently. Organizations are cloning human user accounts to agentic systems, McGladrey told VentureBeat, except agents consume far more permissions than humans would because of the speed, the scale, and the intent. A human employee goes through a background check, an interview, and an onboarding process. Agents skip all three. The onboarding assumptions baked into modern IAM do not apply. Scale compounds the failure. Caulfield pointed to projections where a trillion agents could operate globally. “We barely know how many people are in an average organization,” he said, “let alone the number of agents.” Access control verifies the badge. It does not watch what happens next. Zero trust still applies to agentic AI, Caulfield argued. But only if security teams push it past access and into action-level enforcement. “We really need to shift our thinking to more action-level control,” he told VentureBeat. “What action is that agent taking?” A human employee with authorized access to a system will not execute 500 API calls in three seconds. An agent will. Traditional zero trust verifies that an identity can reach an application. It doesn’t scrutinize what that identity does once inside. Carter Rees, VP of Artificial Intelligence at Reputation, identified the structural reason. The flat authorization plane of an LLM fails to respect user permissions, Rees told VentureBeat. An agent operating on that flat plane does not need to escalate privileges. It already has them. That is why access control alone cannot contain what agents do after authentication. CrowdStrike CTO Elia Zaitsev described the detection gap to VentureBeat. In most default logging configurations, an agent’s activity is indistinguishable from a human. Distinguishing the two requires walking the process tree, tracing whether a browser session was launched by a human or spawned by an agent in the background. Most enterprise logging cannot make that distinction. Caulfield’s identity layer and Zaitsev’s telemetry layer are solving two halves of the same problem. No single vendor closes both gaps. “At any moment in time, that agent can go rogue and can lose its mind,” Caulfield said. “Agents read the wrong website or email, and their intentions can just change overnight.” How the request lifecycle works when agents have their own identity Five vendors shipped agent identity frameworks at RSAC 2026, including Cisco, CrowdStrike, Palo Alto Networks, Microsoft, and Cato Networks. Caulfield walked through how Cisco's identity-layer approach works in practice. The Duo agent identity platform registers agents as first-class identity objects, with their own policies, authentication requirements, and lifecycle management. The enforcement routes all agent traffic through an AI gateway supporting both MCP and traditional REST or GraphQL protocols. When an agent makes a request, the gateway authenticates the user, verifies that the agent is permitted, encodes the authorization into an OAuth token, and then inspects the specific action and determines in real time whether it should proceed. “No solution to agent AI is really complete unless you have both pieces,” Caulfield told VentureBeat. “The identity piece, the access gateway piece. And then the third piece would be observability.” Cisco announced its intent to acquire Astrix Security on May 4, signaling that agent identity discovery is now a board-level investment thesis. The deal also suggests that even vendors building identity platforms recognize that the discovery problem is harder than expected. Six-stage identity maturity model for agentic AI When a company shows up claiming 500 agents in production, Caulfield doesn't accept the number. "How do you know it's 500 and not 5,000?" Most organizations don’t have a source of truth for agents. Caulfield outlined a six-stage engagement model. Discovery first: identify every agent, where it runs, and who deployed it. Onboarding: register agents in the identity directory, tie each one to an accountable human, and define permitted actions. Control and enforcement: place a gateway between agents and resources, inspect every request and response. Behavioral monitoring: record all agent activity, flag anomalies, and build the audit trail. Runtime isolation contains agents on endpoints when they go rogue. Compliance mapping ties agent controls to audit frameworks before the auditor shows up. The six stages are not proprietary to any single vendor. They describe the sequence every enterprise will follow regardless of which platform delivers each stage. Maor's Censys data complicates step one before it even starts. Organizations beginning discovery should assume their agent exposure is already visible to adversaries. Step four has its own problem. Zaitsev's process-tree work shows that even organizations logging agent activity may not be capturing the right data. And step three depends on something Rees found most enterprises lack: a gateway that inspects actions, not just access, because the LLM does not respect the permission boundaries the identity layer sets. Agentic identity prescriptive matrix What to audit at each maturity stage, what operational readiness looks like, and the red flag that means the stage is failing. Use this to evaluate any platform or combination of platforms. Stage What to audit Operational readiness looks like Red flag if missing 1. Discovery Complete inventory of every agent, every MCP server it connects to, and every human accountable for it. A queryable registry that returns agent count, owner, and connection map within 60 seconds of an auditor asking. No registry exists. Agent count is an estimate. No human is accountable for any specific agent. Adversaries can see your agent infrastructure from the public internet before you can. 2. Onboarding Agents are registered as a distinct identity type with their own policies, separate from human and machine identities. Each agent has a unique identity object in the directory, tied to an accountable human, with defined permitted actions and a documented purpose. Agents use cloned human accounts or shared service accounts. Permission sprawl starts at creation. No audit trail ties agent actions to a responsible human. 3. Control A gateway between every agent and every resource it accesses, enforcing action-level policy on every request and every response. Four checkpoints per request: authenticate the user, authorize the agent, inspect the action, inspect the response. No direct agent-to-resource connections exist. Agents connect directly to tools and APIs. The gateway (if it exists) checks access but not actions. The flat authorization plane of the LLM does not respect the permission boundaries the identity layer set. 4. Monitoring Logging that can distinguish agent-initiated actions from human-initiated actions at the process-tree level. SIEM can answer: Was this browser session started by a human or spawned by an agent? Behavioral baselines exist for each agent. Anomalies trigger alerts. Default logging treats agent and human activity as identical. Process-tree lineage is not captured. Agent actions are invisible in the audit trail. Behavioral monitoring is incomplete before it starts. 5. Isolation Runtime containment that limits the blast radius if an agent goes rogue, separate from human endpoint protection. A rogue agent can be contained in its sandbox without taking down the endpoint, the user session, or other agents on the same machine. No containment boundary exists between agents and the host. A single compromised agent can access everything the user can. Blast radius is the entire endpoint. 6. Compliance Documentation that maps agent identities, controls, and audit trails to the compliance framework that the auditor will use. When the auditor asks about agents, the security team produces a control catalog, an audit trail, and a governance policy written for agent identities specifically. Emerging AI-risk frameworks (CSA Agentic Profile) exist, but mainstream audit catalogs (SOC 2, ISO 27001, PCI DSS) have not operationalized agent identities. No control catalog maps to agents. The auditor improvises which human-identity controls apply. The security team answers with improvisation, not documentation. Source: VentureBeat analysis of RSAC 2026 interviews (Caulfield, Zaitsev, Maor) and independent practitioner validation (McGladrey, Rees). May 2026. Compliance frameworks have not caught up “If you were to go through an audit today as a chief security officer, the auditor’s probably gonna have to figure out, hey, there are agents here,” Caulfield told VentureBeat. “Which one of your controls is actually supposed to be applied to it? I don’t see the word agents anywhere in your policies.” McGladrey's practitioner experience confirms the gap. The Cloud Security Alliance published an NIST AI RMF Agentic Profile in April 2026, proposing autonomy-tier classification and runtime behavioral metrics. But SOC 2, ISO 27001, and PCI DSS have not operationalized agent identities. The compliance frameworks McGladrey works with inside enterprises were written for humans. Agent identities do not appear in any control catalog he has encountered. The gap is a lagging indicator; the risk is not. Security director action plan VentureBeat identified five actions from the combined findings of Caulfield, Zaitsev, Maor, McGladrey, and Rees. Run an agent census and assume adversaries already did. Every agent, every MCP server those agents touch, every human accountable. Maor's Censys data confirms agent infrastructure is already visible from the public internet. NIST's NCCoE reached the same conclusion in its February 2026 concept paper on AI agent identity and authorization. Stop cloning human accounts for agents. McGladrey found that enterprises default to copying human user profiles, and permission sprawl starts on day one. Agents need to be a distinct identity type with scope limits that reflect what they actually do. Audit every MCP and API access path. Five vendors shipped MCP gateways at RSAC 2026. The capability exists. What matters is whether agents route through one or connect directly to tools with no action-level inspection. Fix logging so it distinguishes agents from humans. Zaitsev's process-tree method reveals that agent-initiated actions are invisible in most default configurations. Rees found authorization planes so flat that access logs alone miss the actual behavior. Logging has to capture what agents did, not just what they were allowed to reach. Build the compliance case before the auditor shows up. The CSA published a NIST AI RMF Agentic Profile proposing agent governance extensions. Most audit catalogs have not caught up. Caulfield told VentureBeat that auditors will see agents in production and find no controls mapped to them. The documentation needs to exist before that conversation starts.

Technology
VentureBeat· 4d ago

5,000 vibe-coded apps just proved shadow AI is the new S3 bucket crisis

Most enterprise security programs were built to protect servers, endpoints, and cloud accounts. None of them was built to find a customer intake form that a product manager vibe coded on Lovable over a weekend, connected to a live Supabase database, and deployed on a public URL indexed by Google. That gap now has a price tag. New research from Israeli cybersecurity firm RedAccess quantifies the scale. The firm discovered 380,000 publicly accessible assets, including applications, databases, and related infrastructure, built with vibe coding tools from Lovable, Base44, and Replit, as well as deployment platform Netlify. Roughly 5,000 of those assets, about 1.3%, contained sensitive corporate information. CEO Dor Zvi said his team found the exposure while researching shadow AI for customers. Axios independently verified multiple exposed apps, and Wired confirmed the findings separately. Among the verified exposures: a shipping company app detailed which vessels were expected at which ports. An internal health company application listed active clinical trials across the U.K. Full, unredacted customer service conversations for a British cabinet supplier sat on the open web. Internal financial information for a Brazilian bank was accessible to anyone who found the URL. The exposed data also included patient conversations at a children’s long-term care facility, hospital doctor-patient summaries, incident response records at a security company, and ad purchasing strategies. Depending on jurisdiction and the data involved, the healthcare and financial exposures may trigger regulatory obligations under HIPAA, UK GDPR, or Brazil’s LGPD. RedAccess found phishing sites built on Lovable that impersonated Bank of America, FedEx, Trader Joe’s, and McDonald’s. Lovable said it had begun investigating and removing the phishing sites. The defaults are the problem Privacy settings on several vibe coding platforms make apps publicly accessible unless users manually switch them to private. Many of these applications get indexed by Google and other search engines. Anyone can stumble across them. Zvi put it plainly: “I don’t think it’s feasible to educate the whole world around security. My mother is [vibe coding] with Lovable, and no offense, but I don’t think she will think about role-based access.” This is not an isolated finding In October 2025, Escape.tech scanned 5,600 publicly available vibe-coded applications and found more than 2,000 high-impact vulnerabilities, over 400 exposed secrets including API keys and access tokens, and 175 instances of personal data exposure containing medical records and bank account numbers. Every vulnerability Escape found was in a live production system, discoverable within hours. The full report documents the methodology. Escape separately raised an $18 million Series A led by Balderton in March 2026, citing the security gap opened by AI-generated code as a core market thesis. Gartner’s “Predicts 2026” report forecasts that by 2028, prompt-to-app approaches adopted by citizen developers will increase software defects by 2,500%. Gartner identifies a new class of defect where AI generates code that is syntactically correct but lacks awareness of broader system architecture and nuanced business rules. The remediation costs for these deep contextual bugs will consume budgets previously allocated to innovation. Shadow AI is the multiplier IBM’s 2025 Cost of a Data Breach Report found that 20% of organizations experienced breaches linked to shadow AI. Those incidents added $670,000 to the average breach cost, pushing the shadow AI breach average to $4.63 million. Among organizations that reported AI-related breaches, 97% lacked proper access controls. And 63% of breached organizations had no AI governance policy in place. Shadow AI breaches disproportionately exposed customer personally identifiable information at 65%, compared to 53% across all breaches, and affected data distributed across multiple environments 62% of the time. Only 34% of organizations with AI governance policies performed regular audits for unsanctioned AI tools. VentureBeat’s shadow AI research estimated that actively used shadow apps could more than double by mid-2026. Cyberhaven data found 73.8% of ChatGPT workplace accounts in enterprise environments were unauthorized. What to do first The audit framework below gives CISOs a starting point for triaging vibe-coded app risk across five domains. Domain Current State (Most Orgs) Target State First Action Discovery No visibility into vibe-coded apps Automated scanning of vibe coding platform domains Run DNS + certificate transparency scan for Lovable, Replit, Base44, and Netlify subdomains tied to corporate assets Authentication Platform defaults (public by default) SSO/SAML integration required before deployment Block unauthenticated apps from accessing internal data sources Code scanning Zero coverage for citizen-built apps Mandatory SAST/DAST before production Extend the existing AppSec pipeline to cover vibe-coded deployments Data loss prevention No DLP coverage for vibe coding domains DLP policies covering Lovable, Replit, Base44, Netlify Add vibe coding platform domains to existing DLP rules Governance No AI usage policy or shadow AI detection AI governance policy with regular audits for unsanctioned tools Publish an acceptable-use policy for AI coding tools with a pre-deployment review gate The CISO who treats this as a policy problem will write a memo. The CISO who treats this as an architecture problem will deploy discovery scanning across the four largest vibe coding domains, require pre-deployment security review, extend the existing AppSec pipeline to citizen-built apps, and add those domains to DLP rules before the next board meeting. One of those CISOs avoids the next headline. The vibe coding exposure RedAccess documented is not a separate problem from shadow AI. It is shadow AI's production layer. Employees build internal tools on platforms that default to public, skip authentication, and never appear on any asset inventory, which means the applications stay invisible to security teams until a breach surfaces or a reporter finds them first. Traditional asset discovery tools were designed to find servers, containers, and cloud instances. They have no way to find a marketing configurator that a product manager built on Lovable over a weekend, connected to a Supabase database holding live customer records, and shared with three external contractors through a public URL that Google indexed within hours. The detection challenge runs deeper than most security teams realize. Vibe-coded apps deploy on platform subdomains that rotate frequently and often sit behind CDN layers that mask origin infrastructure. Organizations running mature, secure web gateways, CASB, or DNS logging can detect employee access to these domains. But detecting access is not the same as inventorying what was deployed, what data it holds, or whether it requires authentication. Without explicit monitoring of the major vibe coding platforms, the apps themselves generate a limited signal in conventional SIEM or endpoint telemetry. They exist in a gap between network visibility and application inventory that most security stacks were never architected to cover. The platform responses tell the story Replit CEO Amjad Masad said RedAccess gave his company only 24 hours before going to the press. Base44 (via Wix) and Lovable both said RedAccess did not include the URLs or technical specifics needed to verify the findings. None of the platforms denied that the exposed applications existed. Wiz Research separately discovered in July 2025 that Base44 contained a platform-wide authentication bypass. Exposed API endpoints allowed anyone to create a verified account on private apps using nothing more than a publicly visible app_id. The flaw meant that showing up to a locked building and shouting a room number was enough to get the doors open. Wix fixed the vulnerability within 24 hours after Wiz reported it, but the incident exposed how thin the authentication layer is on platforms where millions of apps are being built by users who assume the platform handles security for them. The pattern is consistent across the vibe coding ecosystem. CVE-2025-48757 documented insufficient or missing Row-Level Security policies in Lovable-generated Supabase projects. Certain queries skipped access checks entirely, exposing data across more than 170 production applications. The AI generated the database layer. It did not generate the security policies that should have restricted who could read the data. Lovable disputes the CVE classification, stating that individual customers accept responsibility for protecting their application data. That dispute itself illustrates the core tension: platforms that market to nontechnical builders are shifting security responsibility to users who do not know it exists. What this means for security teams The RedAccess findings complete the picture. Professional agents face credential theft on one layer. Citizen platforms face data exposure on the other. The structural failure is the same. Security review happens after deployment or not at all. Identity and access management systems track human users and service accounts. They do not track the Lovable app a sales operations analyst deployed last Tuesday, connected to a live CRM database, and shared with three external contractors via a public URL. Nobody asks whether the database policies restrict who can read the data or whether the API endpoints require authentication. When those questions go unasked at AI-generation speed, the exposure scales faster than any human review process can match. The question for security leaders is not whether vibe-coded apps are inside their perimeter. The question is how many, holding what data, visible to whom. The RedAccess findings suggest the answer, for most organizations, is worse than anyone in the C-suite currently knows. The organizations that start scanning this week will find them. The ones that wait will read about themselves next.

Technology
Daily Brew· 4d ago

The AI Agent Security Surface: What Gets Exposed When You Add Tools and Memory

A structured framework to map and mitigate the backend attack vectors of agentic workflows.

TechnologyEconomics & Markets
Fortune· 4d ago

Why CEO Bill McDermott says ServiceNow’s 39% stock crash is Saaspocalypse ‘nonsense’ and why AI will make it a trillion-dollar company

In an in-depth interview with Fortune, the ServiceNow CEO explains how AI provides a tailwind to its business and how Wall Street is missing the point.

PaywallTechnology
Bloomberg· 4d ago

Intel CEO Who Won Over Trump and Musk Now Needs a Breakthrough

After Lip-Bu Tan became chief executive officer of Intel Corp. in March of last year, the struggling company’s shares went nowhere for seven months while the chipmaker was getting trounced in the market for artificial intelligence.

PaywallTechnologyEconomics & Markets
FT· 4d ago

Nintendo raises Switch 2 prices as profits disappoint

Sony also misses expectations for fourth quarter amid memory cost pain for both Japanese groups

PaywallTechnologyGeopolitics
Bloomberg· 4d ago

Baidu’s Chip Unit Plans Dual IPO in Shanghai and Hong Kong

Baidu Inc.’s chip unit Kunlunxin is planning an initial public offering on Shanghai’s Nasdaq-style bourse in addition to a separate listing plan in Hong Kong as the Chinese search engine giant looks to taps investor appetite for semiconductor stocks.

TechnologyEconomics & Markets
MacroMicro· 4d ago

The AI Race Is No Longer Zero-Sum: Why Intensifying Competition May Be Extending the Infrastructure Supercycle | Blog | MacroMicro

What You Should Know Recent reports suggesting that OpenAI had missed internal revenue and user growth targets briefly reignited concerns surrounding the sustai...

PaywallTechnologyGeopolitics
Bloomberg· 4d ago

US Said to Suspect Nvidia Chips Smuggled to Alibaba Via Thailand

A key company behind Thailand’s national AI effort is suspected of helping to smuggle billions of dollars worth of Super Micro Computer Inc. servers containing advanced Nvidia Corp. chips to China, with Alibaba Group Holding Ltd. one of multiple end customers, according to people familiar with the matter.

TechnologyEconomics & Markets
Theregister· 4d ago

Akamai surges on big LLM deal as Cloudflare dims

Good times, bad times

TechnologyEconomics & Markets
VentureBeat· 4d ago

Anthropic wants to own your agent's memory, evals, and orchestration — and that should make enterprises nervous

Just a few weeks after announcing Claude Managed Agents, Anthropic has updated the platform with three new capabilities that collapse infrastructure layers like memory, evaluation, and multi-agent orchestration, into a single runtime. This move could threaten the standalone tools that many enterprises cobble together. The new capabilities — 'Dreaming,' 'Outcomes,' and 'Multi-Agent Orchestration' — aim to make agents inside Claude Managed Agents “more capable at handling complex tasks with minimal steering,” Anthropic said in a press release.   Dreaming deals with memory, where agents “reflect” on their many sessions and curate memories so they learns and surface unknown patterns. Outcomes allows teams to define and set specific rubrics to measure an agent's success, while Multi-Agent Orchestration breaks jobs down so a lead agent can delegate to other agents. Claude Managed Agents ideally provides enterprises with a simpler path to deploy agents and embeds orchestration logic in the model layer. It’s an end-to-end platform to manage state, execution graphs, and routing. With the addition of Dreaming, Outcomes and Multi-agent Orchestration, Claude Managed Agents expands capabilities even further and directly competes with tools like LangGraph or CrewAI, as well as external evaluation frameworks, RAG memory architectures, and QA loops. An integration threat Enterprises must now ask: Should we ditch our flexible, modular system in favor of an agent platform that brings almost everything in-house? Anthropic designed Claude Managed Agents to share context, state, and traceability in one place. This means the platform sees every decision agents make, rather than enterprises having to wire separate systems together. It sounds practical to have one platform that does everything. But not all enterprises want a full-service system.  Claude Managed Agents already faces criticism that it encourages vendor lock-in because it owns most of the architecture and tools that govern agents. In the current paradigm, an organization may run Managed Agents but keep multi-agent orchestration, memory, or evaluations in a separate space ensures flexibility.  The platform offers a fully-hosted runtime, which means memory and orchestration run on infrastructure the enterprise does not own. This can become a compliance nightmare for some organizations that have to prove data residency.  Another problem to consider is that enterprises already in the middle of large-scale AI transformations must cobble together workarounds to deal with the constraints of their tech stack. Not every workflow is easily replaceable by switching to Claude Managed Agents.  Dreaming and outcomes against current tools Most enterprises have a fragmented approach to AI deployment. For example, they may use LangGraph or Crew AI for agent routing and workflow management, Pinecone as a vector database for long-term memory, DeepEval for external evaluation, and a human-in-the-loop quality assurance to review some tasks. Anthropic hopes to do away with all of that.  With Dreaming, Anthropic approaches memory by allowing users to actively rewrite it between sessions, so the agent essentially learns from its mistakes. Anthropic says this capability is useful for long-running states and orchestration. Current systems often handle memory persistence by storing embeddings, retrieving relevant context, and adding more state over time.  Outcomes addresses the evaluation portion by detailing expectations for agents. Instead of external quality checks, which are often done by a team of humans, Anthropic is bringing evaluation into the orchestration layer rather than above it.  But it’s the Multi-Agent Orchestration capability that pits Claude Managed Agents against orchestration frameworks from Microsoft, LangChain, CrewAI, and others. Model providers like Anthropic and OpenAI have already begun pushing aggressively into this space, arguing that bringing this to the model layer gives teams better control. Big decisions to make Enterprises face a big decision, and this one could depend on where they are in agent maturity.  If an organization is still experimenting with agents and has not deployed many in production, they may find moving to Claude Managed Agents and configuring Dreaming and Outcomes to their needs much easier. This is the stage of development where, even if enterprises are using a third-party orchestrator like LangChain, they’re still customizing it.  But for those who are already further along in the process, the calculation becomes trickier. It’s now a matter of parallel evaluation and better understanding of their processes.  Businesses, though, will face the same decision even if they don’t intend to use Claude Managed Agents. Anthropic has signaled that other model and platform providers will likely shift their product roadmaps to a similar model that keeps everything locked in the same system — because models may become interchangeable, but the tooling and orchestration infrastructure will not.

TechnologyTechnology & Infrastructure
TechStory· 4d ago

AWS Outage Sparks Fresh Concerns Over Data Center Cooling as Coinbase and Other Services Face Disruptions - TechStory

The outage also spotlighted an increasingly urgent challenge for hyperscale cloud providers: cooling AI infrastructure. Modern AI servers and advanced cloud systems consume enormous amounts of electricity while processing massive volumes of data. That energy consumption generates intense heat ...

TechnologyGeopolitics
Daily Brew· 4d ago

QuTwo Secures €25M for Quantum AI Lab, Aims to Boost Europe's AI Sovereignty

QuTwo secures €25 million in angel funding to establish a European AI lab targeting the quantum computing era.

TechnologyLabor & Society
Theregister· 4d ago

Cloudflare to fire 1,100 staff whose jobs just aren’t AI enough

Around 20 percent of staff get an ‘In one hour, you might not work here anymore’ email

Technology
Siliconrepublic· 4d ago

Cloudflare cuts headcount by 20pc for leaner, AI-powered workforce

Cloudflare previously announced plans to hire more than 1,000 interns to 'ramp up' AI use. Read more: Cloudflare cuts headcount by 20pc for leaner, AI-powered workforce

TechnologyLabor & Society
Reuters· 4d ago

Cloudflare to cut about 20% workforce as AI adoption reshapes operations | Reuters

Cloudflare said on Thursday it would cut about 20% of its ‌workforce as the company restructures operations around the rapid adoption of artificial intelligence tools, and forecast second-quarter revenue slightly below Wall Street expectations.

TechnologyTechnology & Infrastructure
Daily Brew· 4d ago

Anthropic introduces Dreaming

Anthropic has unveiled 'Dreaming,' a new system that allows AI agents to learn from their own mistakes.

TechnologyTechnology & Infrastructure
Top Daily Headlines: IBM Cloud evaporates as datacenter loses power· 4d ago

IBM Cloud evaporates as datacenter loses power

Customers say services were down for at least 4 hours, while status page showed no issues.

TechnologyAdoption & Impact
Theregister· 4d ago

HPE drops first Juniper x Aruba collab – self-driving Wi-Fi

NetAdmins can stay in the loop while they learn to trust AI to tackle some scutwork

TechnologyTechnology & Infrastructure
Top Daily Headlines: IBM Cloud evaporates as datacenter loses power· 4d ago

Chrome silently installs a 4 GB local LLM on your computer

You did remember to opt out of AI, didn't you?

PaywallTechnologyEconomics & Markets
Bloomberg· 4d ago

Sony Announces $3 Billion Buyback as Memory Prices Take Toll

Sony Group Corp. said it will buy back as much as ¥500 billion ($3.2 billion) of its shares after rising memory prices weighed on the entertainment group’s annual outlook.

PaywallTechnologyEconomics & Markets
FT· 4d ago

Anthropic weighs deal for near $1tn valuation as revenue surges

Start-up behind Claude tool is fielding inbound investment offers that could lead to it surpassing rival OpenAI in value

PaywallTechnologyEconomics & Markets
FT· 4d ago

Tech boom brings emerging markets and their rich cousins closer together

Asian manufacturers supplying the picks and shovels of the AI bonanza could be on a more durable footing

PaywallTechnologyEconomics & Markets
FT· 4d ago

Big Tech’s $725bn AI spending spree sends free cash flow to a decade low

Silicon Valley giants have transformed from asset-light cash machines to huge infrastructure investors

Technology
Daily Brew· 4d ago

SpaceX has a $55 billion plan to build AI chips in Texas

SpaceX is reportedly planning a massive investment to develop AI-focused hardware at a new facility in Texas.

TechnologyEconomics & Markets
Bebeez· 4d ago

UK semiconductor startup Quinas Technology secures investment approval from Malta Government Venture Capital

Quinas Technology, a London-based DeepTech semiconductor startup developing memory technologies for AI, data centres, edge computing and advanced digital systems, has secured investment approval from Malta Government Venture Capital (MGVC).  On 7 May 2026, the MGVC approved the investment under its co-investment framework to support the growth of the UK-based company. The funding, the amount […]

PaywallTechnologyTechnology & Infrastructure
NYT· 4d ago

Elon Musk’s SpaceX Plans $55 Billion Investment to Make A.I. Chips

The rocket company’s new semiconductor factory, called Terafab, is part of the billionaire’s increasing efforts to dominate artificial intelligence.

PaywallTechnologyEconomics & Markets
FT· 4d ago

Chris Hohn’s hedge fund slashes $8bn Microsoft stake in warning over AI disruption

TCI has cut its position in tech giant from 10% to 1%

PaywallTechnology
Bloomberg· 4d ago

Agentic AI 2026 Outlook | Insights | Bloomberg Professional Services

Artificial intelligence agents are poised to reshape the application-software market, lowering barriers to software creation while increasing competition and pressuring margins.

Technology
VentureBeat· 4d ago

5% GPU utilization: The $401 billion AI infrastructure problem enterprises can't keep ignoring

For the last 24 months, one narrative justified every over-provisioned data center and bloated IT budget: the GPU scramble. Silicon was the new oil, and H100s traded like contraband. Reserve capacity now or your enterprise would be left behind. The bill is now due, and the CFO is paying attention. Gartner estimates AI infrastructure is adding $401 billion in new spending this year. Real-world audits tell a darker story: average GPU utilization in the enterprise is stuck at 5%.  That utilization floor is driven by a self-reinforcing procurement loop that makes idle GPUs nearly impossible to release. What makes this shift more urgent is the CapEx reality now hitting enterprise balance sheets. Many organizations locked in GPU capacity under traditional three- to five-year depreciation cycles, with the hyperscalers being at five years. That means the infrastructure purchased during the peak of the “GPU scramble” is now a fixed cost, regardless of how much it is actually used. As those assets age, the question is no longer whether the investment was justified. It’s whether it can be made productive. Underutilized GPUs are not just idle resources, they are depreciating assets that must now generate measurable return. This is forcing a shift in mindset: from acquiring capacity to maximizing the economic output of what is already deployed. The scramble was a sideshow For the "Tier 1" enterprise — the Intuits, Mastercards, and Pfizers of the world — access was rarely the true bottleneck. Leveraging deep-pocketed relationships with AWS, Azure, and GCP, these organizations secured capacity reservations that sat idle while internal teams struggled with data gravity, governance, and architectural immaturity. The industry narrative of "scarcity" served as a convenient smokescreen for this inefficiency. While the headlines focused on supply chain delays, the internal reality was a massive productivity gap. Organizations were activity-rich (buying chips) but output-poor (generating near-zero useful tokens). At 5% utilization, the math simply doesn't work. For every dollar spent on silicon, 95 cents is essentially a donation to a cloud provider’s bottom line. In any other department, a 95% waste metric would be a firing offense; in AI infrastructure, it was just called "preparedness." The Q1 tracker: A market in pivot VentureBeat’s Q1 2026 AI Infrastructure & Compute Market Tracker confirms that the panic phase has officially broken. The tracker is directional rather than statistically definitive — January surveyed 53 qualified respondents, and in February there were 39 — but the pattern across both waves is consistent. When we asked IT decision-makers what actually drives their provider choices today, the results show a market in rapid pivot: The access collapse: “Access to GPUs/availability” factor dropped from 20.8% to 15.4% in a single quarter — from primary concern to secondary in 90 days. The pragmatic pivot: “Integration with existing cloud and data stacks” held steady as the top priority at roughly 43% across both waves, while security and compliance requirements surged from 41.5% to 48.7% — nearly closing the gap with integration. The TCO mandate: “Cost per inference/TCO (total cost of ownership)” as a top priority jumped from 34% to 41% in a single quarter, overtaking performance as the dominant procurement lens. The era of the blank check is dead. Inference is where AI becomes a line item.  Training and even fine-tuning were a tactical project; inference is a strategic business model. For most enterprises, the unit economics of that model are currently unsustainable. During the initial pilot phase, flat-fee licenses and bundled token deals allowed for architectural waste. Teams built long-context agents and complex retrieval pipelines because tokens were effectively a sunk cost. As the industry moves toward usage-based pricing in 2026, those same architectures have become liabilities. When metered billing is applied to an infrastructure stack that sits idle 95% of the time, the cost per useful token becomes a line-item emergency the moment a project moves into production. From activity to productivity The shift highlighted in our Q1 data represents more than just a budget correction; it is a fundamental change in how the success of an AI leader is measured. For the last two years, success was about “securing” the stack. In the efficiency era, success is “squeezing” the stack. This is why cost optimization platforms saw the largest planned budget increase in our survey, becoming a top-tier priority as organizations realize that buying more GPUs is often the wrong answer. Increasingly IT users are asking how to stop paying for GPUs they aren't using. They are moving away from measuring GPU activity (how many chips are powered on) and toward GPU productivity (how many useful tokens are generated per dollar spent). The luxury of underutilization is now a liability. The next act of the enterprise AI play is more about finding a way to make the silicon you already have pay for itself. Owning the mint: The choice between token consumer and producer As organizations move from proof-of-concept to production, the focus is shifting away from the latest GPU and toward the architecture of token generation. In this new economic reality, every enterprise must decide its role in the token economy: will you be a token consumer, paying a permanent tax to a model provider, or a token producer, owning the infrastructure and the unit economics that come with it? This choice is not just about cost; it is about how an organization decides to handle complexity. Owning inference infrastructure means overcoming KV cache persistence, understanding the storage architecture, knowing what are tolerable latency guarantees, and addressing power constraints. It also introduces real-world enterprise limitations, power availability, data center footprint, and operational complexity, that directly impact how far and how fast AI can scale. At the core of this challenge is KV cache economics. Storing context in GPU memory delivers performance but comes at a premium, limiting concurrency and driving up cost per token. Offloading KV cache to shared NVMe-based storage can improve reuse and reduce prefill overhead, but introduces tradeoffs in latency and system design. As NVMe costs rise and GPU memory remains scarce, organizations are forced to balance performance against efficiency. For a token producer, managing these tradeoffs, across memory, storage, power, and operations, is simply the cost of doing business at scale. For others, the overhead remains too high, requiring a different path. The specialized cloud pivot VentureBeat’s Q1 tracker shows that the market is already voting on this strategy. The top strategic direction for enterprises is now to move more workloads to specialized AI clouds, a category that grew from 30.2% to 35.9% in our latest survey. These providers — including Coreweave, Lambda, and Crusoe — are evolving. While they initially gained ground by serving model builders and training-heavy workloads, their revenue mix is changing rapidly. Today, training represents roughly 70% of their business volume, but inference customers now make up 30%. We expect that ratio to flip by the end of 2026 as the long tail of enterprise inference begins to scale. These specialized providers are gaining strategic attention because they are not just selling GPU access. They are selling the removal of infrastructure friction. They optimize the full stack — storage, networking, and scheduling — around inference-first economics rather than general-purpose cloud operations. For an organization aiming to be a token producer, these environments offer a more efficient factory floor than traditional hyperscalers. The rise of managed inference For organizations that realize they cannot efficiently build or manage their own inference factories, a different trend is emerging. Our survey found that the intention to evaluate inference outsourcing and managed LLM providers jumped from 13.2% to 23.1% in a single quarter. This nearly 10-percentage-point increase represents a realization that building inference infrastructure internally often creates hidden costs. Providers like Baseten, Anyscale, FireworksAI, and Together AI offer predictable pricing and service-level agreements without requiring the customer to become experts in vLLM tuning or distributed GPU scheduling. In this model, the enterprise remains a token consumer, but one that is actively looking to price away the complexity of the stack. They are learning that managing inference internally is only viable if they have the volume to justify the operational burden. Simplifying the hybrid stack The choice to be a producer is also being made easier by a new layer of hybrid-cloud AI platforms. Solutions from Red Hat, Nutanix, and Broadcom are designed to operationalize open-source inference infrastructure without forcing every company to become a systems integrator. The challenge is that modern inference depends on complex open-source components like vLLM, Triton, and Kubernetes. These systems rely on a rapidly evolving stack, with vLLM for high-throughput serving, Triton for model orchestration, and Ray for distributed execution, each powerful on its own, but complex to integrate, tune, and operate at scale. For most enterprises, the challenge isn’t access to these tools, it’s stitching them together into a reliable, production-grade inference pipeline. The promise of these newer platforms is portability: the ability to build an inference stack once and deploy it anywhere, whether in a hyperscaler, a specialized cloud, or an on-premises data center. Our Q1 2026 AI Infrastructure & Compute Market Tracker confirms that interest in these DIY-but-managed stacks is growing, jumping from 11.3% in January to 17.9% in February, alongside provider adoption, with a steady rise in organizations leaning into open source. This flexibility matters because enterprise AI will not be centralized in one place. Inference workloads will be distributed based on where data lives, how sensitive it is, and where the cost of running it is lowest. The winner in the next phase of the token economy will not be the platform that forces standardization through restriction. It will be the one that delivers standardization through portability, allowing enterprises to switch between being consumers and producers as their needs evolve. The architecture of efficiency: The technical levers of productivity Fixing the 5% utilization wall requires more than just better software; it requires a structural overhaul of the efficiency stack. Many organizations are discovering that high activity is not the same as high productivity. A cluster can run at full tilt but remain economically inefficient if time-to-first-token is too high or if inference requests spend too much time in prefill. Inference economics are determined by how much useful output a cluster generates per unit of cost. This requires a shift from measuring GPU activity — simply having the chips powered on — to measuring GPU productivity. Achieving that productivity depends on three technical levers: the network, the memory, and the storage stack. Networking: The cost of waiting The network is the often-ignored backbone of inference economics. In a distributed environment, the speed at which data moves between compute nodes and storage determines whether a GPU is actually working or merely waiting. RDMA (Remote Direct Memory Access) has become the non-negotiable standard for this move. By allowing data to bypass the CPU and move directly between memory and the GPU, RDMA eliminates the latency spikes that traditional network architectures introduce. In practical terms, an RDMA-enabled architecture can increase the output per GPU by a factor of ten for concurrent workloads. Without this level of networking, an enterprise is effectively paying a "waiting tax" on every chip in the rack. As model context windows expand and multi-node orchestration becomes the norm, the network determines whether a cluster is a high-speed factory or a bottlenecked warehouse. Solving the memory tax: Shared KV cache As models become larger and context windows expand toward the millions of tokens, the cost of repeatedly rebuilding the prompt state has become unsustainable. Large language models rely on key-value (KV) caches to maintain context during a session. Traditionally, these are stored in local GPU memory, which is both expensive and limited. This creates a "memory tax" that crushes unit economics as concurrency rises. To solve this, the industry is moving toward persistent shared KV cache architectures. By storing the cache centrally on high-performance storage rather than redundantly across multiple GPU nodes, organizations can reduce prefill overhead and improve context reuse. Newer architectures are already proving this out. The VAST Data AI Operating System, running on VAST C-nodes using Nvidia BlueField-4 DPUs, allows for pod-scale shared KV cache that collapses legacy storage tiers. Similarly, the HPE Alletra Storage MP X10000 — the first object-based platform to achieve Nvidia-Certified Storage validation — is designed specifically to feed data to inference resources without the coordination tax that causes bottlenecks at scale. WEKA is another provider in this space.  The compression edge Beyond the physical hardware, new algorithmic contributions are redefining what is possible in inference memory. Google’s recent presentation of TurboQuant at ICLR 2026 demonstrates the scale of this shift. TurboQuant provides up to a 6x compression level for the KV cache with zero accuracy loss. Techniques like these allow for building large vector indices with minimal memory footprints and near-zero preprocessing time. For the enterprise, this means more concurrent users on the same hardware estate without the "rebuild storms" that typically cause latency spikes. The caveat: compression standards remain contested — no open-source consensus has emerged, and the space is shaping up as a proprietary stack war between Google and Nvidia. Storage as a financial decision Storage is no longer just a backend decision; it is a financial one. Platforms like Dell PowerScale are now delivering up to 19x faster time-to-first-token compared to traditional approaches, according to Dell. By separating high-performance shared storage and memory-intensive data access from scarce GPU resources, these platforms allow inference to scale more efficiently. When a storage layer can keep GPU-intensive workloads continuously fed with data, it prevents expensive resources from sitting idle. In the efficiency era, the goal is to drive the 5% utilization wall upward by ensuring that every cycle is spent on token generation, not on data movement. But as the stack becomes more efficient, the perimeter becomes more porous. High-productivity tokens are worthless if the data powering them cannot be trusted. Sovereignty and the agentic future: Building the trust foundation The final barrier to achieving return on AI is not a technical bottleneck, but a trust bottleneck. As enterprise AI shifts from simple chatbots to autonomous agents, the risk profile changes. Agents require deep access to internal systems and intellectual property to be useful. Without a sovereign architecture, that access creates a liability that most organizations are not equipped to manage. VentureBeat research into the state of AI governance reveals a stark disconnect. While many organizations believe they have secured their AI environments, 72% of enterprises admit they do not have the level of control and security they think they do. This governance mirage is particularly dangerous as agentic systems move into production. In the last 12 months, 88% of executives reported security incidents related to AI agents. Sovereignty as an architecture principle Data sovereignty is often treated as a geographic or regulatory checkbox. For the strategic enterprise, it must be treated as a core architecture principle. It is about maintaining control, lineage, and explainability over the data that powers an agentic workflow. This requires a new approach to data maturity, modeled on the traditional medallion architecture. In this framework, data moves through layers of usability and trust — from raw ingestion at the bronze level to refined gold and, eventually, platinum-quality operational data. AI inference must follow this same discipline. Agentic systems do not just need available context; they need trusted context. Providing the wrong data to an agent, or exposing sensitive intellectual property to a non-sovereign endpoint, creates both business and regulatory risk. Compartmentalization must be designed into the stack from the start. Organizations need to know which models and agents can access specific data layers, under what conditions, and with what lineage attached. Bringing the AI to the data The fundamental question for the agentic future is whether to bring the data to the AI or the AI to the data. For highly sensitive workloads, moving data to a centralized model endpoint is often the wrong answer. The move toward private AI — where inference happens closer to where trusted data resides — is gaining momentum. This architecture uses sovereign clouds, private environments, or governed enterprise platforms to keep the data perimeter intact. This is where the choice to be a token producer becomes a security advantage. By owning the inference stack, an enterprise can enforce governance and lineage at the infrastructure layer. It ensures that the intellectual property used to ground an agent never leaves the organization's control. The next platform war The battle for AI dominance will not be decided by who owns the largest GPU clusters. It will be won by the companies with the best inference economics and the most trusted data foundation. The organizations that win the efficiency era will be those that deliver the lowest cost per useful token and the fastest path to production. They will be the ones that have moved past the hoarding hangover to focus on productive output. Achieving return on AI requires a shift in mindset. It means moving from a culture of securing the stack to a culture of squeezing the stack. It requires architectural rigor, a focus on token-level ROI and a commitment to sovereignty. When an organization can generate its own tokens efficiently and securely, AI moves from a science project to an economically repeatable business advantage. That is how ROI becomes real. That is where the next generation of enterprise advantage will be built. Rob Strechay is a Contributing VentureBeat analyst and principal at Smuget Consulting, a research and advisory firm focused on data infrastructure and AI systems. Disclosure: Smuget Consulting engages or has engaged in research, consulting, and advisory services with many technology companies, which can include those mentioned in this article. Analysis and opinions expressed herein are specific to the analyst individually, and data and other information that might have been provided for validation, not those of VentureBeat as a whole.

TechnologyTechnology & Infrastructure
Theregister· 4d ago

AWS warns of EC2 ‘impairment’ as power loss hits notorious US-EAST-1 region

Extra aircon found to cool overheating datacenter as users complain their resources are ... nowhere

TechnologyTechnology & Infrastructure
Bebeez· 4d ago

Stellanor completes acquisition of eight data centers from Redcentric

DWS-backed platform now operating eleven UK data centers serving enterprise and AI-ready workloads LONDON, May 5, 2026 /PRNewswire/ — Stellanor Datacenters, the UK’s fastest-growing urban data center company backed by DWS, today announced the successful completion of its acquisition of eight data centers from UK managed services provider Redcentric plc.   The addition of these […]

TechnologyTechnology & Infrastructure
Daily Brew· 4d ago

Apple’s AirPods with cameras for AI are apparently close to production

Reports suggest Apple is nearing production for new AirPods models equipped with cameras to support AI-driven features.

TechnologyTechnology & Infrastructure
Daily Brew· 4d ago

The Canvas Hack Is a New Kind of Ransomware Debacle

A security breach involving the Canvas platform highlights the evolving threat of ransomware attacks.

PaywallTechnology
FT· 4d ago

While AI pumps up portfolios, some companies sound the alarm

The clout of Big Tech groups is obscuring the precarious state of others

Technology
Ticker Report· 4d ago

SoundHound AI Q1 Earnings Call Highlights - Ticker Report

SoundHound AI (NASDAQ:SOUN) reported first-quarter 2026 revenue of $44.2 million, up 52% from a year earlier, as management pointed to broad enterprise demand, automotive growth and continued expansion across financial services, restaurants, healthcare and technology.

Technology
Firstpost· 4d ago

Anthropic says AI demand could fuel 80x growth in 2026 — and even it is struggling to keep up

Anthropic CEO Dario Amodei says surging demand for AI tools could drive the startup to 80x growth in 2026, forcing the company into a frantic race for computing power as it expands partnerships with SpaceX, Google and Amazon.

Technology
ICO Optics· 4d ago

Datadog Jumps 31% as AI Winners Reshape Software Market – ICO Optics

AI adoption is speeding up, and infrastructure is playing a decisive role in model development and reliability. This quarter’s results highlight a few things for the broader market: AI-native platforms with clear monetization strategies are catching investors’ eyes, even while most software ...

Technology
MSN· 4d ago

TSMC posts strong April revenue growth driven by AI chip demand

Taiwan Semiconductor Manufacturing Co (NYSE:TSM) reported a solid increase in April revenue on Friday as continued demand for advanced artificial intelligence semiconductors supported sales growth. The world’s largest contract chip manufacturer said April revenue reached NT$410.73 billion ...

Technology
Kavout· 4d ago

AI's Profit Paradox: Navigating the Next Phase of Tech Investment

Despite massive AI investment, enterprise returns remain elusive, creating a market paradox. Hardware giants like NVIDIA are clear beneficiaries, while software firms face a more complex path. Navigating this phase requires understanding the disconnect between hype and economic reality.

Technology
The Motley Fool· 4d ago

ServiceNow Had Problems Long Before Agentic AI. Here's Why. | The Motley Fool

ServiceNow stock is cratering as competition from agentic AI heats up.

Technology
MIT Technology Review· 4d ago

Musk v. Altman week 2: OpenAI fires back, and Shivon Zilis reveals that Musk tried to poach Sam Altman

In the second week of the landmark trial between Elon Musk and OpenAI, Musk’s motivations for bringing the suit were under scrutiny. Last week, Musk took the stand, alleging that OpenAI CEO Sam Altman and president Greg Brockman had deceived him into donating $38 million to the company. He claimed that they’d promised to maintain…

Technology
Yahoo! Finance· 4d ago

NVIDIA (NVDA) Gains Another Enterprise AI Win As Adoption Continues To Expand

With an upside potential of 35.94%, NVIDIA Corporation (NASDAQ:NVDA) is among the 10 Tech Stocks That Could Make You a Millionaire. On April 27, LiveRamp announced native support for NVIDIA Corporation (NASDAQ:NVDA)’s AI infrastructure, underscoring expanding enterprise adoption of the ...

Technology
GuruFocus· 4d ago

TCI Reduces Microsoft (MSFT) Stake Amid AI Concerns

On May 08, 2026, TCI Fund Management, led by Sir Christopher Hohn, announced a significant reduction in its stake in Microsoft Corp (MSFT), decreasing its owner

Technology
Daily Brew· 4d ago

Musk v. Altman Evidence Shows What Microsoft Executives Thought of OpenAI

Newly surfaced evidence from the Musk v. Altman trial reveals internal Microsoft communications regarding their partnership with OpenAI.

Technology
GuruFocus· 4d ago

Chip Stocks Surge as AI Infrastructure Investment Grows, Nvidia (NVDA) Leads the Way

On May 08, 2026, the Philadelphia Semiconductor Index has risen approximately 18% this year, significantly outpacing the S&P 500 Software and Services Index, wh

Technology
GuruFocus· 4d ago

Nvidia (NVDA): Strategic Investment in AI Infrastructure with Corning (GLW)

On May 07, 2026, Nvidia (NVDA) announced a significant expansion of its investments in AI infrastructure, which includes a multibillion-dollar advance payment t

Technology
The Economic Times· 4d ago

Why venture capital firms want Indian AI founders in Silicon Valley early - The Economic Times

Indian AI startups targeting global markets are now advised by venture capital firms to establish an early US presence, particularly in San Francisco. This strategic shift is driven by the critical need for proximity to customers, capital, talent, and emerging AI trends to effectively scale ...

PaywallTechnology
Bloomberg· 4d ago

Nvidia Names Goldman Sachs Veteran Suzanne Nora Johnson to Board

Nvidia Corp. named former Goldman Sachs Group Inc. Vice Chairman Suzanne Nora Johnson as a director, adding a finance-industry and philanthropy veteran to the chipmaker’s board.

Technology
MarTech· 4d ago

Amplitude and Statsig deal raises questions for customers | MarTech

Amplitude gets the Statsig platform and customers, but OpenAI keeps the team that built the technology.

Technology
Stimson Center· 4d ago

All-In on AI: How the United States and Taiwan Are Deepening Their Chip Partnership • Stimson Center

Taiwan’s cooperation with the United States on the AI supply chain could strengthen the silicon shield.

PaywallTechnologyAdoption & Impact
Daily Brew· 4d ago

Meta's embrace of AI is making its employees miserable

A report details how Meta's aggressive pivot toward AI integration is negatively impacting employee morale and workplace culture.

Technology
Yahoo! Finance· 4d ago

Cloudflare stock sinks as company slashes 20% of its workforce, citing AI

Cloudfare issued a Q2 sales forecast on Thursday that fell short of analyst expectations and said it would slash roughly 1,100 jobs, or about a fifth of its workforce, as it adopts artificial intelligence tools in its operating model.

Technology
The Hans India· 4d ago

Cloudflare to Lay Off 20% Workforce as AI Reshapes Company Operations

Cloudflare plans major global layoffs as rapid AI adoption transforms internal operations, signaling deeper automation trends across the technology industry.

Technology
Seeking Alpha· 4d ago

The reality of AI layoffs and the current labor market ( | Seeking Alpha

Tech layoffs and AI: are job cuts real or “right-sizing”? Get key data from the Challenger report, ADP jobs beat, and what to watch in today’s payrolls—read...

Technology
VentureBeat· 4d ago

OpenAI brings GPT-5-class reasoning to real-time voice — and it changes what voice agents can actually orchestrate

Voice agents have been expensive to run and painful to orchestrate, not because the models can't handle conversation, but because context ceilings forced enterprises to build session resets, state compression, and reconstruction layers into every deployment. OpenAI's three new voice models are designed to reduce that overhead, and they change how engineers can think about building voice into a larger agent stack. GPT-Realtime-2, GPT-Realtime-Translate, and GPT-Realtime-Whisper integrate real-time audio into the model management stack as discrete orchestration primitives — separating conversational reasoning, translation, and transcription into specialized components rather than bundling them in a single voice product. The company said in a blog post that Realtime-2 is its first voice model “with GPT-5 class reasoning” and can handle difficult requests and keep conversations flowing naturally. Realtime-Translate understands more than 70 languages and translates them into 13 others at the speaker's pace, and Realtime-Whisper is its new speech-to-text transcription model. These three actions no longer sit inside a single stack or model. GPT-Realtime-2 could technically handle transcription, but OpenAI is routing distinct tasks to specialized models: Realtime-Translate for multilingual speech and Realtime-Whisper for transcription. Enterprises can assign each task to the appropriate model rather than routing everything through a single, all-encompassing voice system. The new OpenAI models compete against Mistral’s Voxtral models, which also separate transcription and target enterprise use cases.   What enterprises should do More enterprises are seeing the value of voice agents now that more people are becoming comfortable conversing with an AI agent, and also because of the richness of data from voice customer interactions. Organizations evaluating these models will need to consider their orchestration architecture, not just model quality — specifically, whether their stack can route discrete voice tasks to specialized models and manage state across a 128K-token context window.

Technology
yourNEWS· 4d ago

AI Could Soon Consume Nearly Half of Global Data Center Electricity, New Research Warns – [your]NEWS

The rapid growth of AI infrastructure is already complicating climate goals established by major technology companies. Google disclosed in its 2024 environmental report that its greenhouse gas emissions have risen 48% since 2019, driven in large part by increased energy demand tied to AI expansion. “As we further integrate AI into our products, reducing emissions may be challenging ...

TechnologyTechnology & Infrastructure
Theregister· 4d ago

Iran war hits datacenter building supply chains, upping costs

BCS says builders face up to 20% material hikes and patchy deliveries

TechnologyTechnology & Infrastructure
Top Daily Headlines: IBM Cloud evaporates as datacenter loses power· 4d ago

AMD puts out new slottable GPU for AI-curious enterprises

MI350P packs 144 GB of HBM3e and up to 4.6 petaFLOPS of FP4 grunt into a dual slot card.

TechnologyTechnology & Infrastructure
Arxiv· 4d ago

Authorization Propagation in Multi-Agent AI Systems: Identity Governance as Infrastructure

arXiv:2605.05440v1 Announce Type: new Abstract: The security discussion around agentic AI focuses heavily on prompt injection. This paper argues that multi-agent systems also create a distinct authorization problem: maintaining authorization invariants as non-human principals retrieve data, delegate tasks, and synthesize results across changing boundaries. We call this problem authorization propagation. It is not reducible to prompt injection and is not fully addressed by classical access-control models such as RBAC, ABAC, or ReBAC. The paper formalizes authorization propagation as a workflow-level property, identifies three sub-problems (transitive delegation, aggregation inference, and temporal validity), and derives seven structural requirements for authorization architectures in multi-agent AI systems. Recent work on invocation-bound capability tokens, task-scoped authorization envelopes, dependency-graph policy enforcement, and execution-count revocation demonstrates that the field is converging on the problem, but not yet on a complete architecture. The central claim is that identity governance must be treated as infrastructure: evaluated continuously, enforced at every interaction boundary, and designed into the system before orchestration logic is allowed to scale. Preliminary implementation evidence from a production enterprise AI platform shows that ordinary system behavior, not only adversarial action, already produces the failures this model predicts.

TechnologyAdoption & Impact
Bebeez· 4d ago

Pinecone Expands in Europe with New Frankfurt Cloud Region, Delivering the Knowledge Infrastructure for AI to Central European Enterprises

Alongside new availability on AWS eu-central-1, Pinecone announces Nexus knowledge engine, KnowQL query language, Marketplace, Builder tier, and native full-text search FRANKFURT, Germany, May 5, 2026 /PRNewswire/ — Pinecone, the knowledge infrastructure for AI at scale, today announced its expansion into the AWS Europe (Frankfurt) Region (eu-central-1), bringing its full serverless vector database and knowledge infrastructure […]

TechnologyTechnology & Infrastructure
Arxiv· 4d ago

ZAYA1-8B Technical Report

arXiv:2605.05365v1 Announce Type: new Abstract: We present ZAYA1-8B, a reasoning-focused mixture-of-experts (MoE) model with 700M active and 8B total parameters, built on Zyphra's MoE++ architecture. ZAYA1-8B's core pretraining, midtraining, and supervised fine-tuning (SFT) were performed on a full-stack AMD compute, networking, and software platform. With under 1B active parameters, ZAYA1-8B matches or exceeds DeepSeek-R1-0528 on several challenging mathematics and coding benchmarks, and remains competitive with substantially larger open-weight reasoning models. ZAYA1-8B was trained from scratch for reasoning, with reasoning data included from pretraining onward using an answer-preserving trimming scheme. Post-training uses a four-stage RL cascade: reasoning warmup on math and puzzles; a 400-task RLVE-Gym curriculum; math and code RL with test-time compute traces and synthetic code environments built from competitive-programming references; and behavioral RL for chat and instruction following. We also introduce Markovian RSA, a test-time compute method that recursively aggregates parallel reasoning traces while carrying forward only bounded-length reasoning tails between rounds. In TTC evaluation, Markovian RSA raises ZAYA1-8B to 91.9\% on AIME'25 and 89.6\% on HMMT'25 while carrying forward only a 4K-token tail, narrowing the gap to much larger reasoning models including Gemini-2.5 Pro, DeepSeek-V3.2, and GPT-5-High.

TechnologyAdoption & Impact
Arxiv· 4d ago

Operationalizing Ethics for AI Agents: How Developers Encode Values into Repository Context Files

arXiv:2605.05584v1 Announce Type: cross Abstract: As AI coding agents become embedded in software development workflows, developers are beginning to operationalize ethical principles by encoding behavioral rules into repository-level context files for AI agents, such as AGENTS.md files. Rather than examining the ethics of AI agents in the abstract, this vision paper investigates how ethics and values are already being translated for AI agents into actionable instructions that shape agent behavior. Through a preliminary investigation, we find that developers are already embedding guidance related to fairness, accessibility, sustainability, tone, and privacy. These artifacts function as a developer-authored governance layer, translating abstract principles into situated, natural-language directives within development workflows. We outline a research agenda for studying this emerging practice, including how encoded values vary across communities, what governance dynamics emerge when multiple contributors negotiate these files, and whether agents reliably adhere to the constraints specified. Understanding how ethics and values are operationalized for AI agents is essential to ground AI governance in modern software engineering practice.

TechnologyEconomics & Markets
Arxiv· 4d ago

Artificial Aesthetics: The Implicit Economics of Valuing AI-Generated Text

arXiv:2605.05578v1 Announce Type: new Abstract: Aesthetic qualities command measurable premiums in traditional goods markets. However, it remains unclear whether users are willing to pay for such qualities in AI-generated text. This paper estimates the willingness to pay for aesthetic attributes in large language model outputs using an online experiment with N = 117 participants. Participants evaluated responses from four anonymized models across academic, professional, and personal contexts, rated outputs along multiple dimensions, and submitted bids for access using a Becker-DeGroot-Marschak (BDM) mechanism. We find no statistically significant relationship between perceived aesthetic quality and willingness to pay. While participants systematically distinguish between outputs and exhibit consistent preferences over stylistic features, these differences do not translate into higher monetary valuation. Further analysis shows that aesthetic and functional attributes load onto a single latent factor, suggesting that users perceive quality as a unified construct rather than a separable aesthetic dimension. These results imply that, in current large language model (LLM) markets, aesthetic improvements function as baseline expectations rather than sources of price differentiation.

TechnologyTechnology & Infrastructure
Arxiv· 4d ago

From History to State: Constant-Context Skill Learning for LLM Agents

arXiv:2605.05413v1 Announce Type: new Abstract: Large language model (LLM) agents are increasingly used to operate browsers, files, code and tools, making personal assistants a natural deployment target. Yet personal agents face a privacy-cost-capability tension: cloud models execute multi-step workflows well but expose sensitive intermediate context to external APIs, while local models preserve privacy but remain less reliable. Both settings also pay repeatedly for long skill prompts and growing histories. We propose constant-context skill learning, a context-to-weights framework for recurring agent workflows: reusable procedures are learned in lightweight task-family modules, while inference conditions only on the current observation and a compact state block. A deterministic tracker renders this state block from task progress and supplies aligned subgoal rewards, so each module can be trained with step-level SFT and refined through online RL. Across ALFWorld, WebShop, and SciWorld, our agents achieve strong performance across Qwen3-4B, Qwen3-8B and Llama-3.1-8B. With Qwen3-8B, SFT+RL reaches 89.6\% unseen success on ALFWorld, 76.8\% success on WebShop, and 66.4\% unseen success on SciWorld. They match or exceed strong published agent-training results while reducing prompt tokens per turn by 2--7$\times$ relative to controlled ReAct prompting baselines, showing that procedural context can be moved from prompts into weights.

Technology
AIMultiple· 4d ago

Top 20+ AI Chip Makers: NVIDIA & Its Competitors

Despite the restored access to high-end hardware, the added costs and supply chain complexities continue to incentivize the Chinese government and chip industry to develop competitive local alternatives. While Chinese chips currently underperform NVIDIA’s latest technology, these trade barriers ensure that domestic development remains a strategic priority, potentially challenging NVIDIA’s market dominance in the future.10 · While NVIDIA dominates the AI ...

Technology
Benzinga· 4d ago

Super Micro, NVIDIA Chips And Alibaba: Inside The Explosive AI Smuggling Allegations - NVIDIA (NASDAQ:NVD - Benzinga

Washington's export restrictions on NVIDIA chips and the broader crackdown on smuggling have sharply disrupted China's AI hardware market, driving up server prices, tightening supply, and reshaping competition among semiconductor players.

TechnologyGeopolitics
UC Today· 4d ago

European Tech CEOs Want Easier AI Rules: What It Means for UC Security and Compliance Leaders - UC Today

UC Today delivers insights for IT leaders and buyers covering Agentic AI, Agentic AI in the Workplace​, AI Agents, Call Recording, Communication Compliance​, Security and Compliance, collaboration, employee experience and workspace tech.

TechnologyLabor & Society
Daily Brew· 4d ago

Chrome removes claim of On-device AI not sending data to Google Servers

Reports indicate that Google has updated its messaging regarding Chrome's on-device AI features, removing previous claims about data privacy.

TechnologyTechnology & Infrastructure
Daily AI News May 8, 2026: OpenAI Turns Voice AI into Real Capability· 4d ago

Advancing Voice Intelligence with New Models in the API

OpenAI introduced GPT-Realtime-2, GPT-Realtime-Translate, and GPT-Realtime-Whisper to support real-time voice reasoning, translation, and conversational context in AI applications.

TechnologyEconomics & Markets
Daily Brew· 4d ago

GPT-5.5 may burn fewer tokens, but it always burns more cash

An analysis of the economic trade-offs of GPT-5.5, noting that while token efficiency has improved, operational costs continue to rise.

Technology
Information Week· 5d ago

The AI infrastructure boom is coming for enterprise budgets

As chipmakers raise forecasts for CPU market growth and AI vendors increase their spending plans, CIOs may end up being the ones to absorb the cost.

PaywallTechnologyAdoption & Impact
FT· 5d ago

Why software firms are calling time on the SaaSpocalypse

SaaS companies can stay relevant by housing customer data and layering AI on top

TechnologyTechnology & Infrastructure
Daily Brew· 5d ago

Higher usage limits for Claude and a compute deal with SpaceX

Anthropic has announced increased usage limits for Claude and a new compute partnership with SpaceX.

PaywallTechnologyEconomics & Markets
Bloomberg· 5d ago

Cloudflare Forecast Misses Estimates, Reveals Job Cuts

Cloudflare Inc. stock fell on Thursday after the company said it would slash jobs and gave a forecast for revenue that fell short of analysts’ expectations.

Technology
Wccftech· 5d ago

Arm Doubles AGI CPU Revenue Forecast to $2 Billion by 2028 as OpenAI, Cerebras, and Hyperscalers Pile Into Agentic AI Orders

Arm's AGI CPU has seen explosive demand amidst the Agentic AI boom, as the company now reports double the revenue by 2028.

Technology
Domain-b· 5d ago

Asia’s chip giants power the AI rally as hardware demand surges | Domain-b.com

Asian semiconductor giants TSMC, Samsung, and SK hynix are benefiting from booming AI infrastructure demand as HBM memory and advanced chips power the global AI rally.

TechnologyEconomics & Markets
Artificial Intelligence Newsletter | May 7, 2026· 5d ago

Apple Intelligence probe in South Korea stalled for over a year, Seoul YMCA says

The Seoul YMCA has criticized the Korea Fair Trade Commission for failing to advance an investigation into Apple's advertising of 'Apple Intelligence' while consumers remain without compensation.

Technology
Tech Startups· 5d ago

Top Startup and Tech Funding News – May 7, 2025 - Tech Startups

Today is Thursday, May 7, 2026, and we’re back with the latest startup and tech funding news from around the world. Today’s funding activity highlights where investors continue placing their biggest bets: sovereign AI infrastructure, quantum computing, satellite communications, defense ...

TechnologyAdoption & Impact
Arxiv· 5d ago

cotomi Act: Learning to Automate Work by Watching You

arXiv:2605.03231v1 Announce Type: new Abstract: What if a browser agent could learn your work simply by watching you do it? We present cotomi Act, a browser-based computer-using agent that combines reliable multi-step task execution with persistent organizational knowledge learned from user behavior. For execution, an agent scaffold with adaptive lazy observation, verbal-diff-based history compression, coarse-grained actions, and test-time scaling via best-of-N action selection achieves 80.4% on the 179-task WebArena human-evaluation subset, exceeding the reported 78.2% human baseline. For organizational knowledge, a behavior-to-knowledge pipeline passively observes the user's browsing and progressively abstracts it into artifacts (task boards, wiki) exposed through a shared workspace editable by both user and agent. A controlled proxy evaluation confirms that task success improves as behavior-derived knowledge accumulates. In our live demonstration, attendees interact with the system in a real browser, issuing tasks and observing end-to-end autonomous execution and shared knowledge management.

TechnologyAdoption & Impact
Bebeez· 5d ago

France’s OpsMill raises €11.9 million to help enterprises prepare infrastructure data for AI and automation

OpsMill, a Paris-based infrastructure data management company, has raised €11.9 million ($14 million) in Series A funding to grow its engineering and product teams and continue developing data-centric AIOps solutions.  The round was led by IRIS with participation from BGV and existing investors Serena and Partech. The company aims to transform fragmented IT data into […]

PaywallTechnologyEconomics & Markets
Bloomberg· 5d ago

Andreessen Horowitz in Funding Round for Swedish AI Startup Pit

Andreessen Horowitz has led a $16 million funding round for Pit, an artificial intelligence startup based in Sweden.

PaywallTechnologyEconomics & Markets
Bloomberg· 5d ago

Kimi Chatbot Maker Moonshot AI Valued at $20 Billion in Meituan-Led Round

Moonshot AI has raised about $2 billion in its latest funding round, signaling growing investor appetite for Chinese startups rivaling Silicon Valley’s leaders.

PaywallTechnologyEconomics & Markets
Bloomberg· 5d ago

Korea Surpasses Canada as World’s Seventh-Largest Stock Market

South Korea’s equity market has overtaken Canada’s as the world’s seventh largest, propelled by insatiable demand for chips powering artificial intelligence.

PaywallTechnologyEconomics & Markets
Bloomberg· 5d ago

Alibaba Shares Outpace Tencent’s as Chip Exposure Fuels Demand

The torrid stock rally in Asian chipmakers is driving a divergence between China’s two internet giants, with Alibaba Group Holding Ltd. gaining an edge over rival Tencent Holdings Ltd. due to investor enthusiasm about its ambitious semiconductor unit.

PaywallTechnologyEconomics & Markets
Bloomberg· 5d ago

Montage Tops CATL as Priciest Dual-Listed Stock After Chip Rally

Montage Technology Co. has overtaken Contemporary Amperex Technology Co. Ltd. as the most expensive dual-listed stock in Hong Kong relative to its mainland shares, propelled by surging demand for AI chips.

PaywallTechnology
Bloomberg· 5d ago

Anthropic Is Making Its Claude Chatbot More Appealing to Consumers

Artificial intelligence startup Anthropic PBC, which originally aimed its Claude chatbot at businesses, is looking to capitalize on recent inroads with consumers by making the software more appealing to everyday people.

Technology
Artificial Intelligence Newsletter | May 8, 2026· 5d ago

EU lawmakers reach deal on amendments to EU's AI Act

EU lawmakers agreed on amendments to the AI Act, including a ban on AI-generated intimate content and a more centralized governance framework to simplify compliance.

TechnologyTechnology & Infrastructure
Arxiv· 5d ago

Enhancing Agent Safety Judgment: Controlled Benchmark Rewriting and Analogical Reasoning for Deceptive Out-of-Distribution Scenarios

arXiv:2605.03242v1 Announce Type: new Abstract: Tool-using agent systems powered by large language models (LLMs) are increasingly deployed across web, app, operating-system, and transactional environments. Yet existing safety benchmarks still emphasize explicit risks, potentially overstating a model's ability to judge deceptive or ambiguous trajectories. To address this gap, we introduce ROME (Red-team Orchestrated Multi-agent Evolution), a controlled benchmark-construction pipeline that rewrites known unsafe trajectories into more deceptive evaluation instances while preserving their underlying risk labels. Starting from 100 unsafe source trajectories, ROME produces 300 challenge instances spanning contextual ambiguity, implicit risks, and shortcut decision-making. Experiments show that these challenge sets substantially degrade safety-judgment performance, with hidden-risk cases remaining particularly non-trivial even for recent frontier models. We further study ARISE (Analogical Reasoning for Inference-time Safety Enhancement), a retrieval-guided inference-time enhancement that retrieves ReAct-style analogical safety trajectories from an external analogical base and injects them as structured reasoning exemplars. ARISE improves judgment quality without retraining, but is best viewed as a task-specific robustness enhancement rather than a standalone safety guarantee. Together, ROME and ARISE provide practical tools for stress-testing and improving agent safety judgment under deceptive distribution shifts.

TechnologyTechnology & Infrastructure
Arxiv· 5d ago

Evaluating Prompting and Execution-Based Methods for Deterministic Computation in LLMs

arXiv:2605.03227v1 Announce Type: new Abstract: Large Language Models (LLMs) have demonstrated strong capabilities in natural language understanding and reasoning. However, their ability to perform exact, deterministic computation remains unclear. In this work, we systematically evaluate multiple prompting strategies, including Chain-of-Thought (CoT), Least-to-Most decomposition, Program-of-Thought (PoT), and Self-Consistency (SC), on tasks requiring precise and error-free outputs, including binary counting, longest substring detection, and arithmetic evaluation. To support this study, we introduce a synthetic dataset with diverse natural language instructions, enabling controlled evaluation of exact computation across multiple task types. Our results show that standard prompting methods achieve only moderate accuracy on sequence-based tasks. CoT provides limited improvement, while Least-to-Most suffers from error accumulation. In contrast, PoT achieves perfect accuracy by generating executable code and delegating computation to an external interpreter. Self-Consistency improves robustness through majority voting, but incurs substantial computational overhead. We further train a small domain-specific model (CodeT5-small) to generate executable programs, which achieves perfect accuracy on held-out synthetic test data across all tasks with minimal training cost. Overall, our findings suggest that LLMs may simulate reasoning patterns rather than reliably perform exact symbolic computation. For deterministic tasks, combining LLMs with external tools or using specialized models provides a more reliable and efficient solution.

TechnologyTechnology & Infrastructure
Arxiv· 5d ago

Learning Correct Behavior from Examples: Validating Sequential Execution in Autonomous Agents

arXiv:2605.03159v1 Announce Type: new Abstract: As autonomous agents become increasingly sophisticated, validating their sequential behavior presents a significant challenge. Traditional testing approaches require manual specification, exact sequence matching, or thousands of training examples. We present a novel algorithm that automatically learns correct behavior from just 2-10 passing execution traces and validates new executions against this learned model. Our approach combines dominator analysis from compiler theory with multimodal large language model-powered semantic understanding to identify essential states and handle non-deterministic behavior. The system constructs a generalized ground truth model using Prefix Tree Acceptors, merges traces through multi-tiered equivalence detection, and validates new executions via topological subsequence matching. In controlled experiments, our system achieved high accuracy in detecting product bugs and false successes using only 3 training traces. This approach provides explainable validation results with coverage metrics and works across diverse domains including UI testing, code generation, and robotic processes.

PaywallTechnologyAdoption & Impact
Microsoft Blogs· 5d ago

The state of global AI diffusion in 2026 - Microsoft On the Issues

Today we published our latest Global AI Diffusion Report. The global adoption of artificial intelligence continued to rise in the first quarter of 2026. During the quarter, AI usage increased by 1.5 percentage points from 16.3% to 17.8% of the world’s working age population.

Technology
Artificial Intelligence Newsletter | May 8, 2026· 5d ago

Vietnam finalizes AI implementation decree, relaxes incident-reporting timelines

Vietnam released a final decree guiding its AI law, clarifying risk classification and content labeling. The decree eases incident-reporting timelines compared to earlier drafts.

Technology
CIO Dive· 5d ago

Why AI regulation is now an operating model | CIO Dive

CIOs must know where AI is deployed, manage risk across the lifecycle and produce evidence without scrambling.

TechnologyEconomics & Markets
Arxiv· 5d ago

Terminus-4B: Can a Smaller Model Replace Frontier LLMs at Agentic Execution Tasks?

arXiv:2605.03195v1 Announce Type: new Abstract: Modern coding agents increasingly delegate specialized subtasks to subagents, which are smaller, focused agentic loops that handle narrow responsibilities like search, debugging or terminal execution. This architectural pattern keeps the main agent's context window clean by isolating verbose outputs (e.g. build logs, test results, etc.) within the s

TechnologyGeopolitics
BBC· 5d ago

Meta brings High Court challenge over Ofcom fees

Meta said Ofcom's calculations were "disproportionate"; the regulator said it would defend its position.

Technology
Daily Brew· 5d ago

Pennsylvania sues Character.AI chatbot posing as doctor

The state of Pennsylvania has filed a lawsuit against Character.AI regarding a chatbot that was providing medical and psychological advice.

TechnologyTechnology & Infrastructure
Tom's Hardware· 5d ago

Motherboard sales 'collapse' by more than 25% as chipmakers strangle enthusiast PC market to build more AI chips — Asus projected to sell 5 million fewer boards in 2025, Gigabyte, MSI, and ASRock also expected to see reduced sales numbers | Tom's Hardware

Fewer people are buying parts and building new PCs from scratch.

Technology
🤝 Anthropic, SpaceX(AI) become unlikely compute partners· 5d ago

Anthropic, SpaceX(AI) become unlikely compute partners

Anthropic and SpaceX have entered into a partnership focused on compute resources.

TechnologyEconomics & Markets
Theregister· 5d ago

Neocloud IREN buys OpenStack champion Mirantis

Former bitcoin miner plans to build an easier cloudy AI on ramp while remaining a friend to FOSS

TechnologyTechnology & Infrastructure
Arxiv· 5d ago

DAO-enabled decentralized physical AI: A new paradigm for human-machine collaboration

arXiv:2605.04522v1 Announce Type: cross Abstract: We propose DAO-enabled decentralized physical AI (DePAI), a democratic architecture for coordinating humans and autonomous machines in the operation and governance of physical-digital systems. We (1) synthesize foundations in blockchains, decentralized autonomous organizations (DAOs), and cryptoeconomics; (2) connect DAO design with digital-democracy research on deliberation and voting, showing how each can advance the other; (3) position DAO-governed decentralized physical infrastructure networks (DePIN) within a vertically integrated stack that links energy and sensing to connectivity, storage/compute, models, and robots; (4) show how these elements specify workflows that couple machine execution with human oversight, enabling enhanced self-organization of techno-socio-economic systems, which we call DePAI; and (5) analyze risks, including security, centralization, incentive failure, legal exposure, and the crowding-out of intrinsic motivation, and argue for value-sensitive design and continuously adaptive governance. DePAI offers a path to scalable, resilient self-organization that integrates physical infrastructure, AI, and community ownership under transparent rules, on-chain incentives, and permissionless participation, aiming to preserve human autonomy.

PaywallTechnologyTechnology & Infrastructure
NYT· 5d ago

Anthropic’s C.E.O. Says It Could Grow by 80 Times This Year

The chief executive, Dario Amodei, said the rapid growth had exponentially increased the start-up’s need for more computing power.

TechnologyTechnology & Infrastructure
Theregister· 5d ago

Anthropic response to 1-click pwn: Shouldn't have clicked 'ok'

Security biz Adversa AI argues users of AI tools need clearer warnings

Technology
Siliconrepublic· 5d ago

Anthropic joins forces with SpaceX for Colossus capacity

Anthropic has also ‘expressed interest’ to co-develop data centres in space, said SpaceX. Read more: Anthropic joins forces with SpaceX for Colossus capacity

TechnologyEconomics & Markets
Business Today· 5d ago

India’s AI push gathers pace as enterprise tech spending set to rise 6 8% in 2026: Bain - BusinessToday

Bain said around 40% of enterprise technology budgets in 2026 will be allocated toward change initiatives, and nearly 40%-45% of that spend is expected to go toward AI and data-led transformation projects.

TechnologyEconomics & Markets
Guardian· 5d ago

Europe’s AI translation industry told it risks reputation by partnering with US firms

Partnership between top startup DeepL and Amazon comes amid concern about Silicon Valley’s monopoly over digital infrastructure AI companies in Europe risk losing their world-leading status in the field of machine translation, industry figures have said, after the decision by one of the continent’s leading startups to partner with Amazon’s cloud computing division provoked alarm. While businesses in the EU have generally lagged behind the US and China in AI adoption, a small group of European companies have cornered the global market for high-quality machine translations for professional use. Continue reading...

TechnologyEconomics & Markets
Investing.com· 5d ago

Morning Bid: No stopping AI frenzy in Asia By Reuters

All intellectual property rights are reserved by the providers and/or the exchange providing the data contained in this website. Fusion Media may be compensated by the advertisers that appear on the website, based on your interaction with the advertisements or advertisers. © 2007-2026 - Fusion ...

Technology
Daily Brew· 5d ago

Anthropic's $200B Google Cloud Deal Boosts Alphabet

Anthropic is reportedly set to invest $200 billion in Google Cloud over five years, marking a significant shift in cloud-and-AI financing.

Technology
Theregister· 5d ago

Mozilla boasts Mythos boosted Firefox bug cull

Yet it remains unclear if Anthropic's uber model was effective, or if better model middleware is what makes the difference

TechnologyEconomics & Markets
The Information· 5d ago

Ex-OpenAI Researcher’s Six-Week-Old Startup Targets Funding at $4 Billion Valuation

Nvidia’s investment spree in startups that use its AI chips has stoked traditional venture capital firms’ interest in those companies. That has paved the way for brand-new startups taking on Anthropic and OpenAI to raise back-to-back funding rounds, sometimes in a matter of weeks.

TechnologyAdoption & Impact
Computer Weekly· 5d ago

Nutanix CEO maps out agentic AI strategy, targets VMware defectors | Computer Weekly

Rajiv Ramaswami talks up the Nutanix’s agentic AI play, the growing demand for sovereign cloud capabilities, and why decoupling storage from HCI is hastening migrations from VMware

PaywallTechnologyEconomics & Markets
FT· 5d ago

Old IT makes its bid for AI relevance

Pendulum is swinging back to companies in areas such as servers, more general chips and software

Technology
Reuters· 5d ago

Asia's tech giants give AI bull run a new centre of gravity | Reuters

Just as the world's AI bulls looked to be running out of puff, a fresh investor frenzy has hit Asia's tech names, making Seoul's stock market the world's hottest and delivering bonuses of half a million dollars to workers at one ​Korean chipmaker.

Technology
Siliconrepublic· 5d ago

Moonshot AI valued at $20bn after $2bn raise for Kimi creator

China’s AI chatbot labs are currently attracting big investors. Read more: Moonshot AI valued at $20bn after $2bn raise for Kimi creator

PaywallTechnologyEconomics & Markets
FT· 5d ago

The mysterious $53bn ‘other income’ boost to AI hyperscaler earnings

Quantum entanglement

AI Daily Brief — leaders actually read it.

Free email — not hiring or booking. Optional BPAI updates for company news. Unsubscribe anytime.

Include

No spam. Unsubscribe anytime. Privacy policy.