DD
MM
YYYY

PAGES

DD
MM
YYYY

spot_img

PAGES

Home Blog Page 6

Inside Nvidia’s Blackwell Rollout at Microsoft: Internal Email Flags “Wasteful” Cooling as AI Data Center Demands Surge

0

As Nvidia races to deploy its most powerful GPUs into Microsoft’s global data-center network, an internal note from early fall shows that the rollout hasn’t been entirely seamless.

In the email, an Nvidia Infrastructure Specialists (NVIS) team member questioned whether Microsoft’s cooling strategy at one facility was “wasteful,” underscoring the mounting resource pressures tied to the AI boom.

The observation came during the installation of GB200 Blackwell servers supporting OpenAI workloads — a reminder of how intertwined Nvidia, Microsoft, and OpenAI have become as the global AI arms race accelerates.

Nvidia announced its Blackwell architecture in March 2024, touting it as roughly twice as powerful as the preceding Hopper generation. The GB200 series was the first wave of Blackwell systems shipped to hyperscalers, followed now by the newer GB300 generation already making its way into top-tier data centers.

The email described the setup of two GB200 NVL72 racks — each holding 72 GPUs — installed for OpenAI via Microsoft’s cloud infrastructure. Given the extreme heat density created by these multi-GPU clusters, Nvidia’s systems rely on liquid cooling inside the racks. But a second facility-wide cooling layer is still required to expel the heat, and this is where the Nvidia staffer raised concerns.

Microsoft’s approach “seems wasteful due to the size and lack of facility water use,” the Nvidia employee wrote, though the note acknowledged that the design offered flexibility and fault tolerance.

Shaolei Ren, an associate professor of electrical and computer engineering at the University of California who studies data-center resource use, offered context consistent with the internal critique. He said the Nvidia staffer was likely referring to the building-level cooling stage. In some facilities, Microsoft uses an air-based system at this second stage rather than a water-based one.

“This type of cooling system tends to be using more energy,” Ren said, “but it doesn’t use water.”

Ren added that operators face a “trade-off” between water use and energy consumption. Air cooling consumes more power, but it avoids the public pushback that often comes with heavy water usage — now one of the most sensitive environmental issues facing hyperscalers and AI developers around the world.

Microsoft confirmed that the installation used a closed-loop liquid cooling heat exchanger inside an air-cooled facility.

“Microsoft’s liquid cooling heat exchanger unit is a closed-loop system that we deploy in existing air-cooled data centers to enhance cooling capacity,” a spokesperson told Business Insider.

The company said the hybrid approach allows it to scale AI infrastructure using its existing footprint while maintaining efficient heat dissipation and “optimizing power delivery to meet the demands of AI and hyperscale systems.”

The debate around cooling is no longer technical background noise — it is central to the politics and environmental footprint of AI expansion. In regions from Europe to the American Southwest, community groups and local governments have pushed back against hyperscale data centers over water use, energy intensity, and strain on local grids. Ren noted that companies weigh energy costs, water costs, and “publicity cost” in deciding which cooling strategy to adopt.

Microsoft insists it remains on track to meet its self-imposed 2030 goals to become “carbon negative, water positive, and zero waste.” It has also announced a “zero-water” cooling design for future facilities, along with advances in on-chip cooling intended to reduce thermal load at the processor level.

The Nvidia memo, while flagging cooling as an area of inefficiency, also described common early-deployment challenges. Blackwell’s first large-scale rollouts require close coordination between Nvidia and hyperscaler staff, and the email said on-site support was critical. Validation documentation needed extensive rewriting, and handover processes between Microsoft and Nvidia “required a lot more solidification.”

Still, the note suggested that Nvidia’s production hardware quality has improved significantly from the early qualification samples customers received before the formal launch. Both NVL72 racks deployed at the facility achieved a 100% pass rate on compute performance tests.

Nvidia, in its public response, said Blackwell systems “deliver exceptional performance, reliability, and energy efficiency” and that companies such as Microsoft have already deployed “hundreds of thousands” of GB200 and GB300 NVL72 systems.

The episode offers a glimpse into the intense, resource-hungry infrastructure race beneath today’s AI boom. Nvidia is under pressure to deliver chips fast enough. Microsoft is under pressure to build cooling and power capacity fast enough. And the entire industry is under pressure to justify its environmental footprint as AI workloads grow at a pace no one predicted even three years ago.

The friction is almost inevitable. Blackwell is far more powerful — and far more thermally demanding — than anything that came before it. As deployments scale into the hundreds of thousands, cooling will remain one of the most contested, expensive, and politically sensitive aspects of the AI infrastructure buildout.

And if the industry’s trajectory is any guide, the next-generation chips coming after Blackwell will only intensify that battle.

AI Reshapes Retail: Conversational Commerce To Drive $263 Billion in Holiday Sales

0

The laborious process of holiday shopping, traditionally a complex matrix of deciding, price comparing, and review checking, is being fundamentally disrupted by the rise of conversational Artificial Intelligence (AI).

Consumers are now seamlessly integrating generative AI platforms like OpenAI’s ChatGPT, Google’s Gemini, and Perplexity into their buying journeys, treating them as personalized, omniscient shopping assistants. This shift is not merely a trend; it is a seismic commercial event poised to reshape the digital retail landscape and redefine the meaning of online presence.

According to market data published by CNBC, AI has quickly become an indispensable driver of sales and customer engagement this holiday season. Salesforce projects that AI will propel a staggering $263 billion in global online holiday sales, accounting for a significant 21% of all holiday orders globally in 2025. This massive figure is supported by high consumer adoption, with various surveys indicating that between 40% and 83% of consumers are planning to utilize AI tools for their shopping needs this year.

The value of an AI-fueled customer is proving demonstrably higher for retailers. Data from Adobe shows that shoppers arriving on retail websites via generative AI platforms are 30% more likely to make a purchase and are 14% more engaged—evidenced by spending more time on the site and having a lower immediate bounce rate—compared to those from traditional non-AI sources. Consequently, these AI-driven visits are generating 8% more revenue per session.

This phenomenon is transforming product discovery. For many shoppers, the AI assistant acts as a curator, unearthing relevant brands that might otherwise be overlooked. This is leading to significant shifts in purchasing patterns, as evidenced by one retail tech CEO who reported that approximately half of the gifts she bought came from brands she had never encountered before the AI guided her search.

The new customer journey starts with a conversational query, such as, “Where can I find the best gift under $50 for a tech-savvy teenager who cares about sustainability?” The AI then serves up a tailored, product-level conclusion, rather than a list of generic links.

The New SEO: Pivoting to Answer Engine Optimization (AEO)

The profound shift in customer search behavior is compelling retailers to abandon outdated digital strategies centered around traditional Search Engine Optimization (SEO) in favor of Answer Engine Optimization (AEO). SEO, a game of keyword density and link building designed to rank links on a search results page, is proving ineffective in the age of conversational AI, where the goal is to be the single, definitive answer the AI provides.

AEO focuses on semantics, content structure, and credibility to ensure a brand’s products are chosen by the AI model. Since AI platforms prioritize providing a single, highly relevant, and trustworthy answer, brands are now engaged in a critical restructuring of their content:

  • Conversational Structure: Content must be optimized for natural language and provide direct answers to questions in a concise, authoritative format, often using an upfront summary or rich answer blocks.
  • Rich Data and Trust: Retailers must provide highly detailed, trustworthy data, including real-time inventory, clear product specifications, brand certifications, and aggregated customer feedback, rather than simply relying on basic attributes.
  • Solution-Focused Content: Brands like Ethique Beauty have shifted their content strategy from describing products (e.g., “shampoo bar”) to addressing customer problems (e.g., “solution for oily scalp,” or “how to sleep with curly hair”). This approach aligns with the conversational nature of AI queries, which often begin with a problem or need rather than a specific product name.

Retailers are now directing funds away from traditional paid social media and search campaigns, which are seeing performance declines, and into the infrastructure required for AEO. This strategic pivot, while requiring significant internal and consulting investment, yields a high return on investment because the customer delivered by the AI is typically more qualified and further along the purchasing funnel.

The Retail Giants’ AI Strategy Split and Internal Tools

The country’s biggest retailers are pursuing distinct and competitive strategies to win the AI shopper, primarily defined by their relationship with external large language model (LLM) platforms:

Retailer External Strategy (Partnership) Internal AI Assistant & Features Key Advantage
Walmart Strategic partnership with OpenAI (ChatGPT), enabling single-item Instant Checkout directly from the chat. Sparky: Conversational agent on its app for party planning, review summaries, and in-store navigation. Wide net casting via ChatGPT integration, capturing early-stage advice seekers.
Target Strategic partnership with OpenAI (ChatGPT), offering a multi-item purchasing app within ChatGPT, including groceries and selecting fulfillment (Drive Up/Pickup). Target Gift Finder: AI-powered tool for higher-engagement, larger cart sizes, trained to understand conversational searches. Deep functionality and omnichannel experience (groceries, pickup) within the chat interface.
Amazon Exclusionary Stance against external LLMs (blocking crawlers, sending cease-and-desist to Perplexity AI). Rufus: In-house conversational shopping assistant that provides product comparisons, personalized deals, and can automate purchases based on price alerts. Control over its vast product data and proprietary AI agents with high automation (agentic AI).

Amazon’s aggressive approach to blocking external AI crawlers, coupled with its focus on its powerful internal assistant Rufus, signals a strong belief that its massive data moat is a competitive advantage that should not be shared. Conversely, Walmart and Target are betting that placing their products where the customer is starting the conversation—in ChatGPT—will provide superior reach, even if it means sharing the customer experience.

Walmart CEO Doug McMillon has championed this, stating that agentic AI will be a core growth driver, helping people “save time and have more fun shopping.” Target has observed that its Gift Finder tool is already driving higher engagement and larger shopping carts than its previous tools, highlighting the superior performance of conversational, problem-solving AI.

Where AI Falls Short

Despite the momentum, AI shopping tools are not yet universally perfect. Resistance to adoption persists due to functional and psychological barriers, including perceived complexity, risk concerns, and value deficits compared to traditional methods.

When the technology fails to meet conversational expectations, the results can be frustrating. One startup founder, Diana Tan, provided detailed body type, preference, and budget information to a chatbot for a capsule wardrobe recommendation. The tool repeatedly suggested “boring basics” like black shirts and gray pants, leading her to feel like she was talking to a “demented grandmother” and ultimately abandoning the tool.

For consumers who enjoy the process of shopping, the highly efficient, hyper-focused nature of AI can eliminate the serendipitous discovery of browsing. As Tan noted, it “takes the joy out of shopping.”

These instances show that while AI excels at research and optimizing for known intent, it still struggles with tasks requiring deep creative nuance, abstract style sense, or maintaining a human-like, flexible conversation flow. This dichotomy means retailers must continue to maintain a dual strategy, optimizing for both the hyper-efficient AI shopper and the traditional browser, until the AI’s creative and empathetic capabilities mature.

World App Reinvents as ‘Super App,’ Blending Encrypted Chat, Global Crypto Payments, and Biometric Identity

0

World, the ambitious biometric identity verification project co-founded by OpenAI CEO Sam Altman and Alex Blania, has unveiled a major upgrade to its mobile application, transforming it into a decentralized “super app” for the AI era.

The newest version of the World App, developed by the startup Tools for Humanity, deeply integrates high-utility features, including a Signal-equivalent encrypted chat service and vastly expanded cryptocurrency payment capabilities, with its core mission: to provide digital “proof of human” tools to distinguish real people from increasingly sophisticated bots and AI-generated digital fakery.

At a gathering at World’s headquarters, Altman and Blania framed the project as a necessary solution in an age of AI, growing out of conversations about creating a new economic model built on Web3 principles. Altman acknowledged the fundamental difficulty, stating, “It’s really hard to both identify unique people and do that in a privacy-preserving way.”

The flagship new feature is World Chat, a fully end-to-end encrypted messenger that utilizes the XMTP protocol and is designed to meet or exceed the security standards set by privacy-focused messengers like Signal, notably without requiring a phone number or collecting messaging metadata.

World Chat is built to foster trust by leveraging the underlying World ID system to visually and cryptographically verify the identity of chat participants. The platform uses a clear, color-coded system for verification status.

Blue Speech Bubbles with a unique World ID gem indicate a contact has been verified as a unique human via the biometric Orb system.

Gray Speech Bubbles denote an unverified user.

This feature, originally launched in beta in March, is explicitly designed to incentivize verification by allowing users to instantly confirm the authenticity of their chat partner, addressing the rising problem of synthetic accounts, spam, and digital impersonation. According to Tiago Sada, World’s Chief Product Officer, the chat tool was added to create a more interactive experience, fulfilling the demand for a “more social World app” while providing a secure communication method akin to WhatsApp or Telegram, but with Signal-level encryption. Furthermore, messages, media, and crypto payments can all be sent directly from within the chat interface, which also supports Mini Apps for group functionality like polls and games.

Global Financial Access and Seamless Crypto Integration

The second major pillar of the upgrade is a dramatic expansion of the app’s financial utility, converting the World App into a comprehensive global crypto wallet and payment system. The expanded digital payment system allows for seamless Venmo-like capabilities to send and receive over 100 cryptocurrency tokens, including USDC, EURC, and wrapped Bitcoin and Ethereum.

This integration significantly lowers the barrier to financial access for global users:

  • Users can now create virtual bank accounts to receive paychecks directly into the World App.
  • Users can easily make deposits from their external bank accounts.
  • All funds can be instantly converted into crypto via the native USDC and Circle’s Cross-Chain Transfer Protocol (CCTP V2) integration, simplifying global financial access.

Crucially, users do not need to be verified by the World ID authentication system to utilize these expanded financial and transfer features. The platform is actively working on integration for merchant payments and supports the growing ecosystem of over 300 Mini Apps that leverage the World App wallet for decentralized finance (DeFi), gaming, and social media.

Scaling the Biometric “Proof of Human” Network

Despite the focus on new utility, the project’s ultimate success hinges on scaling its unique identity verification layer, the World ID. This requires users to undergo an iris scan via the Orb—a specialized, purpose-built device utilizing NVIDIA Jetson processors and AI models—which converts the iris’s unique pattern into a secure, encrypted digital code (an IrisCode). This World ID then allows users to interact with services that require “Proof of Personhood,” without revealing their personal data through the use of zero-knowledge proofs (ZKPs).

While the company aims to verify billions of people, Tools for Humanity claims to have already surpassed 27 million users and is rapidly expanding. To overcome the logistical and psychological barriers of the fixed-location Orb, the company is actively developing solutions to streamline verification.

The introduction of the Orb Minis—portable, smartphone-like verification devices—will allow verification to scale by a factor of ten, bringing the technology directly into people’s hands and homes. The company plans to deploy 7,500 Orbs across the United States alone by the end of 2025 as part of a significant global expansion effort.

The long-term plan involves potentially turning the Orb Minis into mobile point-of-sale devices or licensing the ID sensor technology to device manufacturers, which would drop the verification barrier significantly and accelerate global adoption.

The rollout of these social and financial features aims to build retention and utility, making the World App an indispensable daily platform and furthering its mission to establish a global, privacy-preserving identity primitive for the age of advanced AI.

Do Kwon to Be Sentenced in New York After TerraUSD Crash That Wiped Out $40bn and Shook the Global Crypto Industry

0

Do Kwon, the South Korean cryptocurrency entrepreneur whose TerraUSD and Luna tokens imploded in 2022 and erased an estimated $40 billion in market value, is scheduled to be sentenced on Thursday in a New York federal court.

The hearing before U.S. District Judge Paul A. Engelmayer is set for 11 a.m. local time (1600 GMT) in Manhattan, marking a decisive moment in one of the largest crypto fraud cases to reach the U.S. courts.

Kwon, 34, co-founded the Singapore-based Terraform Labs and developed TerraUSD, a so-called algorithmic stablecoin that he claimed would hold its $1 value even in volatile markets. He also created Luna, the token that was designed to help maintain TerraUSD’s peg. He previously pleaded guilty to two criminal counts — conspiracy to defraud and wire fraud — and admitted he misled investors about the stability mechanisms supporting the coin.

“I made false and misleading statements about why it regained its peg by failing to disclose a trading firm’s role in restoring that peg,” Kwon said in court. “What I did was wrong.”

The sentencing follows a sweeping set of charges filed by U.S. prosecutors in January, which originally included nine counts covering securities fraud, wire fraud, commodities fraud, and money laundering conspiracy. Prosecutors say Kwon’s actions inflicted massive losses on investors around the world and set off a chain reaction in the crypto market. When TerraUSD collapsed in 2022, the shockwave contributed to the downfall of some of the industry’s most prominent firms and accelerated a slide in digital asset prices that wiped out tens of billions more.

Kwon’s case sits at the center of a broader crackdown on cryptocurrency executives after the market’s dramatic downturn. Multiple companies failed in rapid succession in 2022, and several top industry figures have since faced criminal prosecution or regulatory action. The Justice Department has argued that these cases demonstrate the extent to which fraud had spread across the sector during its boom years.

Prosecutors have asked the court to impose a sentence of at least 12 years, saying the magnitude of the losses and the market disruption justify a substantial prison term. They argue that Kwon’s manipulation of TerraUSD’s value in 2021 — when he told investors that a computer algorithm known as the Terra Protocol had restored the coin’s peg, even though a high-frequency trading firm had secretly been enlisted to buy millions of dollars’ worth of tokens — was a deliberate and calculated deception.

Kwon’s lawyers are pushing for a sentence of no more than five years, saying he needs to return to South Korea to face separate criminal charges there. As part of his plea agreement, U.S. prosecutors said they would not oppose a request for Kwon to be transferred abroad after he serves half of his U.S. sentence.

His legal problems extend beyond the criminal case in New York. In 2024, Kwon agreed to pay an $80 million civil fine and accept a ban on crypto-related activities as part of a sweeping $4.55 billion settlement reached with the U.S. Securities and Exchange Commission alongside Terraform Labs. He also remains wanted in South Korea, where authorities have pursued him since the crash and accused him of violating financial laws.

The sentencing marks the culmination of a saga that reshaped global attitudes toward digital assets. TerraUSD’s collapse became a turning point for regulators worldwide, prompting tighter scrutiny over stablecoins, greater demands for transparency, and mounting pressure on exchanges and crypto issuers to prove the legitimacy of their reserves and business models.

For many investors burned in the collapse, Thursday’s hearing is the closest step yet toward accountability in a market that operated for years with minimal oversight.

Google Launches Gemini Deep Research, Locking Into a New-Era AI Arms Race with OpenAI Hours After GPT-5.2 Launch

0

Google on Thursday rolled out a revamped version of its Gemini Deep Research agent, rebuilding the tool on top of Gemini 3 Pro, the company’s latest flagship foundation model.

The new system is designed not only to generate long-form research reports, but also to function as an embeddable research engine that other developers can plug directly into their own applications.

The timing was not lost on anyone in the industry. What once looked like a broad field of AI competitors is becoming narrowed into a head-to-head sprint between two giants that now announce upgrades within the same news cycle, each determined to stay a step ahead in the escalating race for model dominance.

Google’s upgraded Deep Research system is no longer just a tool for generating polished research reports. It has broadened into a programmable research agent suited for large-scale information synthesis, built to survive massive context loads and long reasoning cycles that would normally push an LLM into error territory.

The company is positioning it as a backbone for due diligence review, drug toxicity evaluation, and other high-precision analytical tasks. The major engineering shift was the introduction of Google’s new Interactions API — a feature aimed at letting developers embed Deep Research-grade capabilities into their own products and control the agent layer more directly.

Google also disclosed plans to integrate Deep Research into a series of its core services, including Search, Finance, the Gemini app, and NotebookLM. It is edging toward an internet where people no longer hunt for information themselves; they dispatch an agent trained to do it for them. Google leaned heavily on the claim that Gemini 3 Pro is its “most factual” model so far, tuned to lower hallucination rates in long-chain reasoning tasks where a single incorrect step can ruin the entire output.

To back its performance claims, Google introduced a new benchmark called DeepSearchQA, intended to test how well an agent handles complex, multi-step information missions. It was released as open source. The company also tested the system on Humanity’s Last Exam, a notoriously difficult general-knowledge benchmark filled with obscure tasks, and on BrowserComp, which evaluates an AI agent’s ability to carry out browser-based operations.

In addition to the widely-praised Gemini 3, Deep Research topped the leaderboard on Google’s own benchmark and on Humanity’s. But OpenAI’s ChatGPT 5 Pro landed close behind in both tests and overtook Google on BrowserComp. Even those results barely had time to settle. Hours later, OpenAI announced GPT-5.2, which the company said outperforms rivals — including Google — across a wide suite of standard benchmarks.

This back-to-back rollout marked another escalation in a rivalry that has grown unusually public. The industry now watches releases from both companies like heavyweight rounds. Each upgrade arrives with the sense of a reply to the other.

Google introduced a package that pushes its agentic vision further, giving developers a programmable research companion and moving the company closer to agent-based search. OpenAI countered with a model built to strengthen its leadership in reasoning, speed, and multimodal performance.

The speed of these releases also signals how aggressively both companies are trying to define the next phase of AI development. The focus is shifting from raw model power to advanced agents that can plan, browse, run tools, complete tasks over long stretches, and handle increasingly complex workloads with minimal supervision. Whoever builds the most reliable and scalable agent layer stands to redefine how people use software, how information moves across the web, and how organizations make decisions.

For now, Google seems to be pushing deeper into agentic research systems integrated across its own ecosystem, while OpenAI continues to refine its core model lineup and deploy it across consumer and enterprise channels. Each announcement seems to trigger the next. Each benchmark update sets off another. And as these systems continue to evolve in public view, the industry’s center of gravity keeps shifting toward a true duel for AI supremacy.