DD
MM
YYYY

PAGES

DD
MM
YYYY

spot_img

PAGES

Home Blog Page 7

Anthropic in Early Stages of Exploring Possibilities of Designing its Own AI Chips

0

Anthropic is in the very early stages of exploring the possibility of designing its own AI chips. The company hasn’t committed to the idea, formed a dedicated team, or settled on any specific architecture. It could still decide to continue solely buying chips from existing suppliers.

Sources described the discussions as preliminary, driven by the chronic shortage of high-end AI accelerators needed to train and run ever-larger models. Anthropic currently relies on a diversified mix of hardware: NVIDIA GPUs including recent use of Blackwell for at least one major model like Mythos.

Google’s TPUs via a major expansion on Google Cloud, potentially up to ~1 million TPUs in partnership with Broadcom. Amazon’s Trainium and Inferentia chips through its primary cloud and training partnership on AWS, including the massive Project Rainier cluster. This multi-vendor strategy provides resilience, but surging demand for Claude with Anthropic’s annualized revenue reportedly tripling to a $30B+ run rate is straining supply and driving up costs.

Designing in-house silicon could give Anthropic more control over performance, power efficiency, and long-term economics—reducing what some call the Nvidia tax on margins and availability. This isn’t isolated. Other frontier labs and hyperscalers are pursuing similar paths: Meta and OpenAI already have custom chip projects underway.

Google (TPUs), Amazon (Trainium/Inferentia), and Microsoft (with Maia) have long invested in custom AI silicon. Partnerships like Anthropic’s with Broadcom for custom TPUs show they’re already leaning into semi-custom designs before going fully in-house. Designing a competitive AI chip from scratch is extremely expensive (hundreds of millions of dollars) and technically demanding. Success isn’t guaranteed—NVIDIA still dominates due to its CUDA software ecosystem, scale, and iterative hardware improvements.

Many attempts at custom AI accelerators have underperformed or been abandoned. If Anthropic moves forward, it could: Lower long-term compute costs. Optimize hardware specifically for Claude’s architecture and safety-focused training methods. Further diversify away from any single supplier. However, execution risks are high, and it would take years to reach production scale.

For now, the report signals strategic caution amid explosive AI growth rather than an imminent break from NVIDIA or its cloud partners. This fits the ongoing vertical integration push in AI: labs realizing that software model performance is increasingly bottlenecked by hardware access and cost. The compute race is shifting from who has the most GPUs toward who can build or control the best silicon stack.

We’ll likely see more such explorations as inference and training demands continue to outpace supply. Custom chips could reduce long-term dependence on expensive Nvidia GPUs and ease shortages. Optimization for Claude’s architecture might improve trainin and inference efficiency, power usage, and performance-per-watt, lowering the massive compute bills that frontier labs face.

Greater control over hardware tailored to safety-focused or specific model needs, potentially accelerating development cycles. However, success is far from guaranteed—designing a competitive AI accelerator can cost ~$500 million upfront, plus years of engineering, manufacturing likely via TSMC or similar, and software ecosystem building.

High execution risk. Failure could waste capital. Near-term, Anthropic continues diversifying via deals like expanded Google TPUs with Broadcom, scaling to multi-gigawatt capacity and CoreWeave for Nvidia-based cloud

Tether’s QVAC SDK Is a Pivotal Step Making AI More Private, Resilient and Accessible

0

Tether, the company behind the USDT stablecoin has just released QVAC SDK, an open-source, cross-platform toolkit for building and running local, decentralized AI directly on devices.

QVAC from Tether’s dedicated AI team is positioned as a universal building block for the Stable Intelligence Era — a future with billions of devices, autonomous machines, and AI agents running intelligence privately and without relying on centralized cloud providers. AI models run offline on the device itself for privacy, speed, and resilience — no API keys, no cloud dependency, and no Big Tech oversight.

Cross-platform from a single codebase: Supports iOS, Android, Windows, macOS, and Linux with no code changes. Built on a llama.cpp fork called QVAC Fabric, enabling text generation, speech processing, visual recognition, translation, and more. Uses the Holepunch protocol stack for peer-to-peer model distribution and delegated inference.
Decentralized training and fine-tuning, plus specialized toolkits for robotics and brain-computer interfaces. It also includes Fabric LLM, a LoRA fine-tuning framework for edge devices.

Tether’s CEO Paolo Ardoino has called centralized AI a dead end. This move expands Tether beyond stablecoins into open infrastructure for on-device AI, aligning with growing demands for privacy, decentralization echoing Web3 and DePIN narratives, and resilience against cloud outages or censorship. It builds on Tether’s earlier QVAC efforts, like releasing large open synthetic educational datasets for AI training.

By offering a single codebase that runs seamlessly across iOS, Android, Windows, macOS, and Linux, QVAC lowers barriers for building offline-capable AI apps. It builds on a llama.cpp fork with unified support for text generation, speech, vision, translation and more. Integration with the Holepunch protocol enables peer-to-peer model distribution, delegated inference, and future decentralized training and fine-tuning via swarms.

This reduces reliance on centralized servers and improves resilience. It extends llama.cpp’s strengths; broad model compatibility via GGUF with cross-platform abstractions and planned LoRA fine-tuning on edge devices. It competes with platform-specific solutions like Apple’s MLX (strong on Apple Silicon performance and fine-tuning) but aims for true hardware-agnostic, privacy-first deployment.

Early feedback notes it simplifies integration but will need optimization to match cloud-scale performance on smaller models. Planned expansions into robotics and brain-computer interfaces could accelerate specialized on-device AI in hardware-heavy fields. Faster prototyping of private AI features without API costs or vendor lock-in. Open-source nature invites community contributions and audits.

AI runs locally without sending sensitive inputs to remote servers. This directly addresses concerns over surveillance, data breaches, and censorship by Big Tech. Users get instant, always-available AI for everyday tasks like writing, finance planning, or translation—even without internet. Lower latency, no subscription fees for inference, and reduced exposure to cloud outages or policy changes.

Challenge to cloud AI dominance: Accelerates the shift from hyperscale cloud models to edge computing. It could pressure centralized providers by highlighting privacy and cost drawbacks. Fosters transparency and innovation, contrasting with closed models from major labs. Analysts see it as part of a broader trend toward responsible, decentralized AI adoption.

Tether’s diversification: Signals the stablecoin giant’s evolution into infrastructure beyond finance. It builds on prior QVAC efforts and could create synergies with DeFi, agents, or autonomous systems. Aligns strongly with decentralized physical infrastructure networks (DePIN), Web3, and autonomous AI agents. Local AI could enable trustless, on-device agents that interact with blockchains/dApps without central intermediaries—potentially powering smarter wallets, DeFi tools, or P2P economies.

Tether is already exploring related areas. Success here could enhance USDT’s utility in AI-powered crypto applications and attract developer talent to the broader crypto space. This requires sustained R&D investment. Some commentary notes potential capital strain if USDT market cap faces continued pressure, as resources are diverted from core stablecoin operations. Returns are uncertain and hinge on ecosystem growth.

On-device models may lag cloud performance on complex tasks; security and optimization responsibility shifts to developers and users; achieving critical mass for P2P swarms will take time. QVAC SDK is seen as a pivotal step in making AI more private, resilient, and accessible—potentially influencing how future intelligent systems are built amid rising concerns over centralization.

It’s not a complete replacement for cloud AI yet but adds a powerful local/decentralized alternative. If adoption grows, it could reshape developer tools, user expectations, and even regulatory conversations around data control. Short-term impact is mostly in crypto and tech circles; longer-term effects will depend on community contributions and practical use cases.

OpenAI’s Chief Scientist Says AI is close to reaching Human-level Intelligence

0

Fresh comments from chief scientist Jakub Pachocki suggest OpenAI believes it is moving materially closer to one of its most ambitious internal milestones: building systems capable of functioning at the level of a human research intern, a development that could reshape not only AI research itself but the future economics of science and technical work.

OpenAI is moving closer to one of the most consequential milestones it has publicly outlined in the race toward advanced artificial intelligence: the creation of systems that can operate at the level of a human research intern.

Speaking on the Unsupervised Learning podcast, chief scientist Jakub Pachocki said recent progress across coding, mathematical reasoning, and physics-related problem solving suggests the company’s internal roadmap remains on track.

“I definitely see this as a signal that something here is on track,” Pachocki said, pointing to recent technical breakthroughs as evidence that models are beginning to handle increasingly complex, multi-step work with less direct human intervention.

The significance of that remark lies not in the headline ambition alone, but in what OpenAI now sees as the core metric of progress. Rather than focusing purely on benchmark scores or isolated task performance, Pachocki framed autonomy in terms of time horizon.

“The way I would distinguish a research intern from a full automated researcher is the span of time that we would have it work mostly autonomously,” he said.

That is an important shift in how frontier labs are increasingly defining intelligence. The question is no longer whether a model can solve a single problem correctly. It is whether it can sustain coherent work over hours, days, or potentially weeks without constant human correction.

This concept, often described in the industry as “long-horizon autonomy,” is fast becoming one of the most important frontiers in AI development.

At an internal livestream last October, Pachocki laid out a two-stage roadmap: an “AI research intern” by September 2026, followed by a fully autonomous AI researcher by March 2028. Sam Altman later acknowledged the uncertainty around the target, writing that OpenAI “may totally fail” at the goal, but said transparency was necessary given the scale of its implications.

Pachocki pointed to the “explosive growth of coding tools,” particularly agents such as Codex, which he said are already handling much of the company’s internal programming work.

“We’ve seen this explosive growth of coding tools,” he said. “For most people, the act of programming has changed quite a bit.”

This is one of the most revealing parts of the interview. OpenAI is effectively describing a feedback loop in which AI tools are increasingly being used to improve the very systems that produce them. If coding agents are already automating substantial portions of internal software work, the logical next step is the automation of research workflows themselves: experiment design, evaluation pipelines, model comparisons, literature synthesis, and iterative testing.

Pachocki made this progression explicit.

“For more specific technical ideas, like I have this particular idea how to improve the models, how to run this evaluation differently, I think we have the pieces that we mostly just need to put together,” he said.

That phrase, “put the pieces together,” may sound modest, but it points to a major industry inflection point. Many of the component capabilities already exist in fragmented form: coding agents, reasoning systems, verification tools, web-enabled research agents, and increasingly capable math solvers.

The challenge now is orchestration, which has birthed an open question. Can these systems chain together tasks reliably enough to mimic the workflow of a junior researcher?

Pachocki was careful not to overstate where the technology currently stands.

“I don’t expect we’ll have systems where you just tell them, ‘go improve your model capability, go solve alignment,’ and they will do it, not this year,” he said.

That caveat is important because it sharply distinguishes between intern-level assistance and true scientific autonomy. A research intern, in this framing, is not an independent scientist. It is a system capable of executing bounded, technically sophisticated tasks over longer durations with minimal supervision.

Junior-level technical work across AI labs, universities, biotech firms, and enterprise R&D units could increasingly be augmented or partially automated. This could compress experimentation cycles from weeks to days, allowing frontier labs to iterate faster than smaller competitors. It may also widen the competitive moat around firms with the compute, data, and engineering infrastructure to deploy such systems at scale.

The “AI research intern” is believed to be an indication of a move from AI as a tool for users to AI as an active participant in the research process itself. It is expected to mark a transition from copilots to semi-autonomous scientific agents if realized.

However, the most important insight from Pachocki’s remarks is that OpenAI is increasingly measuring progress by sustained autonomy rather than isolated intelligence. That is regarded as a more difficult benchmark, but also a more meaningful one.

US Spot Bitcoin and Ethereum ETFs Added Solid Net Inflows

0

U.S. spot Bitcoin (BTC) and Ethereum (ETH) ETFs have been seeing solid net inflows recently. BTC ETFs recorded ~$358M net inflows led heavily by BlackRock’s IBIT at +$269M, with Fidelity’s FBTC adding +$53M and smaller contributions from others.

ETH ETFs added ~$85M net; BlackRock’s ETHA was the standout at +$91M, offset by some outflows from Fidelity and others. April 6, 2026, BTC ETFs saw a strong ~$471M inflow; one of the largest daily figures of 2026 so far, the 6th-biggest. ETH added around $120M. This was another notably positive day. Earlier in the week/month, flows have been mixed but trending more positive after weaker periods in Q1 2026.

April has shown stabilization or recovery in demand compared to prior outflows. These inflows reflect institutional and retail interest via regulated vehicles, often led by major issuers like BlackRock; IBIT for BTC and ETHA for ETH dominate daily volumes and flows.

BTC ETFs: Cumulative net inflows remain strongly positive long-term; tens of billions since launch, though 2026 started with some volatility and net outflows in parts of Q1. Recent days signal renewed buying as BTC hovers near or above $68K–$70K levels amid macro easing.

ETH ETFs: More variable due to ETH’s higher beta nature. They’ve seen streaks of both inflows and outflows including Grayscale-related redemptions early on. Total AUM for ETH products is in the $10B–$20B range depending on the period, representing a smaller slice of ETH’s market cap than BTC ETFs do for Bitcoin. Lower perceived risk, ETF maturity, potential rate environment signals, and crypto’s correlation with broader risk assets.

However, Grayscale products GBTC and ETHE have historically seen outflows as investors rotate to lower-fee alternatives. Consistent positive ETF flows are generally bullish for spot prices as they represent sticky institutional capital entering the market without direct custody hassles. That said, crypto remains volatile — inflows don’t guarantee short-term pumps, especially with macro factors like Fed policy, employment data, etc. in play.

Bitcoin has repeatedly tested or briefly surpassed the $73,000 level in 2026, most notably in early-to-mid March during a sharp recovery rally, and again more recently in April e.g., intraday tops near or above $73k on April 9–10 amid volatile trading. As of April 10, 2026, BTC trades around $71,000–$72,800, showing resilience but facing resistance in the $73k–$75k zone. This price action occurs against a backdrop of geopolitical tensions, mixed U.S. economic data and ongoing ETF flows.

Reaching or approaching $73k often coincided with renewed buying in U.S. spot Bitcoin ETFs. Examples include: Multi-day streaks with hundreds of millions in net inflows. BlackRock’s IBIT frequently led, absorbing large portions of daily flows. Combined BTC + ETH ETF inflows in the $400M+ range on some sessions.

These flows act as a structural bid, helping stabilize or propel price even when spot selling or macro headwinds appear. However, flows remain volatile—strong inflow days alternate with modest outflows. Brief breaches of $73k signaled a shift from consolidation and sparked short squeezes, rising open interest, and higher trading volumes often $60B–$75B+ daily.

Altcoins and crypto-related stocks tended to surge on the coattails. Investors viewed BTC as a geopolitical hedge amid Middle East tensions, decoupling somewhat from traditional risk assets at times. It also drew attention as a potential inflation or dollar-weakness play. Crossing $73k improved trader confidence and attracted media coverage, drawing in retail and institutional attention.

Breaking and holding above it cleanly could open paths toward $75k–$80k, while failures led to pullbacks toward $68k–$70k support.
Increased volatility: Moves to $73k+ often involved 7–9% daily swings, with leverage amplifying liquidations both long and short. Higher prices correlated with ETF accumulation offsetting some large-holder distribution. Resilience to Iran-related news was highlighted—BTC climbed despite or because of risk-off moves elsewhere.

Weak U.S. indicators paradoxically supported BTC as a scarce asset: Even at $73k peaks, BTC remained down significantly from its 2025 all-time high ~$126k, underscoring a recovery phase rather than euphoric new highs. Stronger ETF AUM growth, potential for more corporate and  institutional adoption, and bullish narratives around regulatory developments.

Profit-taking emerged quickly on some days, and failure to hold $73k led to consolidation. ETH often moved in tandem but with higher beta. Overall crypto market cap reacted positively but remained sensitive to macro shifts. BTC hovers just below or testing the level again, supported by recent ETF inflows but capped by spot selling and geopolitical uncertainty.

Touching $73k has historically in 2026 reinforced institutional interest via ETFs, lifted sentiment, and highlighted BTC’s role as a resilient asset—but it has not yet triggered a sustained breakout due to overhead supply and external pressures. Crypto remains highly volatile; price levels like this are psychological milestones more than fundamental turning points on their own.

Fed’s Rescue Playbook Is Gone, Ruchir Sharma Warns as Oil Shock Reignites Inflation Risks

0

Investors hoping the Federal Reserve will once again step in to cushion the U.S. economy at the first sign of trouble may need to reset expectations.

Ruchir Sharma, chairman of Rockefeller Capital and one of Wall Street’s most closely watched macro investors, says the era of the so-called “Fed put” is effectively over, arguing that persistent inflation and a fresh oil-price shock have left the central bank with little room to cut interest rates even if economic growth weakens.

His warning comes at a delicate moment for financial markets, where investors are trying to assess whether the recent surge in energy costs linked to the Middle East conflict will tip the U.S. economy into a slower-growth, higher-inflation environment.

“I think the Fed is out,” Sharma said in an interview with CNN this week, referring to the prospect of rate cuts.

“When you got any shock to the economy, central banks would rush to the aid of the economy at the slightest hint of trouble. I don’t see that happening this time.”

That is a striking shift from the policy regime that dominated the post-financial-crisis era and was reinforced during the pandemic, when the Fed moved aggressively to slash rates and flood markets with liquidity at the earliest sign of stress.

This time, Sharma argues, inflation is the binding constraint.

Fresh data from the U.S. Bureau of Labor Statistics show that headline inflation accelerated to 3.3% year on year in March, up sharply from 2.4% in February, the highest reading since mid-2024. The increase was driven overwhelmingly by energy, with gasoline prices posting one of the largest monthly jumps in recent years.

More importantly for the Fed, inflation has now remained above its 2% target for roughly 60 consecutive months, reinforcing concerns inside the central bank that price pressures are proving more persistent than expected.

This means the Fed is now confronting a classic policy dilemma. Growth risks are rising, but inflation is still too elevated to justify an easy pivot. That sharply limits policymakers’ ability to use rate cuts as an economic shock absorber.

In market terms, the “Fed put” refers to the long-standing assumption that the central bank would step in to support asset prices and growth whenever financial conditions tightened materially.

Sharma’s argument is that this safety net no longer exists in its traditional form, and the bond market appears to agree.

The 10-year U.S. Treasury yield rose to about 4.30%, up roughly 34 basis points from levels seen before the Iran conflict intensified. Rising long-term yields signal that investors are demanding greater compensation for inflation risk and fiscal uncertainty.

That rise in yields is particularly notable because it suggests the market is becoming less tolerant of Washington’s expanding fiscal deficit.

Sharma pointed directly to that issue.

“The bond markets this time aren’t tolerating that,” he said.

“The doomsayers have been around a long time worrying about the debt and the deficits, and the question many people ask is, ‘So what? The bond market is not really reacting to any of this.’ That’s changing now.”

This introduces a second major constraint on the economy: the potential disappearance of what some investors call the “Trump put.” That term refers to the expectation that the administration might respond to economic weakness with fiscal stimulus, tax relief, or other growth-supportive measures.

But if Treasury yields continue to rise, any new fiscal package risks worsening bond-market anxiety by increasing borrowing needs at a time when investors are already focused on deficit sustainability. In effect, both traditional support mechanisms, monetary easing and fiscal stimulus, are facing resistance.

Markets are beginning to reflect this reality. According to CME FedWatch, traders now assign only about a 29% probability of a rate cut later this year, a dramatic repricing from the much higher odds seen a month ago.

That said, sentiment remains fluid as some market participants slightly raised cut expectations after the latest inflation report showed that core CPI remained relatively contained at 0.2% month on month, suggesting the recent spike is still largely an energy story rather than broad-based overheating.

However, some analysts believe that if oil prices stabilize after the ceasefire, the Fed may eventually regain room to ease. But if energy costs remain elevated and feed into transportation, goods, and services inflation over the next six to eight weeks, the central bank may be forced to remain hawkish for longer.

That is the heart of Sharma’s thesis. He indicated that this is not 2020. Back then, inflation was low, and policymakers had enormous flexibility. Today, the Fed must balance growth risks against a renewed inflation shock, which means that volatility is likely to remain elevated.

Equities can no longer assume that weaker data automatically translates into easier policy. Instead, bad economic news may simply be bad news, marking a fundamental change in the investment landscape.

Sharma’s warning, stripped to its essence, is that the U.S. economy is entering a period where policy support is far less automatic than investors have become accustomed to over the past decade. If growth therefore stumbles under the weight of higher oil prices, markets may find that the usual rescue mechanisms are no longer readily available.