Connect with us
DAPA Banner

Crypto World

Stanford flags rising opacity at the frontier

Published

on

Stanford flags rising opacity at the frontier

The AI models at the frontier of performance are also the least transparent about how they are built and tested, according to Stanford HAI’s 2026 AI Index released Monday, which found that companies are sharing progressively less about training data and benchmark performance even as their models become more powerful and more widely deployed.

Summary

  • Stanford’s report documents that AI companies are sharing less information about how their models are trained, and that independent testing sometimes contradicts what companies report; “a lot of companies are not releasing how their models do in certain benchmarks, particularly the responsible-AI benchmarks,” the report states, citing growing opacity at the exact moment when accountability matters most.
  • The benchmarks designed to measure AI progress are themselves failing: some are poorly constructed, with a popular math benchmark carrying a 42 percent error rate, while others can be gamed by models trained on the benchmark test data itself, meaning strong scores do not reliably indicate stronger or safer models in real-world deployment.
  • US trust in the government to regulate AI sits at just 31 percent, the lowest of any country surveyed in the index; globally, the EU is trusted more than either the US or China to regulate AI effectively, a finding that reflects both the EU AI Act’s full enforcement in January 2026 and the absence of a comparable federal framework in the US.

SiliconAngle reported that the 2026 index documents a world where AI adoption is accelerating at historic speed while “public trust in AI oversight and transparency hits new lows.” The two trends are directly related: as AI tools reach more than half the global population and generate $172 billion in annual consumer value in the US alone, the lack of visibility into how the most powerful models are built and evaluated creates a governance gap that neither regulators nor the public can easily close without the data to work from.

The benchmark problem is not abstract. If a model scores well because it was trained on test data, that score provides no meaningful signal about how the model will perform on novel tasks in deployment. For complex use cases like AI agents and robots, the report notes that benchmarks barely exist yet, meaning the most consequential AI applications are being deployed with almost no standardized external validation.

Advertisement

The opacity operates at multiple levels. At the training level, companies have reduced disclosure about the datasets, filtering methods, and human feedback processes used to build their models. At the evaluation level, they are choosing which benchmarks to publish results on, a selection that naturally favors the tests on which their models perform well. At the deployment level, independent researchers testing the same models sometimes find results that contradict what companies have publicly stated. The Stanford report does not name specific companies but documents the pattern as industry-wide.

Why This Matters More Now Than It Did Two Years Ago

Two years ago, frontier AI models were research tools used primarily by developers and researchers. Today they are integrated into customer service systems, hiring workflows, medical information delivery, financial advice, and legal research. The gap between benchmark performance and real-world performance is no longer an academic concern; it determines whether the systems that millions of people interact with daily are actually doing what their developers claim. The report’s finding that responsible-AI benchmarks are the category companies most often decline to publish results on is precisely the category that matters most for those real-world applications.

What Regulatory and Industry Standards Currently Exist

As crypto.news has reported, the AI infrastructure buildout is advancing faster than the governance structures designed to evaluate it, a tension that is visible in both investment markets and public policy debates. As crypto.news has noted, the competitive pressure among frontier AI labs to release capable models quickly creates structural incentives against transparency, because publishing benchmark weaknesses or training methodology details can be exploited by competitors. Stanford’s report frames that dynamic as the central accountability problem of the current AI era, with 47 countries now having introduced AI-specific legislation but only 23 having enacted laws with active enforcement mechanisms.

Advertisement

Source link

Continue Reading
Click to comment

You must be logged in to post a comment Login

Leave a Reply

Crypto World

SEC Approves Elimination of Pattern Day Trader Rule and $25,000 Minimum: FINRA

Published

on

SEC Approves Elimination of Pattern Day Trader Rule and $25,000 Minimum: FINRA

The SEC granted accelerated approval to FINRA’s rule change eliminating the Pattern Day Trader designation and its $25,000 minimum equity requirement for day traders.

The U.S. Securities and Exchange Commission on Tuesday approved FINRA’s proposed rule change eliminating the Pattern Day Trader designation, the $25,000 minimum equity requirement, and all related day-trading buying power provisions under FINRA Rule 4210. The accelerated approval removes longstanding restrictions that have governed retail day trading for decades.

The SEC simultaneously approved new intraday margin standards requiring broker-dealers to monitor and address real-time risk exposure in customer margin accounts. The regulatory shift represents a substantial change to day-trading accessibility and compliance frameworks for retail investors in U.S. equity markets.

Sources: WatcherGuru | WatcherGuru

Advertisement

This article was generated automatically by The Defiant’s AI news system from publicly available sources.

Source link

Continue Reading

Crypto World

Global recession inevitable if Strait of Hormuz stays shut

Published

on

Global recession inevitable if Strait of Hormuz stays shut

Ken Griffin, chief executive officer of Citadel Advisors LLC, at the Semafor World Economy Summit during the International Monetary Fund (IMF) and World Bank Spring meetings in Washington, DC, US, on Tuesday, April 14, 2026.

Aaron Schwartz | Bloomberg | Getty Images

Citadel CEO Ken Griffin said Tuesday that the global economy is headed toward a recession if the Strait of Hormuz stays shut for much longer.

Advertisement

“Let’s assume [the strait is] shut down for the next six to 12 months — the world’s going to end up in a recession,” Griffin said on stage at the Semafor World Economy conference in Washington, D.C. “There’s no way to avoid that.”

As a result, the world is going to see a massive shift toward alternative fuel sources, including wind, solar and nuclear, he added. To be sure, the hedge fund leader thinks the consequences of the war would have been worse if the U.S. delayed any strikes until Iran’s military capabilities had grown.

Stocks have managed to rebound back to where they were before the U.S. first attacked Iran in February, but the optimistic sentiment among investors is contingent on the duration of the war in the Middle East. Many expect risks of an escalation in tensions between the two countries are not at all priced into the market.

Global economies especially in Asia remain vulnerable to spikes in oil prices, which remain elevated at around $100 a barrel. That’s off their highs during the conflict, but remain far above where they were before the war, at just below $70 a barrel.

Advertisement
Choose CNBC as your preferred source on Google and never miss a moment from the most trusted name in business news.

Source link

Continue Reading

Crypto World

Paxos Labs Raises $12M to Launch Crypto Yield and Lending Platform

Published

on

Paxos Labs Raises $12M to Launch Crypto Yield and Lending Platform

Paxos Labs has raised $12 million in a strategic funding round led by Blockchain Capital to expand its Amplify platform, a suite of tools that lets companies offer crypto yield, lending and stablecoin issuance through a single integration.

The Amplify suite includes three modules — Earn, Borrow and Mint — allowing platforms to generate yield on digital assets, enable crypto-backed loans and issue branded stablecoins with a single integration designed to unlock additional features over time.

According to Tuesday’s announcement, the platform provides a single SDK with configurable controls, while Paxos Labs manages liquidity, counterparty vetting and backend operations, and shares a portion of generated revenue with integrating partners.

The company said partners including Aleo, Hyperbeat and Toku are already using the platform, with Hyperbeat reporting more than $510,000 in assets under management since launching on April 9. The raise also included participation from Robot Ventures, Maelstrom and Uniswap.

Advertisement