Data reveals that “Claude got less intelligent” isn’t an urban legend; an AI model’s instability can become a corporate risk.

ChainNewsAbmedia

After AI became a standard tool for enterprises, a phenomenon once dismissed as a “matter of feel” is quickly coming to the surface: LLMs (large language models) are getting “dumber.” Netizen Wisely Chen points out that so-called “LLM loss of intelligence” is not an urban legend—it has already been continuously tracked through data, and it is now having a real impact on enterprise work processes.

He cites his own experience as an example. On April 15, Anthropic’s Claude services suffered an across-the-board downgrade, including claude.ai, the API, and Claude Code, all showing “Degraded Performance.” This is not simply slower response times or occasional errors; instead, the response quality clearly collapsed, and even situations arose where it couldn’t be used normally, causing all three of his development tasks that day to be delayed in full.

For individual developers, this kind of situation may only mean reduced efficiency, but for enterprise IT teams, the impact is multiplied. When a team has multiple engineers simultaneously relying on AI tools for coding, document writing, and workflow automation, a single model downgrade means overall productivity dips collectively at the same time—turning into a measurable loss of time and costs.

Does it feel like AI is getting dumber? Data confirms it has “already degraded”

Wisely Chen notes that claims like “GPT is getting dumber” and “Claude isn’t as good as before” have circulated in the community for a long time, but have long lacked objective data to back them up. Only until recently, with the emergence of platforms that continuously monitor model quality, has this phenomenon been quantified for the first time.

Among them, StupidMeter runs 24-hour automated tests on mainstream models including OpenAI, Anthropic, Google, and more, tracking metrics such as correctness, reasoning ability, and stability. Unlike traditional one-off benchmarks, these systems are more similar to how enterprises monitor an API or service availability—observing fluctuations in how the model performs in real usage environments.

The results are quite straightforward: most mainstream models are currently in a warning or degraded state, with only a few models maintaining normal performance. This means that model quality is unstable—a widespread industry phenomenon rather than a problem limited to a single product.

LLMs quietly lose intelligence, impacting enterprise AI workflow stability

For enterprises, this kind of change means AI has shifted from a “tool for improving efficiency” to a “variable that affects stability.” If a company’s daily workflows—everything from writing code, to doing code reviews, to producing documents and analysis reports—are already highly dependent on LLMs, then when the model’s reasoning ability drops or response quality deteriorates on a given day, these issues won’t occur locally like traditional software bugs. Instead, they will seep into every part of the workflow that uses AI at the same time.

More importantly, these fluctuations are often hard to predict and difficult to detect in real time. Most enterprises do not have mechanisms to continuously monitor model quality. They usually only realize the problem comes from the model after output results turn abnormal or team efficiency declines. In such a scenario, “loss of intelligence” is no longer just a subjective user experience—it becomes a systemic risk that directly affects the rhythm of business operations.

When AI becomes water and electricity, stability becomes the new key metric

Wisely Chen compares the role of LLMs to “water and electricity for modern enterprises.” When AI has been deeply integrated into day-to-day operations and becomes an indispensable foundational capability, the importance of stability naturally rises as well.

In the past, when enterprises evaluated AI tools, they mainly focused on model capability, price, and features. But as the “loss of intelligence” phenomenon comes into view, another more critical metric is emerging: stability. When model quality may change without notification, enterprises are no longer just “using AI”—they must start taking on a new form of infrastructure risk. Even more bleak is that if you only look at cutting-edge large language models, as long as the compute power problem hasn’t been solved, it may continue to happen.

This article first appeared in Lianxin ABMedia: Data reveals “Claude loss of intelligence” is not an urban legend—AI model instability becomes an enterprise risk.

Disclaimer: The information on this page may come from third parties and does not represent the views or opinions of Gate. The content displayed on this page is for reference only and does not constitute any financial, investment, or legal advice. Gate does not guarantee the accuracy or completeness of the information and shall not be liable for any losses arising from the use of this information. Virtual asset investments carry high risks and are subject to significant price volatility. You may lose all of your invested principal. Please fully understand the relevant risks and make prudent decisions based on your own financial situation and risk tolerance. For details, please refer to Disclaimer.

Related Articles

Meta Stock Rises 1.73% as Company Plans 8,000-Job Layoff Starting May 20

Meta Platforms plans to cut about 8,000 jobs, or 10% of its workforce, starting May 20, despite rising stock prices. The company, with over $200 billion in revenue, is focusing on AI investments amid significant restructuring, aligning with industry trends of layoffs.

GateNews4h ago

Google’s annual report says Gemini achieves millisecond interception, blocking 99% of scam ads

The article discusses how Google strengthens ad safety through its generative AI system, Gemini. The report shows that the speed at which it blocks noncompliant ads has been reduced to milliseconds, with a blocking rate of 99%. Last year, Google removed 8.3 billion ad listings and suspended 24.9 million accounts, indicating a significant rise in the number of scam ads. Experts point out that this is a contest between AI and AI, and that in the future there will still be challenges in dealing with both legal and illegal activities brought about by AI.

ChainNewsAbmedia6h ago

Ethereum Co-founder Lubin: AI Will Be Critical Turning Point for Crypto, But Tech Giant Monopoly Poses Systemic Risk

Ethereum co-founder Joseph Lubin emphasized the transformative potential of AI for the cryptocurrency sector while cautioning against the risks of centralization among tech giants. He envisions AI-driven autonomous transactions on blockchain and highlights the convergence of traditional finance with DeFi.

GateNews8h ago

Elon Musk Pushes 'Universal High Income' Checks as Ultimate Solution for AI Unemployment

Elon Musk advocates for a Universal High Income to combat AI-induced unemployment, envisioning a future with ample goods and zero inflation. In contrast, experts like Sam Altman raise concerns about job loss and propose protective measures for workers.

Coinpedia8h ago

DeepSeek Reportedly Launches First External Fundraising Round, Targets $10B+ Valuation and $300M+

DeepSeek, a Chinese AI startup, is negotiating its first external funding round, aiming for at least $300 million at a $10 billion valuation. Despite previous rejections of investment offers, its fundraising discussions are now reportedly underway.

GateNews8h ago

ChatGPT ads move into Australia and New Zealand: Free and Go users first, paid plans stay ad-free

OpenAI expanded ChatGPT advertising on April 17, 2023 to Australia, New Zealand, and Canada for Free and Go users, with no ads for paid users. This marks the second pathway toward AI commercialization and takes into account business and regulatory risks, where the presence of ads can promote paid conversions.

ChainNewsAbmedia11h ago
Comment
0/400
No comments