休斯顿,我们有麻烦了:Anthropic 乘着人工智能的浪潮 – BIML
Anthropic Rides an Artificial Wave

原始链接: https://berryvilleiml.com/2025/11/14/houston-we-have-a-problem-anthropic-rides-an-artificial-wave/

这篇文章批判了围绕Anthropic声称发现“首次人工智能策划的网络间谍活动”的炒作。作者认为,报道中的攻击依赖于现成的开源工具和蛮力方法——这些技术已经驱动网络犯罪多年——而不是独特的“自主型人工智能”能力。 核心问题是媒体甚至安全专业人员在网络安全*和*人工智能/机器学习方面的综合专业知识不足,导致不加批判的报道。作者强调需要具体的证据(“给我看日志!”),并警告不要将大型语言模型(LLM)人格化,它们擅长*模拟*角色,但缺乏真正的意图。 最终,作者驳斥了这一说法被夸大,指出自动化攻击已经存在了几十年。虽然承认机器学习安全的重要性,但他们提倡一种脚踏实地、现实的方法,超越耸人听闻的叙述。

## AI驱动的黑客攻击与安全问题 - 摘要 Hacker News上的讨论强调了人们对人工智能快速发展,特别是像Anthropic的Claude这样的大型语言模型(LLM)对网络安全的影响日益增长的担忧。一家构建AI驱动的渗透测试代理的初创公司报告称,每次Claude迭代(从Sonnet 3.5到4.5)都实现了显著的能力飞跃,甚至在一个小时内在一个实时网络上实现了权限提升。 核心担忧在于,人工智能正在降低复杂网络攻击的门槛。虽然没有引入*新的*攻击方法,但它极大地提高了速度、规模和自动化程度,使技术水平较低的攻击者能够执行复杂的行动。专家们担心攻击者和防御者之间即将爆发一场军备竞赛,AI代理可能会绕过杀毒软件并自动化漏洞利用。 对话还涉及攻击者使用领先(且昂贵)的美国AI模型,而存在可比较的中国选择的讽刺,以及对提高安全专业知识的需求。一些人建议专注于使用AI代理进行“CTF式”演练,以主动寻找漏洞,而不是仅仅依赖传统的报告。关于开源AI开发是否受到封闭、商业驱动的模型阻碍,以及AI可能加剧现有安全挑战的潜力,存在争论。
相关文章

原文

I’ll tip my hat to the new Constitution
Take a bow for the new revolution
Smile and grin at the change all around
Pick up my guitar and play
Just like yesterday
Then I’ll get on my knees and pray
We don’t get fooled again

Out there in the smoking rubble of the fourth estate, it is hard enough to cover cyber cyber. Imagine, then, piling on the AI bullshit. Can anybody cut through the haze? Apparently for the WSJ and the NY Times, the answer is no.

Yeah, it’s Anthropic again. This time writing a blog-post level document titled “Disrupting the first reported AI-orchestrated cyber espionage campaign” and getting the major tech press all wound around the axle about it.

The root of the problem here is that expertise in cyber cyber is rare AND expertise in AI/ML is rare…but expertise in both fields? Not only is it rare, but like hydrogen-7, which has a half-life of about 10^-24 seconds, it disappears pretty fast as both fields progress. Even superstar tech reporters can’t keep everything straight.

Lets start with the end. What question should the press have asked Anthropic about their latest security story? How about, “which parts of these attacks could ONLY be accomplished with agentic AI?” From our little perch at BIML, it looks like the answer is a resounding none.

Now that we know the ending, lets look at both sides of the beginning. Security first. Unfortunately, brute force, cloud-scale, turnkey software exploit is what has been driving the ransomware cybercrime wave for at least a decade now. All of the offensive security tool technology used by the attackers Anthropic describes is available as open source frameworks, leading experts like Kevin Beaumont to label the whole thing, “vibe usage of open source attack frameworks.” Would existing controls work against this? Apparently not for “a handful” of the thirty companies Anthropic claims were successfully attacked. LOL.

By now those of us old enough to know better than to call ourselves security experts have learned how to approach claims like the ones Anthropic is making skeptically. “Show me the logs,” we yell as we shake our canes in the air. Seriously. Where is the actual evidence? Who has seen it. Do we credulously repeat whatever security vendors tell us as it it is the gods’ honest truth? No we do not. Who was successfully attacked? Did the reporters chase them down? Who was on the list of 30?

AI second. It is all too easy to exaggerate claims in today’s superheated AI universe. One of the most trivial (and intellectually lazy) ways to do this is to use anthropomorphic language when we are describing what LLMs do. LLMs don’t “think” or “believe” or “have intentionality” like humans do. (FWIW, Anthropic is very much guilty of this and they are not getting any better.) LLMs do do a great job of role playing though. So dressing one up as a black hat nation state hacker and sending it lumbering off into the klieg lights is easy.

So who did it? How do we prove that beyond a reasonable doubt? Hilariously, the real attacks here appear to be asking an LLM to pretend to be a white hat red team member dressed in a Where’s Waldo shirt and weilding a SSRF attack. Wake me up when it’s over.

Ultimately, is this really the “first documented case of a cyberattack largely executed without human intervention at scale”…no, that was the script kiddies in the ’90s.

Lets be extremely clear here. Machine Learning Security is absolutely critical. We have lots of work to do. So lets ground ourselves in reality and get to it.

联系我们 contact @ memedata.com