命令行 – AI 编码的控制范围
Command Lines

原始链接: https://www.wreflection.com/p/command-lines-ai-coding

## AI 编码助手崛起:快速发展的市场 (2025) AI 编码助手正在改变软件开发,其影响类似于 20 世纪 50 年代编译器对软件开发的影响,使工程师能够专注于高级逻辑,而不是低级代码。这种转变正在推动爆炸式增长——市场规模在一年多时间内扩大了 5 倍,像 Cursor 这样的公司实现了创纪录的 10 亿美元年收入。 市场目前分为两种主要用户类型:“免提”用户(非工程师原型设计概念)和“动手”用户(专业开发者将 AI 集成到他们的工作流程中)。主要参与者包括 Cursor、AWS Kiro、Google Antigravity 和 GitHub Copilot,初创公司和科技巨头正在激烈竞争。 **模型质量至关重要。** 虽然用户界面和工作流程集成很重要,但生成可靠、可用于生产的代码的能力正在推动采用。公司正在大力投资于利用现有基础模型 *和* 开发自己的模型,例如 Cursor 的新 Composer-2。 这种演变甚至影响了传统的开发者资源,如 Stack Overflow,因为 AI 越来越多地提供即时解决方案。最终目标是完全自主的软件生成,而胜者将是那些提供卓越模型质量、满足特定需求并建立持久用户忠诚度的人。

## AI 工具与软件开发格局的变化 - Hacker News 总结 Hacker News 的讨论围绕着拥抱 AI 驱动编码工具的开发者与持怀疑态度的开发者之间日益扩大的分歧。用户报告称,公司内部存在差距,一些人使用 CodeRabbit.ai 等工具进行代码审查,体验到显著的速度提升,而另一些人则质疑其真正的好处,认为“代理编码”在质量或速度方面并没有比传统方法有明显改善。 人们对速度声明的有效性表示担忧,因为每个项目都是独一无二的。关于 AI 生成的代码是否符合人类开发者的原始意图,也存在争论。讨论涉及 Stack Overflow 等资源潜在的衰落对模型训练的影响,以及 AI 开发工具市场拥挤,暗示任何单一参与者不太可能占据主导地位。 最后,围绕为传统编码(“Trad Coding”)创造术语展开了有趣的辩论,一些人警告说,由于与“Trad wife”运动相关的潜在负面含义,应谨慎使用该术语。总体情绪凸显了软件开发实践中快速变化和不确定性时期。
相关文章

原文

In the early 1950s, Grace Hopper coined the term “compiler” and built one of the first versions with her A-0 system. The compilers that followed abstracted away machine code, letting programmers focus on higher-level logic instead of lower-level hardware details. Today, AI coding assistants are enabling a similar change, letting software engineers focus on higher-order work by generating code from natural language prompts. Everyone from big tech to well-funded startups is competing to capture this shift. Yesterday Google announced Antigravity, their new AI coding assistant, and the day before, AWS announced the general availability of their AI coding tool, Kiro. Last week, Cursor, the standout startup in this space, raised $2.3B in their series-D round at a valuation of $29.3B.

Two lines in Cursor’s press release stood out to me. The first:

We’ve also crossed $1B in annualized revenue, counting millions of developers.

This disclosure means Anysphere Inc. (Cursor’s parent company) is the fastest company in history to reach $1B in annual recurring revenue (ARR). Yes, faster than OpenAI, and faster than Anthropic.

Engineers are trying every new AI coding tool. As a result, the AI-coding tool market is growing exponentially (+5x in just over a year). But it’s still early. As I wrote in Why Some AI Wrappers Build Billion-dollar Businesses, companies spend several hundred billion dollars a year on software engineering, and AI has the potential to unlock productivity gains across that entire spend.

Software developers represent roughly 30% of the workforce at the world’s five largest market cap companies, all of which are technology firms as of October 2025. Development tools that boost productivity by even modest percentages unlock billions in value.

In my view, this nascent market is splitting based on three types of users.

On one end is Handcrafted Coding. These are engineers who actively decline to use LLMs, either because of skepticism about quality or insistence on full control of every code. They argue that accepting AI suggestions creates technical debt you cannot see until it breaks in production. This segment continues to decline as the quality of AI coding models improves.

The opposite end is Vibe Coding. These are typically non-engineers, who use AI to build concepts and prototypes. They prompt the model hoping for an end-to-end solution, accept the output with minimal review, and trust that it works. The user describes what they want and lets the model figure out the implementation details of how to build it.

In the middle sits Architect + AI Coding. The engineer uses the AI/LLM as a pair programmer exploring system designs, analyzing data models, and reviewing API details. When the work is something entirely new or something that needs careful handling, the human programmer still codes those pieces by hand. But for boilerplate code, package installations, generic User Interface (UI) components, and any kind of code that is typically found on the internet, they assign it to the model. The engineer stays in command of what is important to them and delegates what is not.

Based on the user types, I think, the AI coding market splits into two.

  1. Hands-off: Non-engineers (product managers, designers, marketers, other internal employees) use these tools to vibe code early product concepts. They look to AI as the lead engineer to spin-up concepts/prototypes of apps, websites, and tools by simply prompting the AI to make something for them. Lovable, Vercel, Bolt, Figma Make, and Replit fit here. Code from these users, as of now, are not typically pushed to prod.

  2. Hands-on: Professional software engineers use these tools in their existing workflow to ship production code. They use AI as an assistant to write boilerplate code, refactor existing services, wire new features or UI screens, and triage bugs in codebases. Cursor, Claude Code, OpenAI Codex, Github Copilot, Cline, AWS Kiro play here. These products live where the work is done, and integrate into the engineer’s workflow. This is, at least as of now, the bigger market segment.

To see an evaluation of all the major AI coding tools currently in the market, checkout this breakdown by Peter Yang, who runs the newsletter Behind The Craft.

That brings me to the second thing in Cursor’s press release that stood out to me:

Our in-house models now generate more code than almost any other LLMs in the world.

While I am not convinced about that claim, what I am convinced about is that Cursor is still growing despite its previous reliance on foundation models. From Why Some AI Wrappers Build Billion-dollar Businesses again:

But Cursor and other such tools depend almost entirely on accessing Anthropic, OpenAI and Gemini models, until open-source open-weight and in-house models match or exceed frontier models in quality. Developer forums are filled with complaints about rate limits from paying subscribers. In my own projects, I exhausted my Claude credits in Cursor mid-project and despite preferring Cursor’s user interface and design, I migrated to Claude Code (and pay ten times more to avoid rate limits). The interface may be better, but model access proved decisive.

Cursor’s new in-house model Composer-2, which just launched last month, is a good example of how this model versus application competition is evolving. Cursor claims (without any external benchmarks, I must say) that Composer-2 is almost as good as frontier models but 4x faster. It’s still early to say how true that is. Open-source models have not yet come close to the top spots in SWE-bench verified or in private evals.

Source: Introducing Claude Sonnet 4.5, Anthropic, 2025.

To me, model quality is the most decisive factor in these AI coding wars. And in my view, that’s why Claude Code has already overtaken Cursor, and OpenAI’s Codex is close behind, despite both having launched a year or so later.

Even though the newcomers Cursor, Claude Code, and OpenAI Codex are the talk of the (developer) town, incumbents such as Microsoft with Github Copilot, AWS with Kiro, and Google with Antigravity, can utilize their existing customer relationships, bundle their offerings with their existing suites, and/or provide their option as the default in their tech stack to compete. As an example, Cursor charges $20–$40 monthly per user for productive usage, while Google Antigravity launched free with generous limits for individual users. Github Copilot still leads this market, proving once again that enterprise bundling and distribution has structural advantages. This is the classic Microsoft Teams vs. Slack Dynamic.

One way for startups to compete is by winning individual users who may use a coding tool with or without formal approval, and then be the tool’s advocate inside the organization. That organic interest and adoption eventually forces IT and security teams to officially review the tool and then eventually sanction its usage.

Yet, even as these newer tools capture developer mindshare, the underlying developer tools market is changing. Both the IDEs developers choose and the resources they we consult have changed dramatically. StackOverflow, once the default for programmers stuck on a programming issue, has seen its traffic and number of questions decline dramatically since ChatGPT’s launch, suggesting that AI is already replacing some traditional developer resources.

Source: Developer Tools 2.0, Sequoia, 2023

Just as compilers freed programmers from writing assembly code, AI tools are freeing software engineers from the grunt work of writing boilerplate and routine code, and letting them focus on higher-order thinking. Eventually, one day, AI may get so good that it will generate applications on demand and create entire software ecosystems autonomously. Both hands-off and hands-on AI coding tools, as well as incumbents and newcomers, see themselves as the path to that fully autonomous software generation, even if they are taking different approaches. The ones who get there will be those who deliver the best model quality that ships code reliably, go deep enough to ship features that foundation models can’t care enough to replicate, and become sticky enough that users will not leave even when they can.

If you enjoyed this post, please consider sharing it on Twitter/X or LinkedIn, and tag me when you do.

Share

联系我们 contact @ memedata.com