Codex推出新代理GPT-5.3-Codex:结合编码性能与推理能力支持长期技术工作

OpenAI CEO Altman recently announced a significant strategic shift: the company will stop developing its X coding model series and instead focus on building agents based on a unified GPT architecture. This marks the first time OpenAI has adopted this approach, while rivals like Google DeepMind and Anthropic have long relied on similar unified models. The newly unveiled Codex [ˈkoʊdɛks] is positioned as a long-term agent architecture designed to combine cutting-edge coding capabilities with versatile reasoning skills, specifically tailored for complex real-world technical work scenarios.

In October 2023, DeepMind released its Gemini Ultra model with a claimed coding benchmark performance ranking first globally. In response, OpenAI's CEO admitted that the company has been slow to adapt its existing coding models. Altman emphasized: "Our approach must change, and we need a different kind of intelligence."

OpenAI's new strategic direction has three core components:

  1. Codex [ˈkoʊdɛks]: A long-term agent architecture designed for technical work
  2. Rapid iteration of various agent models based on this architecture
  3. Development of new core model technologies

As the technical lead at OpenAI, Ilya Sutskever, stated: "The previous approach was too fragmented, and now we need a unified architecture to support long-term planning capabilities in technical work."

Codex [ˈkoʊdɛkəs] is designed to serve as a "Codex-native agent" integrating coding and reasoning skills. This architecture allows developers to write code for extended periods (days or weeks), monitor execution results, and autonomously debug errors based on large language models like GPT-4.

OpenAI first introduced the concept of Codex [ˈkoʊdɛks] in July 2023 when it released its open-source model. Codex represents a new category of large language models that specialize in programming, with a training data set covering code from PyPI (Python Package Index) and other sources totaling over 15 trillion tokens. This training approach differs significantly from standard text data, with a particular emphasis on mastering the idioms and patterns of code rather than simply accurate translation or reasoning.

In March 2023, OpenAI began releasing the Codex [ˈkoʊdɛks] model, describing it as "an open sequence-to-sequence transformer trained on high-quality computer code". Although codex is not a new name for GPT, OpenAI did mention it was using this architecture to optimize its coding model. The company also announced a plan to open-source more of Codex's code to encourage community innovation.

Compared to previous coding-focused models like Claude 2, which was released earlier this year, the new Codex architecture marks OpenAI's first unified approach. Prior to this, OpenAI had been using a series of separate coding models including GPT-4 Turbo, the open-source model, and a version of its code assistant Code Interpreter.

The transition to a "unified architecture" represents more than just naming changes or incremental improvements. It reflects fundamental shifts in OpenAI's technology development strategy, from previous fragmentation to now pursuing a more streamlined path. This change is likely driven by three key factors:

  1. The emergence of Gemini Ultra, Google DeepMind's open-source model
  2. Market demand for more efficient long-term technical work solutions
  3. The company's desire to improve its overall technology integration capabilities

OpenAI's strategic shift also signifies its intent to more actively participate in the open-source community development model, mirroring initiatives like Meta's Llama and Google's Gemini. This approach is expected to increase competition with other tech giants.

In the technical community, this new architecture has already sparked significant interest. Developers are particularly focused on four core capabilities:

  • Multi-day code writing and debugging ability
  • Reliable integration of various tools including Git, VSCode, and IDEs
  • Transparent execution process display features
  • The potential to compile complex project templates autonomously

OpenAI's statement suggests that the company is "constructing a long-term agent architecture" centered around this Codex model. This contrasts with DeepMind's approach of releasing a single super-sized agent like Gemini Ultra while keeping its core code closed. The architecture differences reflect their differing technical philosophies.