From 29d154cb139d63ef4896dd5fa7ebc5b303fb7fea Mon Sep 17 00:00:00 2001 From: Michael Bolin Date: Tue, 27 May 2025 09:12:55 -0700 Subject: [PATCH] fix: use o4-mini as the default model (#1135) Rollback of https://github.com/openai/codex/pull/972. --- codex-rs/README.md | 2 +- codex-rs/core/src/flags.rs | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/codex-rs/README.md b/codex-rs/README.md index 705d3130..a0e3f584 100644 --- a/codex-rs/README.md +++ b/codex-rs/README.md @@ -32,7 +32,7 @@ The `config.toml` file supports the following options: The model that Codex should use. ```toml -model = "o3" # overrides the default of "codex-mini-latest" +model = "o3" # overrides the default of "o4-mini" ``` ### model_provider diff --git a/codex-rs/core/src/flags.rs b/codex-rs/core/src/flags.rs index c21ef670..e8cc973c 100644 --- a/codex-rs/core/src/flags.rs +++ b/codex-rs/core/src/flags.rs @@ -3,7 +3,7 @@ use std::time::Duration; use env_flags::env_flags; env_flags! { - pub OPENAI_DEFAULT_MODEL: &str = "codex-mini-latest"; + pub OPENAI_DEFAULT_MODEL: &str = "o4-mini"; pub OPENAI_API_BASE: &str = "https://api.openai.com/v1"; /// Fallback when the provider-specific key is not set.