From 7e9ecfbc6a9b0c96147afed4026b5c91df9472a9 Mon Sep 17 00:00:00 2001 From: pakrym-oai Date: Thu, 7 Aug 2025 09:07:51 -0700 Subject: [PATCH] Rename the model (#1942) --- codex-rs/core/src/client.rs | 9 --------- codex-rs/core/src/model_family.rs | 4 ++-- codex-rs/core/src/openai_model_info.rs | 2 +- 3 files changed, 3 insertions(+), 12 deletions(-) diff --git a/codex-rs/core/src/client.rs b/codex-rs/core/src/client.rs index 0fa143fd..ed05fb5d 100644 --- a/codex-rs/core/src/client.rs +++ b/codex-rs/core/src/client.rs @@ -127,15 +127,6 @@ impl ModelClient { let auth_mode = auth.as_ref().map(|a| a.mode); - if self.config.model_family.family == "2025-08-06-model" - && auth_mode != Some(AuthMode::ChatGPT) - { - return Err(CodexErr::UnexpectedStatus( - StatusCode::BAD_REQUEST, - "2025-08-06-model is only supported with ChatGPT auth, run `codex login status` to check your auth status and `codex login` to login with ChatGPT".to_string(), - )); - } - let store = prompt.store && auth_mode != Some(AuthMode::ChatGPT); let full_instructions = prompt.get_full_instructions(&self.config.model_family); diff --git a/codex-rs/core/src/model_family.rs b/codex-rs/core/src/model_family.rs index cadbceca..1245a030 100644 --- a/codex-rs/core/src/model_family.rs +++ b/codex-rs/core/src/model_family.rs @@ -89,9 +89,9 @@ pub fn find_family_for_model(slug: &str) -> Option { simple_model_family!(slug, "gpt-oss") } else if slug.starts_with("gpt-3.5") { simple_model_family!(slug, "gpt-3.5") - } else if slug.starts_with("2025-08-06-model") { + } else if slug.starts_with("gpt-5") { model_family!( - slug, "2025-08-06-model", + slug, "gpt-5", supports_reasoning_summaries: true, ) } else { diff --git a/codex-rs/core/src/openai_model_info.rs b/codex-rs/core/src/openai_model_info.rs index 0ce94267..a072d409 100644 --- a/codex-rs/core/src/openai_model_info.rs +++ b/codex-rs/core/src/openai_model_info.rs @@ -77,7 +77,7 @@ pub(crate) fn get_model_info(model_family: &ModelFamily) -> Option { max_output_tokens: 4_096, }), - "2025-08-06-model" => Some(ModelInfo { + "gpt-5" => Some(ModelInfo { context_window: 200_000, max_output_tokens: 100_000, }),