This is a substantial PR to add support for the chat completions API, which in turn makes it possible to use non-OpenAI model providers (just like in the TypeScript CLI): * It moves a number of structs from `client.rs` to `client_common.rs` so they can be shared. * It introduces support for the chat completions API in `chat_completions.rs`. * It updates `ModelProviderInfo` so that `env_key` is `Option<String>` instead of `String` (for e.g., ollama) and adds a `wire_api` field * It updates `client.rs` to choose between `stream_responses()` and `stream_chat_completions()` based on the `wire_api` for the `ModelProviderInfo` * It updates the `exec` and TUI CLIs to no longer fail if the `OPENAI_API_KEY` environment variable is not set * It updates the TUI so that `EventMsg::Error` is displayed more prominently when it occurs, particularly now that it is important to alert users to the `CodexErr::EnvVar` variant. * `CodexErr::EnvVar` was updated to include an optional `instructions` field so we can preserve the behavior where we direct users to https://platform.openai.com if `OPENAI_API_KEY` is not set. * Cleaned up the "welcome message" in the TUI to ensure the model provider is displayed. * Updated the docs in `codex-rs/README.md`. To exercise the chat completions API from OpenAI models, I added the following to my `config.toml`: ```toml model = "gpt-4o" model_provider = "openai-chat-completions" [model_providers.openai-chat-completions] name = "OpenAI using Chat Completions" base_url = "https://api.openai.com/v1" env_key = "OPENAI_API_KEY" wire_api = "chat" ``` Though to test a non-OpenAI provider, I installed ollama with mistral locally on my Mac because ChatGPT said that would be a good match for my hardware: ```shell brew install ollama ollama serve ollama pull mistral ``` Then I added the following to my `~/.codex/config.toml`: ```toml model = "mistral" model_provider = "ollama" ``` Note this code could certainly use more test coverage, but I want to get this in so folks can start playing with it. For reference, I believe https://github.com/openai/codex/pull/247 was roughly the comparable PR on the TypeScript side.
152 lines
4.5 KiB
Rust
152 lines
4.5 KiB
Rust
use std::time::Duration;
|
||
|
||
use codex_core::Codex;
|
||
use codex_core::ModelProviderInfo;
|
||
use codex_core::config::Config;
|
||
use codex_core::protocol::InputItem;
|
||
use codex_core::protocol::Op;
|
||
use serde_json::Value;
|
||
use tokio::time::timeout;
|
||
use wiremock::Match;
|
||
use wiremock::Mock;
|
||
use wiremock::MockServer;
|
||
use wiremock::Request;
|
||
use wiremock::ResponseTemplate;
|
||
use wiremock::matchers::method;
|
||
use wiremock::matchers::path;
|
||
|
||
/// Matcher asserting that JSON body has NO `previous_response_id` field.
|
||
struct NoPrevId;
|
||
|
||
impl Match for NoPrevId {
|
||
fn matches(&self, req: &Request) -> bool {
|
||
serde_json::from_slice::<Value>(&req.body)
|
||
.map(|v| v.get("previous_response_id").is_none())
|
||
.unwrap_or(false)
|
||
}
|
||
}
|
||
|
||
/// Matcher asserting that JSON body HAS a `previous_response_id` field.
|
||
struct HasPrevId;
|
||
|
||
impl Match for HasPrevId {
|
||
fn matches(&self, req: &Request) -> bool {
|
||
serde_json::from_slice::<Value>(&req.body)
|
||
.map(|v| v.get("previous_response_id").is_some())
|
||
.unwrap_or(false)
|
||
}
|
||
}
|
||
|
||
/// Build minimal SSE stream with completed marker.
|
||
fn sse_completed(id: &str) -> String {
|
||
format!(
|
||
"event: response.completed\n\
|
||
data: {{\"type\":\"response.completed\",\"response\":{{\"id\":\"{}\",\"output\":[]}}}}\n\n\n",
|
||
id
|
||
)
|
||
}
|
||
|
||
#[tokio::test(flavor = "multi_thread", worker_threads = 2)]
|
||
async fn keeps_previous_response_id_between_tasks() {
|
||
#![allow(clippy::unwrap_used)]
|
||
|
||
// Mock server
|
||
let server = MockServer::start().await;
|
||
|
||
// First request – must NOT include `previous_response_id`.
|
||
let first = ResponseTemplate::new(200)
|
||
.insert_header("content-type", "text/event-stream")
|
||
.set_body_raw(sse_completed("resp1"), "text/event-stream");
|
||
|
||
Mock::given(method("POST"))
|
||
.and(path("/v1/responses"))
|
||
.and(NoPrevId)
|
||
.respond_with(first)
|
||
.expect(1)
|
||
.mount(&server)
|
||
.await;
|
||
|
||
// Second request – MUST include `previous_response_id`.
|
||
let second = ResponseTemplate::new(200)
|
||
.insert_header("content-type", "text/event-stream")
|
||
.set_body_raw(sse_completed("resp2"), "text/event-stream");
|
||
|
||
Mock::given(method("POST"))
|
||
.and(path("/v1/responses"))
|
||
.and(HasPrevId)
|
||
.respond_with(second)
|
||
.expect(1)
|
||
.mount(&server)
|
||
.await;
|
||
|
||
// Environment
|
||
// Update environment – `set_var` is `unsafe` starting with the 2024
|
||
// edition so we group the calls into a single `unsafe { … }` block.
|
||
unsafe {
|
||
std::env::set_var("OPENAI_REQUEST_MAX_RETRIES", "0");
|
||
std::env::set_var("OPENAI_STREAM_MAX_RETRIES", "0");
|
||
}
|
||
let model_provider = ModelProviderInfo {
|
||
name: "openai".into(),
|
||
base_url: format!("{}/v1", server.uri()),
|
||
// Environment variable that should exist in the test environment.
|
||
// ModelClient will return an error if the environment variable for the
|
||
// provider is not set.
|
||
env_key: Some("PATH".into()),
|
||
env_key_instructions: None,
|
||
wire_api: codex_core::WireApi::Responses,
|
||
};
|
||
|
||
// Init session
|
||
let mut config = Config::load_default_config_for_test();
|
||
config.model_provider = model_provider;
|
||
let ctrl_c = std::sync::Arc::new(tokio::sync::Notify::new());
|
||
let (codex, _init_id) = Codex::spawn(config, ctrl_c.clone()).await.unwrap();
|
||
|
||
// Task 1 – triggers first request (no previous_response_id)
|
||
codex
|
||
.submit(Op::UserInput {
|
||
items: vec![InputItem::Text {
|
||
text: "hello".into(),
|
||
}],
|
||
})
|
||
.await
|
||
.unwrap();
|
||
|
||
// Wait for TaskComplete
|
||
loop {
|
||
let ev = timeout(Duration::from_secs(1), codex.next_event())
|
||
.await
|
||
.unwrap()
|
||
.unwrap();
|
||
if matches!(ev.msg, codex_core::protocol::EventMsg::TaskComplete) {
|
||
break;
|
||
}
|
||
}
|
||
|
||
// Task 2 – should include `previous_response_id` (triggers second request)
|
||
codex
|
||
.submit(Op::UserInput {
|
||
items: vec![InputItem::Text {
|
||
text: "again".into(),
|
||
}],
|
||
})
|
||
.await
|
||
.unwrap();
|
||
|
||
// Wait for TaskComplete or error
|
||
loop {
|
||
let ev = timeout(Duration::from_secs(1), codex.next_event())
|
||
.await
|
||
.unwrap()
|
||
.unwrap();
|
||
match ev.msg {
|
||
codex_core::protocol::EventMsg::TaskComplete => break,
|
||
codex_core::protocol::EventMsg::Error { message } => {
|
||
panic!("unexpected error: {message}")
|
||
}
|
||
_ => (),
|
||
}
|
||
}
|
||
}
|