This release represents a comprehensive transformation of the codebase from Codex to LLMX, enhanced with LiteLLM integration to support 100+ LLM providers through a unified API. ## Major Changes ### Phase 1: Repository & Infrastructure Setup - Established new repository structure and branching strategy - Created comprehensive project documentation (CLAUDE.md, LITELLM-SETUP.md) - Set up development environment and tooling configuration ### Phase 2: Rust Workspace Transformation - Renamed all Rust crates from `codex-*` to `llmx-*` (30+ crates) - Updated package names, binary names, and workspace members - Renamed core modules: codex.rs → llmx.rs, codex_delegate.rs → llmx_delegate.rs - Updated all internal references, imports, and type names - Renamed directories: codex-rs/ → llmx-rs/, codex-backend-openapi-models/ → llmx-backend-openapi-models/ - Fixed all Rust compilation errors after mass rename ### Phase 3: LiteLLM Integration - Integrated LiteLLM for multi-provider LLM support (Anthropic, OpenAI, Azure, Google AI, AWS Bedrock, etc.) - Implemented OpenAI-compatible Chat Completions API support - Added model family detection and provider-specific handling - Updated authentication to support LiteLLM API keys - Renamed environment variables: OPENAI_BASE_URL → LLMX_BASE_URL - Added LLMX_API_KEY for unified authentication - Enhanced error handling for Chat Completions API responses - Implemented fallback mechanisms between Responses API and Chat Completions API ### Phase 4: TypeScript/Node.js Components - Renamed npm package: @codex/codex-cli → @valknar/llmx - Updated TypeScript SDK to use new LLMX APIs and endpoints - Fixed all TypeScript compilation and linting errors - Updated SDK tests to support both API backends - Enhanced mock server to handle multiple API formats - Updated build scripts for cross-platform packaging ### Phase 5: Configuration & Documentation - Updated all configuration files to use LLMX naming - Rewrote README and documentation for LLMX branding - Updated config paths: ~/.codex/ → ~/.llmx/ - Added comprehensive LiteLLM setup guide - Updated all user-facing strings and help text - Created release plan and migration documentation ### Phase 6: Testing & Validation - Fixed all Rust tests for new naming scheme - Updated snapshot tests in TUI (36 frame files) - Fixed authentication storage tests - Updated Chat Completions payload and SSE tests - Fixed SDK tests for new API endpoints - Ensured compatibility with Claude Sonnet 4.5 model - Fixed test environment variables (LLMX_API_KEY, LLMX_BASE_URL) ### Phase 7: Build & Release Pipeline - Updated GitHub Actions workflows for LLMX binary names - Fixed rust-release.yml to reference llmx-rs/ instead of codex-rs/ - Updated CI/CD pipelines for new package names - Made Apple code signing optional in release workflow - Enhanced npm packaging resilience for partial platform builds - Added Windows sandbox support to workspace - Updated dotslash configuration for new binary names ### Phase 8: Final Polish - Renamed all assets (.github images, labels, templates) - Updated VSCode and DevContainer configurations - Fixed all clippy warnings and formatting issues - Applied cargo fmt and prettier formatting across codebase - Updated issue templates and pull request templates - Fixed all remaining UI text references ## Technical Details **Breaking Changes:** - Binary name changed from `codex` to `llmx` - Config directory changed from `~/.codex/` to `~/.llmx/` - Environment variables renamed (CODEX_* → LLMX_*) - npm package renamed to `@valknar/llmx` **New Features:** - Support for 100+ LLM providers via LiteLLM - Unified authentication with LLMX_API_KEY - Enhanced model provider detection and handling - Improved error handling and fallback mechanisms **Files Changed:** - 578 files modified across Rust, TypeScript, and documentation - 30+ Rust crates renamed and updated - Complete rebrand of UI, CLI, and documentation - All tests updated and passing **Dependencies:** - Updated Cargo.lock with new package names - Updated npm dependencies in llmx-cli - Enhanced OpenAPI models for LLMX backend This release establishes LLMX as a standalone project with comprehensive LiteLLM integration, maintaining full backward compatibility with existing functionality while opening support for a wide ecosystem of LLM providers. 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude <noreply@anthropic.com> Co-Authored-By: Sebastian Krüger <support@pivoine.art>
318 lines
11 KiB
Rust
318 lines
11 KiB
Rust
use crate::types::CodeTaskDetailsResponse;
|
|
use crate::types::PaginatedListTaskListItem;
|
|
use crate::types::RateLimitStatusPayload;
|
|
use crate::types::RateLimitWindowSnapshot;
|
|
use crate::types::TurnAttemptsSiblingTurnsResponse;
|
|
use anyhow::Result;
|
|
use llmx_core::auth::LlmxAuth;
|
|
use llmx_core::default_client::get_llmx_user_agent;
|
|
use llmx_protocol::protocol::RateLimitSnapshot;
|
|
use llmx_protocol::protocol::RateLimitWindow;
|
|
use reqwest::header::AUTHORIZATION;
|
|
use reqwest::header::CONTENT_TYPE;
|
|
use reqwest::header::HeaderMap;
|
|
use reqwest::header::HeaderName;
|
|
use reqwest::header::HeaderValue;
|
|
use reqwest::header::USER_AGENT;
|
|
use serde::de::DeserializeOwned;
|
|
|
|
#[derive(Clone, Copy, Debug, PartialEq, Eq)]
|
|
pub enum PathStyle {
|
|
/// /api/llmx/…
|
|
LlmxApi,
|
|
/// /wham/…
|
|
ChatGptApi,
|
|
}
|
|
|
|
impl PathStyle {
|
|
pub fn from_base_url(base_url: &str) -> Self {
|
|
if base_url.contains("/backend-api") {
|
|
PathStyle::ChatGptApi
|
|
} else {
|
|
PathStyle::LlmxApi
|
|
}
|
|
}
|
|
}
|
|
|
|
#[derive(Clone, Debug)]
|
|
pub struct Client {
|
|
base_url: String,
|
|
http: reqwest::Client,
|
|
bearer_token: Option<String>,
|
|
user_agent: Option<HeaderValue>,
|
|
chatgpt_account_id: Option<String>,
|
|
path_style: PathStyle,
|
|
}
|
|
|
|
impl Client {
|
|
pub fn new(base_url: impl Into<String>) -> Result<Self> {
|
|
let mut base_url = base_url.into();
|
|
// Normalize common ChatGPT hostnames to include /backend-api so we hit the WHAM paths.
|
|
// Also trim trailing slashes for consistent URL building.
|
|
while base_url.ends_with('/') {
|
|
base_url.pop();
|
|
}
|
|
if (base_url.starts_with("https://chatgpt.com")
|
|
|| base_url.starts_with("https://chat.openai.com"))
|
|
&& !base_url.contains("/backend-api")
|
|
{
|
|
base_url = format!("{base_url}/backend-api");
|
|
}
|
|
let http = reqwest::Client::builder().build()?;
|
|
let path_style = PathStyle::from_base_url(&base_url);
|
|
Ok(Self {
|
|
base_url,
|
|
http,
|
|
bearer_token: None,
|
|
user_agent: None,
|
|
chatgpt_account_id: None,
|
|
path_style,
|
|
})
|
|
}
|
|
|
|
pub async fn from_auth(base_url: impl Into<String>, auth: &LlmxAuth) -> Result<Self> {
|
|
let token = auth.get_token().await.map_err(anyhow::Error::from)?;
|
|
let mut client = Self::new(base_url)?
|
|
.with_user_agent(get_llmx_user_agent())
|
|
.with_bearer_token(token);
|
|
if let Some(account_id) = auth.get_account_id() {
|
|
client = client.with_chatgpt_account_id(account_id);
|
|
}
|
|
Ok(client)
|
|
}
|
|
|
|
pub fn with_bearer_token(mut self, token: impl Into<String>) -> Self {
|
|
self.bearer_token = Some(token.into());
|
|
self
|
|
}
|
|
|
|
pub fn with_user_agent(mut self, ua: impl Into<String>) -> Self {
|
|
if let Ok(hv) = HeaderValue::from_str(&ua.into()) {
|
|
self.user_agent = Some(hv);
|
|
}
|
|
self
|
|
}
|
|
|
|
pub fn with_chatgpt_account_id(mut self, account_id: impl Into<String>) -> Self {
|
|
self.chatgpt_account_id = Some(account_id.into());
|
|
self
|
|
}
|
|
|
|
pub fn with_path_style(mut self, style: PathStyle) -> Self {
|
|
self.path_style = style;
|
|
self
|
|
}
|
|
|
|
fn headers(&self) -> HeaderMap {
|
|
let mut h = HeaderMap::new();
|
|
if let Some(ua) = &self.user_agent {
|
|
h.insert(USER_AGENT, ua.clone());
|
|
} else {
|
|
h.insert(USER_AGENT, HeaderValue::from_static("llmx-cli"));
|
|
}
|
|
if let Some(token) = &self.bearer_token {
|
|
let value = format!("Bearer {token}");
|
|
if let Ok(hv) = HeaderValue::from_str(&value) {
|
|
h.insert(AUTHORIZATION, hv);
|
|
}
|
|
}
|
|
if let Some(acc) = &self.chatgpt_account_id
|
|
&& let Ok(name) = HeaderName::from_bytes(b"ChatGPT-Account-Id")
|
|
&& let Ok(hv) = HeaderValue::from_str(acc)
|
|
{
|
|
h.insert(name, hv);
|
|
}
|
|
h
|
|
}
|
|
|
|
async fn exec_request(
|
|
&self,
|
|
req: reqwest::RequestBuilder,
|
|
method: &str,
|
|
url: &str,
|
|
) -> Result<(String, String)> {
|
|
let res = req.send().await?;
|
|
let status = res.status();
|
|
let ct = res
|
|
.headers()
|
|
.get(CONTENT_TYPE)
|
|
.and_then(|v| v.to_str().ok())
|
|
.unwrap_or("")
|
|
.to_string();
|
|
let body = res.text().await.unwrap_or_default();
|
|
if !status.is_success() {
|
|
anyhow::bail!("{method} {url} failed: {status}; content-type={ct}; body={body}");
|
|
}
|
|
Ok((body, ct))
|
|
}
|
|
|
|
fn decode_json<T: DeserializeOwned>(&self, url: &str, ct: &str, body: &str) -> Result<T> {
|
|
match serde_json::from_str::<T>(body) {
|
|
Ok(v) => Ok(v),
|
|
Err(e) => {
|
|
anyhow::bail!("Decode error for {url}: {e}; content-type={ct}; body={body}");
|
|
}
|
|
}
|
|
}
|
|
|
|
pub async fn get_rate_limits(&self) -> Result<RateLimitSnapshot> {
|
|
let url = match self.path_style {
|
|
PathStyle::LlmxApi => format!("{}/api/llmx/usage", self.base_url),
|
|
PathStyle::ChatGptApi => format!("{}/wham/usage", self.base_url),
|
|
};
|
|
let req = self.http.get(&url).headers(self.headers());
|
|
let (body, ct) = self.exec_request(req, "GET", &url).await?;
|
|
let payload: RateLimitStatusPayload = self.decode_json(&url, &ct, &body)?;
|
|
Ok(Self::rate_limit_snapshot_from_payload(payload))
|
|
}
|
|
|
|
pub async fn list_tasks(
|
|
&self,
|
|
limit: Option<i32>,
|
|
task_filter: Option<&str>,
|
|
environment_id: Option<&str>,
|
|
) -> Result<PaginatedListTaskListItem> {
|
|
let url = match self.path_style {
|
|
PathStyle::LlmxApi => format!("{}/api/llmx/tasks/list", self.base_url),
|
|
PathStyle::ChatGptApi => format!("{}/wham/tasks/list", self.base_url),
|
|
};
|
|
let req = self.http.get(&url).headers(self.headers());
|
|
let req = if let Some(lim) = limit {
|
|
req.query(&[("limit", lim)])
|
|
} else {
|
|
req
|
|
};
|
|
let req = if let Some(tf) = task_filter {
|
|
req.query(&[("task_filter", tf)])
|
|
} else {
|
|
req
|
|
};
|
|
let req = if let Some(id) = environment_id {
|
|
req.query(&[("environment_id", id)])
|
|
} else {
|
|
req
|
|
};
|
|
let (body, ct) = self.exec_request(req, "GET", &url).await?;
|
|
self.decode_json::<PaginatedListTaskListItem>(&url, &ct, &body)
|
|
}
|
|
|
|
pub async fn get_task_details(&self, task_id: &str) -> Result<CodeTaskDetailsResponse> {
|
|
let (parsed, _body, _ct) = self.get_task_details_with_body(task_id).await?;
|
|
Ok(parsed)
|
|
}
|
|
|
|
pub async fn get_task_details_with_body(
|
|
&self,
|
|
task_id: &str,
|
|
) -> Result<(CodeTaskDetailsResponse, String, String)> {
|
|
let url = match self.path_style {
|
|
PathStyle::LlmxApi => format!("{}/api/llmx/tasks/{}", self.base_url, task_id),
|
|
PathStyle::ChatGptApi => format!("{}/wham/tasks/{}", self.base_url, task_id),
|
|
};
|
|
let req = self.http.get(&url).headers(self.headers());
|
|
let (body, ct) = self.exec_request(req, "GET", &url).await?;
|
|
let parsed: CodeTaskDetailsResponse = self.decode_json(&url, &ct, &body)?;
|
|
Ok((parsed, body, ct))
|
|
}
|
|
|
|
pub async fn list_sibling_turns(
|
|
&self,
|
|
task_id: &str,
|
|
turn_id: &str,
|
|
) -> Result<TurnAttemptsSiblingTurnsResponse> {
|
|
let url = match self.path_style {
|
|
PathStyle::LlmxApi => format!(
|
|
"{}/api/llmx/tasks/{}/turns/{}/sibling_turns",
|
|
self.base_url, task_id, turn_id
|
|
),
|
|
PathStyle::ChatGptApi => format!(
|
|
"{}/wham/tasks/{}/turns/{}/sibling_turns",
|
|
self.base_url, task_id, turn_id
|
|
),
|
|
};
|
|
let req = self.http.get(&url).headers(self.headers());
|
|
let (body, ct) = self.exec_request(req, "GET", &url).await?;
|
|
self.decode_json::<TurnAttemptsSiblingTurnsResponse>(&url, &ct, &body)
|
|
}
|
|
|
|
/// Create a new task (user turn) by POSTing to the appropriate backend path
|
|
/// based on `path_style`. Returns the created task id.
|
|
pub async fn create_task(&self, request_body: serde_json::Value) -> Result<String> {
|
|
let url = match self.path_style {
|
|
PathStyle::LlmxApi => format!("{}/api/llmx/tasks", self.base_url),
|
|
PathStyle::ChatGptApi => format!("{}/wham/tasks", self.base_url),
|
|
};
|
|
let req = self
|
|
.http
|
|
.post(&url)
|
|
.headers(self.headers())
|
|
.header(CONTENT_TYPE, HeaderValue::from_static("application/json"))
|
|
.json(&request_body);
|
|
let (body, ct) = self.exec_request(req, "POST", &url).await?;
|
|
// Extract id from JSON: prefer `task.id`; fallback to top-level `id` when present.
|
|
match serde_json::from_str::<serde_json::Value>(&body) {
|
|
Ok(v) => {
|
|
if let Some(id) = v
|
|
.get("task")
|
|
.and_then(|t| t.get("id"))
|
|
.and_then(|s| s.as_str())
|
|
{
|
|
Ok(id.to_string())
|
|
} else if let Some(id) = v.get("id").and_then(|s| s.as_str()) {
|
|
Ok(id.to_string())
|
|
} else {
|
|
anyhow::bail!(
|
|
"POST {url} succeeded but no task id found; content-type={ct}; body={body}"
|
|
);
|
|
}
|
|
}
|
|
Err(e) => anyhow::bail!("Decode error for {url}: {e}; content-type={ct}; body={body}"),
|
|
}
|
|
}
|
|
|
|
// rate limit helpers
|
|
fn rate_limit_snapshot_from_payload(payload: RateLimitStatusPayload) -> RateLimitSnapshot {
|
|
let Some(details) = payload
|
|
.rate_limit
|
|
.and_then(|inner| inner.map(|boxed| *boxed))
|
|
else {
|
|
return RateLimitSnapshot {
|
|
primary: None,
|
|
secondary: None,
|
|
};
|
|
};
|
|
|
|
RateLimitSnapshot {
|
|
primary: Self::map_rate_limit_window(details.primary_window),
|
|
secondary: Self::map_rate_limit_window(details.secondary_window),
|
|
}
|
|
}
|
|
|
|
fn map_rate_limit_window(
|
|
window: Option<Option<Box<RateLimitWindowSnapshot>>>,
|
|
) -> Option<RateLimitWindow> {
|
|
let snapshot = match window {
|
|
Some(Some(snapshot)) => *snapshot,
|
|
_ => return None,
|
|
};
|
|
|
|
let used_percent = f64::from(snapshot.used_percent);
|
|
let window_minutes = Self::window_minutes_from_seconds(snapshot.limit_window_seconds);
|
|
let resets_at = Some(i64::from(snapshot.reset_at));
|
|
Some(RateLimitWindow {
|
|
used_percent,
|
|
window_minutes,
|
|
resets_at,
|
|
})
|
|
}
|
|
|
|
fn window_minutes_from_seconds(seconds: i32) -> Option<i64> {
|
|
if seconds <= 0 {
|
|
return None;
|
|
}
|
|
|
|
let seconds_i64 = i64::from(seconds);
|
|
Some((seconds_i64 + 59) / 60)
|
|
}
|
|
}
|