From 8f79e89db2edfbb2c6d665b0dcb0acc923b914ce Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Sebastian=20Kr=C3=BCger?= Date: Mon, 17 Nov 2025 11:01:31 +0100 Subject: [PATCH] fix: Add max_tokens field to all test ModelProviderInfo initializations MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit - Fixed all test files that initialize ModelProviderInfo - Added max_tokens: None to maintain compatibility - Ensures all tests compile with new max_tokens field 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude --- llmx-rs/core/src/client.rs | 7 +++++++ llmx-rs/core/src/config/mod.rs | 1 + llmx-rs/core/tests/chat_completions_payload.rs | 1 + llmx-rs/core/tests/chat_completions_sse.rs | 1 + llmx-rs/core/tests/responses_headers.rs | 2 ++ llmx-rs/core/tests/suite/client.rs | 3 +++ llmx-rs/core/tests/suite/stream_error_allows_next_turn.rs | 1 + llmx-rs/core/tests/suite/stream_no_completed.rs | 1 + 8 files changed, 17 insertions(+) diff --git a/llmx-rs/core/src/client.rs b/llmx-rs/core/src/client.rs index b06e7005..f949c2ba 100644 --- a/llmx-rs/core/src/client.rs +++ b/llmx-rs/core/src/client.rs @@ -1123,6 +1123,7 @@ mod tests { request_max_retries: Some(0), stream_max_retries: Some(0), stream_idle_timeout_ms: Some(1000), + max_tokens: None, requires_openai_auth: false, }; @@ -1187,6 +1188,7 @@ mod tests { request_max_retries: Some(0), stream_max_retries: Some(0), stream_idle_timeout_ms: Some(1000), + max_tokens: None, requires_openai_auth: false, }; @@ -1224,6 +1226,7 @@ mod tests { request_max_retries: Some(0), stream_max_retries: Some(0), stream_idle_timeout_ms: Some(1000), + max_tokens: None, requires_openai_auth: false, }; @@ -1263,6 +1266,7 @@ mod tests { request_max_retries: Some(0), stream_max_retries: Some(0), stream_idle_timeout_ms: Some(1000), + max_tokens: None, requires_openai_auth: false, }; @@ -1298,6 +1302,7 @@ mod tests { request_max_retries: Some(0), stream_max_retries: Some(0), stream_idle_timeout_ms: Some(1000), + max_tokens: None, requires_openai_auth: false, }; @@ -1333,6 +1338,7 @@ mod tests { request_max_retries: Some(0), stream_max_retries: Some(0), stream_idle_timeout_ms: Some(1000), + max_tokens: None, requires_openai_auth: false, }; @@ -1437,6 +1443,7 @@ mod tests { request_max_retries: Some(0), stream_max_retries: Some(0), stream_idle_timeout_ms: Some(1000), + max_tokens: None, requires_openai_auth: false, }; diff --git a/llmx-rs/core/src/config/mod.rs b/llmx-rs/core/src/config/mod.rs index 3be6914d..44ea4251 100644 --- a/llmx-rs/core/src/config/mod.rs +++ b/llmx-rs/core/src/config/mod.rs @@ -2809,6 +2809,7 @@ model_verbosity = "high" request_max_retries: Some(4), stream_max_retries: Some(10), stream_idle_timeout_ms: Some(300_000), + max_tokens: None, requires_openai_auth: false, }; let model_provider_map = { diff --git a/llmx-rs/core/tests/chat_completions_payload.rs b/llmx-rs/core/tests/chat_completions_payload.rs index 4d65bac6..9c619320 100644 --- a/llmx-rs/core/tests/chat_completions_payload.rs +++ b/llmx-rs/core/tests/chat_completions_payload.rs @@ -58,6 +58,7 @@ async fn run_request(input: Vec) -> Value { request_max_retries: Some(0), stream_max_retries: Some(0), stream_idle_timeout_ms: Some(5_000), + max_tokens: None, requires_openai_auth: false, }; diff --git a/llmx-rs/core/tests/chat_completions_sse.rs b/llmx-rs/core/tests/chat_completions_sse.rs index 688328da..81f11867 100644 --- a/llmx-rs/core/tests/chat_completions_sse.rs +++ b/llmx-rs/core/tests/chat_completions_sse.rs @@ -58,6 +58,7 @@ async fn run_stream_with_bytes(sse_body: &[u8]) -> Vec { request_max_retries: Some(0), stream_max_retries: Some(0), stream_idle_timeout_ms: Some(5_000), + max_tokens: None, requires_openai_auth: false, }; diff --git a/llmx-rs/core/tests/responses_headers.rs b/llmx-rs/core/tests/responses_headers.rs index f541408f..595d30ad 100644 --- a/llmx-rs/core/tests/responses_headers.rs +++ b/llmx-rs/core/tests/responses_headers.rs @@ -47,6 +47,7 @@ async fn responses_stream_includes_subagent_header_on_review() { request_max_retries: Some(0), stream_max_retries: Some(0), stream_idle_timeout_ms: Some(5_000), + max_tokens: None, requires_openai_auth: false, }; @@ -135,6 +136,7 @@ async fn responses_stream_includes_subagent_header_on_other() { request_max_retries: Some(0), stream_max_retries: Some(0), stream_idle_timeout_ms: Some(5_000), + max_tokens: None, requires_openai_auth: false, }; diff --git a/llmx-rs/core/tests/suite/client.rs b/llmx-rs/core/tests/suite/client.rs index 28bb552d..12aba54f 100644 --- a/llmx-rs/core/tests/suite/client.rs +++ b/llmx-rs/core/tests/suite/client.rs @@ -712,6 +712,7 @@ async fn azure_responses_request_includes_store_and_reasoning_ids() { request_max_retries: Some(0), stream_max_retries: Some(0), stream_idle_timeout_ms: Some(5_000), + max_tokens: None, requires_openai_auth: false, }; @@ -1195,6 +1196,7 @@ async fn azure_overrides_assign_properties_used_for_responses_url() { request_max_retries: None, stream_max_retries: None, stream_idle_timeout_ms: None, + max_tokens: None, requires_openai_auth: false, }; @@ -1272,6 +1274,7 @@ async fn env_var_overrides_loaded_auth() { request_max_retries: None, stream_max_retries: None, stream_idle_timeout_ms: None, + max_tokens: None, requires_openai_auth: false, }; diff --git a/llmx-rs/core/tests/suite/stream_error_allows_next_turn.rs b/llmx-rs/core/tests/suite/stream_error_allows_next_turn.rs index 9596b585..e4eacf7d 100644 --- a/llmx-rs/core/tests/suite/stream_error_allows_next_turn.rs +++ b/llmx-rs/core/tests/suite/stream_error_allows_next_turn.rs @@ -72,6 +72,7 @@ async fn continue_after_stream_error() { request_max_retries: Some(1), stream_max_retries: Some(1), stream_idle_timeout_ms: Some(2_000), + max_tokens: None, requires_openai_auth: false, }; diff --git a/llmx-rs/core/tests/suite/stream_no_completed.rs b/llmx-rs/core/tests/suite/stream_no_completed.rs index 740de9e4..acf3382b 100644 --- a/llmx-rs/core/tests/suite/stream_no_completed.rs +++ b/llmx-rs/core/tests/suite/stream_no_completed.rs @@ -80,6 +80,7 @@ async fn retries_on_early_close() { request_max_retries: Some(0), stream_max_retries: Some(1), stream_idle_timeout_ms: Some(2000), + max_tokens: None, requires_openai_auth: false, };