From 53e6e3b18aa221feb65cbe73a91bfbd458c6e214 Mon Sep 17 00:00:00 2001 From: Morgan Ewing Date: Tue, 22 Jul 2025 14:53:01 +1000 Subject: [PATCH] =?UTF-8?q?=F0=9F=93=9D=20docs:=20update=20model=20names?= =?UTF-8?q?=20in=20reasoning=20example?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Update example to use current model naming conventions: - Change Grok model to grok-3-mini - Change Claude model to claude-4-sonnet --- examples/openrouter_reasoning.rs | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/examples/openrouter_reasoning.rs b/examples/openrouter_reasoning.rs index 4f5e258..ea38d5b 100644 --- a/examples/openrouter_reasoning.rs +++ b/examples/openrouter_reasoning.rs @@ -12,7 +12,7 @@ async fn main() -> Result<(), Box> { // Example 1: Using reasoning with effort let mut req = ChatCompletionRequest::new( - "x-ai/grok-2-1212".to_string(), // Grok model that supports reasoning + "x-ai/grok-3-mini".to_string(), // Grok model that supports reasoning vec![chat_completion::ChatCompletionMessage { role: chat_completion::MessageRole::user, content: chat_completion::Content::Text(String::from("Explain quantum computing in simple terms.")), @@ -36,7 +36,7 @@ async fn main() -> Result<(), Box> { // Example 2: Using reasoning with max_tokens let mut req2 = ChatCompletionRequest::new( - "anthropic/claude-3.7-sonnet".to_string(), // Claude model that supports max_tokens + "anthropic/claude-4-sonnet".to_string(), // Claude model that supports max_tokens vec![chat_completion::ChatCompletionMessage { role: chat_completion::MessageRole::user, content: chat_completion::Content::Text(String::from("What's the most efficient sorting algorithm?")),