Compare commits

...

2 Commits

Author SHA1 Message Date
be9fdd9a52 fix(gemini): implement dynamic API versioning and support Gemini 3
Some checks failed
CI / Check (push) Has been cancelled
CI / Clippy (push) Has been cancelled
CI / Formatting (push) Has been cancelled
CI / Test (push) Has been cancelled
CI / Release Build (push) Has been cancelled
- Switch to v1beta endpoint for 'preview' and 'thinking' models.
- Update model version checks to include gemini-3 as a known version.
- Use get_base_url helper to construct dynamic URLs for both streaming and non-streaming requests.
2026-03-05 15:24:47 +00:00
e89658fd87 fix(gemini): add model mapping fallback and token clamping
- Map unknown Gemini model names to the configured default model to prevent 400 errors.
- Clamp max_tokens to a safe limit of 8192 for Gemini models.
- Clean up message filtering and role injection for better client compatibility.
2026-03-05 15:23:26 +00:00

View File

@@ -439,6 +439,17 @@ impl GeminiProvider {
} }
} }
/// Determine the appropriate base URL for the model.
/// "preview" models often require the v1beta endpoint.
fn get_base_url(&self, model: &str) -> String {
if model.contains("preview") || model.contains("thinking") {
self.config.base_url.replace("/v1", "/v1beta")
} else {
self.config.base_url.clone()
}
}
}
#[async_trait] #[async_trait]
impl super::Provider for GeminiProvider { impl super::Provider for GeminiProvider {
fn name(&self) -> &str { fn name(&self) -> &str {
@@ -454,7 +465,22 @@ impl super::Provider for GeminiProvider {
} }
async fn chat_completion(&self, request: UnifiedRequest) -> Result<ProviderResponse, AppError> { async fn chat_completion(&self, request: UnifiedRequest) -> Result<ProviderResponse, AppError> {
let model = request.model.clone(); let mut model = request.model.clone();
// Normalize model name: If it's a known Gemini model version, use it;
// otherwise, if it starts with gemini- but is an unknown legacy version,
// fallback to the default model to avoid 400 errors.
// We now allow gemini-3+ as valid versions.
let is_known_version = model.starts_with("gemini-1.5") ||
model.starts_with("gemini-2.0") ||
model.starts_with("gemini-2.5") ||
model.starts_with("gemini-3");
if !is_known_version && model.starts_with("gemini-") {
tracing::info!("Mapping unknown Gemini model {} to default {}", model, self.config.default_model);
model = self.config.default_model.clone();
}
let tools = Self::convert_tools(&request); let tools = Self::convert_tools(&request);
let tool_config = Self::convert_tool_config(&request); let tool_config = Self::convert_tool_config(&request);
let (contents, system_instruction) = Self::convert_messages(request.messages.clone()).await?; let (contents, system_instruction) = Self::convert_messages(request.messages.clone()).await?;
@@ -464,9 +490,14 @@ impl super::Provider for GeminiProvider {
} }
let generation_config = if request.temperature.is_some() || request.max_tokens.is_some() { let generation_config = if request.temperature.is_some() || request.max_tokens.is_some() {
// Some Gemini models (especially 1.5) have lower max_output_tokens limits (e.g. 8192)
// than what clients like opencode might request. Clamp to a safe maximum.
// Note: Gemini 2.0+ supports much higher limits, but 8192 is a safe universal floor.
let max_tokens = request.max_tokens.map(|t| t.min(8192));
Some(GeminiGenerationConfig { Some(GeminiGenerationConfig {
temperature: request.temperature, temperature: request.temperature,
max_output_tokens: request.max_tokens, max_output_tokens: max_tokens,
}) })
} else { } else {
None None
@@ -480,7 +511,8 @@ impl super::Provider for GeminiProvider {
tool_config, tool_config,
}; };
let url = format!("{}/models/{}:generateContent", self.config.base_url, model); let base_url = self.get_base_url(&model);
let url = format!("{}/models/{}:generateContent", base_url, model);
let response = self let response = self
.client .client
@@ -579,7 +611,19 @@ impl super::Provider for GeminiProvider {
&self, &self,
request: UnifiedRequest, request: UnifiedRequest,
) -> Result<BoxStream<'static, Result<ProviderStreamChunk, AppError>>, AppError> { ) -> Result<BoxStream<'static, Result<ProviderStreamChunk, AppError>>, AppError> {
let model = request.model.clone(); let mut model = request.model.clone();
// Normalize model name: fallback to default if unknown Gemini model is requested
let is_known_version = model.starts_with("gemini-1.5") ||
model.starts_with("gemini-2.0") ||
model.starts_with("gemini-2.5") ||
model.starts_with("gemini-3");
if !is_known_version && model.starts_with("gemini-") {
tracing::info!("Mapping unknown Gemini model {} to default {}", model, self.config.default_model);
model = self.config.default_model.clone();
}
let tools = Self::convert_tools(&request); let tools = Self::convert_tools(&request);
let tool_config = Self::convert_tool_config(&request); let tool_config = Self::convert_tool_config(&request);
let (contents, system_instruction) = Self::convert_messages(request.messages.clone()).await?; let (contents, system_instruction) = Self::convert_messages(request.messages.clone()).await?;
@@ -589,9 +633,13 @@ impl super::Provider for GeminiProvider {
} }
let generation_config = if request.temperature.is_some() || request.max_tokens.is_some() { let generation_config = if request.temperature.is_some() || request.max_tokens.is_some() {
// Some Gemini models (especially 1.5) have lower max_output_tokens limits (e.g. 8192)
// than what clients like opencode might request. Clamp to a safe maximum.
let max_tokens = request.max_tokens.map(|t| t.min(8192));
Some(GeminiGenerationConfig { Some(GeminiGenerationConfig {
temperature: request.temperature, temperature: request.temperature,
max_output_tokens: request.max_tokens, max_output_tokens: max_tokens,
}) })
} else { } else {
None None
@@ -605,9 +653,10 @@ impl super::Provider for GeminiProvider {
tool_config, tool_config,
}; };
let base_url = self.get_base_url(&model);
let url = format!( let url = format!(
"{}/models/{}:streamGenerateContent?alt=sse", "{}/models/{}:streamGenerateContent?alt=sse",
self.config.base_url, model, base_url, model,
); );
// (no fallback_request needed here) // (no fallback_request needed here)
@@ -622,7 +671,7 @@ impl super::Provider for GeminiProvider {
// Prepare clones for HTTP fallback usage inside non-streaming paths. // Prepare clones for HTTP fallback usage inside non-streaming paths.
let http_client = self.client.clone(); let http_client = self.client.clone();
let http_api_key = self.api_key.clone(); let http_api_key = self.api_key.clone();
let http_base = self.config.base_url.clone(); let http_base = base_url.clone();
let gemini_request_clone = gemini_request.clone(); let gemini_request_clone = gemini_request.clone();
let es_result = reqwest_eventsource::EventSource::new( let es_result = reqwest_eventsource::EventSource::new(