mirror of
https://github.com/openai/codex.git
synced 2026-05-17 09:43:19 +00:00
Fixes #21070. This is a small cleanup around model metadata handling for gateway/provider model names. It follows the report and proposed direction from @dkbush by keeping the fallback metadata warning useful without repeating it every turn, and by tightening the existing provider-prefix lookup path. - Track fallback metadata warning slugs in session state so each unresolved model warns once per session. - Keep warning emission outside the session-state lock and preserve the existing warning text. - Allow one-segment provider prefixes with hyphenated provider IDs, while preserving the multi-segment rejection behavior. - Add focused coverage for warning dedupe and hyphenated provider-prefix metadata matching. Testing: - Ran `just fmt`. - Ran `git diff --check`. - Added tests for the new warning dedupe and provider-prefix lookup behavior.
464 lines
16 KiB
Rust
464 lines
16 KiB
Rust
use super::cache::ModelsCacheManager;
|
|
use crate::collaboration_mode_presets::builtin_collaboration_mode_presets;
|
|
use crate::config::ModelsManagerConfig;
|
|
use crate::model_info;
|
|
use async_trait::async_trait;
|
|
use codex_login::AuthManager;
|
|
use codex_protocol::config_types::CollaborationModeMask;
|
|
use codex_protocol::error::Result as CoreResult;
|
|
use codex_protocol::openai_models::ModelInfo;
|
|
use codex_protocol::openai_models::ModelPreset;
|
|
use codex_protocol::openai_models::ModelsResponse;
|
|
use std::fmt;
|
|
use std::path::PathBuf;
|
|
use std::sync::Arc;
|
|
use std::time::Duration;
|
|
use tokio::sync::RwLock;
|
|
use tokio::sync::TryLockError;
|
|
use tracing::Instrument as _;
|
|
use tracing::error;
|
|
use tracing::info;
|
|
|
|
const MODEL_CACHE_FILE: &str = "models_cache.json";
|
|
const DEFAULT_MODEL_CACHE_TTL: Duration = Duration::from_secs(300);
|
|
|
|
/// Remote endpoint used by the OpenAI-compatible model manager.
|
|
///
|
|
/// Implementations own provider-specific auth and transport details. The model
|
|
/// manager owns refresh policy, cache behavior, and catalog merging; it calls
|
|
/// this endpoint only when it decides a remote refresh should happen.
|
|
#[async_trait]
|
|
pub trait ModelsEndpointClient: fmt::Debug + Send + Sync {
|
|
/// Returns whether this provider can authenticate command-scoped requests.
|
|
fn has_command_auth(&self) -> bool;
|
|
|
|
/// Returns whether the currently resolved auth can use Codex backend-only models.
|
|
async fn uses_codex_backend(&self) -> bool;
|
|
|
|
/// Fetches the latest remote model catalog and optional ETag.
|
|
async fn list_models(
|
|
&self,
|
|
client_version: &str,
|
|
) -> CoreResult<(Vec<ModelInfo>, Option<String>)>;
|
|
}
|
|
|
|
/// Strategy for refreshing available models.
|
|
#[derive(Debug, Clone, Copy, PartialEq, Eq)]
|
|
pub enum RefreshStrategy {
|
|
/// Always fetch from the network, ignoring cache.
|
|
Online,
|
|
/// Only use cached data, never fetch from the network.
|
|
Offline,
|
|
/// Use cache if available and fresh, otherwise fetch from the network.
|
|
OnlineIfUncached,
|
|
}
|
|
|
|
impl RefreshStrategy {
|
|
const fn as_str(self) -> &'static str {
|
|
match self {
|
|
Self::Online => "online",
|
|
Self::Offline => "offline",
|
|
Self::OnlineIfUncached => "online_if_uncached",
|
|
}
|
|
}
|
|
}
|
|
|
|
impl fmt::Display for RefreshStrategy {
|
|
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
|
f.write_str(self.as_str())
|
|
}
|
|
}
|
|
|
|
type SharedModelsEndpointClient = Arc<dyn ModelsEndpointClient>;
|
|
|
|
/// Coordinates model discovery plus cached metadata on disk.
|
|
#[async_trait]
|
|
pub trait ModelsManager: fmt::Debug + Send + Sync {
|
|
/// List all available models, refreshing according to the specified strategy.
|
|
///
|
|
/// Returns model presets sorted by priority and filtered by auth mode and visibility.
|
|
async fn list_models(&self, refresh_strategy: RefreshStrategy) -> Vec<ModelPreset> {
|
|
async move {
|
|
let catalog = self.raw_model_catalog(refresh_strategy).await;
|
|
self.build_available_models(catalog.models)
|
|
}
|
|
.instrument(tracing::info_span!(
|
|
"list_models",
|
|
refresh_strategy = %refresh_strategy
|
|
))
|
|
.await
|
|
}
|
|
|
|
/// Return the active raw model catalog, refreshing according to the specified strategy.
|
|
async fn raw_model_catalog(&self, refresh_strategy: RefreshStrategy) -> ModelsResponse;
|
|
|
|
/// Return the current in-memory remote model catalog without refreshing or loading cache state.
|
|
async fn get_remote_models(&self) -> Vec<ModelInfo>;
|
|
|
|
/// Attempt to return the current in-memory remote model catalog without blocking.
|
|
///
|
|
/// Returns an error if the internal lock cannot be acquired.
|
|
fn try_get_remote_models(&self) -> Result<Vec<ModelInfo>, TryLockError>;
|
|
|
|
/// Return the auth manager used for picker filtering.
|
|
fn auth_manager(&self) -> Option<&AuthManager>;
|
|
|
|
/// Build picker-ready presets from the active catalog snapshot.
|
|
fn build_available_models(&self, mut remote_models: Vec<ModelInfo>) -> Vec<ModelPreset> {
|
|
remote_models.sort_by(|a, b| a.priority.cmp(&b.priority));
|
|
|
|
let mut presets: Vec<ModelPreset> = remote_models.into_iter().map(Into::into).collect();
|
|
let uses_codex_backend = self
|
|
.auth_manager()
|
|
.is_some_and(AuthManager::current_auth_uses_codex_backend);
|
|
presets = ModelPreset::filter_by_auth(presets, uses_codex_backend);
|
|
|
|
ModelPreset::mark_default_by_picker_visibility(&mut presets);
|
|
|
|
presets
|
|
}
|
|
|
|
/// List collaboration mode presets.
|
|
///
|
|
/// Returns a static set of presets seeded with the configured model.
|
|
fn list_collaboration_modes(&self) -> Vec<CollaborationModeMask>;
|
|
|
|
/// Attempt to list models without blocking, using the current cached state.
|
|
///
|
|
/// Returns an error if the internal lock cannot be acquired.
|
|
fn try_list_models(&self) -> Result<Vec<ModelPreset>, TryLockError> {
|
|
let remote_models = self.try_get_remote_models()?;
|
|
Ok(self.build_available_models(remote_models))
|
|
}
|
|
|
|
// todo(aibrahim): should be visible to core only and sent on session_configured event
|
|
/// Get the model identifier to use, refreshing according to the specified strategy.
|
|
///
|
|
/// If `model` is provided, returns it directly. Otherwise selects the default based on
|
|
/// auth mode and available models.
|
|
async fn get_default_model(
|
|
&self,
|
|
model: &Option<String>,
|
|
refresh_strategy: RefreshStrategy,
|
|
) -> String {
|
|
async move {
|
|
if let Some(model) = model.as_ref() {
|
|
return model.to_string();
|
|
}
|
|
default_model_from_available(self.list_models(refresh_strategy).await)
|
|
}
|
|
.instrument(tracing::info_span!(
|
|
"get_default_model",
|
|
model.provided = model.is_some(),
|
|
refresh_strategy = %refresh_strategy
|
|
))
|
|
.await
|
|
}
|
|
|
|
// todo(aibrahim): look if we can tighten it to pub(crate)
|
|
/// Look up model metadata, applying remote overrides and config adjustments.
|
|
async fn get_model_info(&self, model: &str, config: &ModelsManagerConfig) -> ModelInfo {
|
|
async move {
|
|
let remote_models = self.get_remote_models().await;
|
|
construct_model_info_from_candidates(model, &remote_models, config)
|
|
}
|
|
.instrument(tracing::info_span!("get_model_info", model = model))
|
|
.await
|
|
}
|
|
|
|
/// Refresh models if the provided ETag differs from the cached ETag.
|
|
///
|
|
/// Uses `Online` strategy to fetch latest models when ETags differ.
|
|
async fn refresh_if_new_etag(&self, etag: String);
|
|
}
|
|
|
|
/// Shared model manager handle used across runtime services.
|
|
pub type SharedModelsManager = Arc<dyn ModelsManager>;
|
|
|
|
/// OpenAI-compatible model manager backed by bundled models, cache, and `/models`.
|
|
#[derive(Debug)]
|
|
pub struct OpenAiModelsManager {
|
|
remote_models: RwLock<Vec<ModelInfo>>,
|
|
etag: RwLock<Option<String>>,
|
|
cache_manager: ModelsCacheManager,
|
|
endpoint_client: SharedModelsEndpointClient,
|
|
auth_manager: Option<Arc<AuthManager>>,
|
|
}
|
|
|
|
/// Static model manager backed by an authoritative in-process catalog.
|
|
#[derive(Debug)]
|
|
pub struct StaticModelsManager {
|
|
remote_models: Vec<ModelInfo>,
|
|
auth_manager: Option<Arc<AuthManager>>,
|
|
}
|
|
|
|
impl OpenAiModelsManager {
|
|
/// Construct an OpenAI-compatible remote model manager.
|
|
pub fn new(
|
|
codex_home: PathBuf,
|
|
endpoint_client: Arc<dyn ModelsEndpointClient>,
|
|
auth_manager: Option<Arc<AuthManager>>,
|
|
) -> Self {
|
|
let cache_path = codex_home.join(MODEL_CACHE_FILE);
|
|
let cache_manager = ModelsCacheManager::new(cache_path, DEFAULT_MODEL_CACHE_TTL);
|
|
let remote_models = load_remote_models_from_file().unwrap_or_default();
|
|
Self {
|
|
remote_models: RwLock::new(remote_models),
|
|
etag: RwLock::new(None),
|
|
cache_manager,
|
|
endpoint_client,
|
|
auth_manager,
|
|
}
|
|
}
|
|
}
|
|
|
|
impl StaticModelsManager {
|
|
/// Construct a static model manager from an authoritative catalog.
|
|
pub fn new(auth_manager: Option<Arc<AuthManager>>, model_catalog: ModelsResponse) -> Self {
|
|
Self {
|
|
remote_models: model_catalog.models,
|
|
auth_manager,
|
|
}
|
|
}
|
|
}
|
|
|
|
#[async_trait]
|
|
impl ModelsManager for OpenAiModelsManager {
|
|
async fn raw_model_catalog(&self, refresh_strategy: RefreshStrategy) -> ModelsResponse {
|
|
if let Err(err) = self.refresh_available_models(refresh_strategy).await {
|
|
error!("failed to refresh available models: {err}");
|
|
}
|
|
ModelsResponse {
|
|
models: self.get_remote_models().await,
|
|
}
|
|
}
|
|
|
|
async fn get_remote_models(&self) -> Vec<ModelInfo> {
|
|
self.remote_models.read().await.clone()
|
|
}
|
|
|
|
fn try_get_remote_models(&self) -> Result<Vec<ModelInfo>, TryLockError> {
|
|
Ok(self.remote_models.try_read()?.clone())
|
|
}
|
|
|
|
fn auth_manager(&self) -> Option<&AuthManager> {
|
|
self.auth_manager.as_deref()
|
|
}
|
|
|
|
fn list_collaboration_modes(&self) -> Vec<CollaborationModeMask> {
|
|
builtin_collaboration_mode_presets()
|
|
}
|
|
|
|
async fn refresh_if_new_etag(&self, etag: String) {
|
|
let current_etag = self.get_etag().await;
|
|
if current_etag.clone().is_some() && current_etag.as_deref() == Some(etag.as_str()) {
|
|
if let Err(err) = self.cache_manager.renew_cache_ttl().await {
|
|
error!("failed to renew cache TTL: {err}");
|
|
}
|
|
return;
|
|
}
|
|
if let Err(err) = self.refresh_available_models(RefreshStrategy::Online).await {
|
|
error!("failed to refresh available models: {err}");
|
|
}
|
|
}
|
|
}
|
|
|
|
impl OpenAiModelsManager {
|
|
/// Refresh available models according to the specified strategy.
|
|
async fn refresh_available_models(&self, refresh_strategy: RefreshStrategy) -> CoreResult<()> {
|
|
if !self.should_refresh_models().await {
|
|
if matches!(
|
|
refresh_strategy,
|
|
RefreshStrategy::Offline | RefreshStrategy::OnlineIfUncached
|
|
) {
|
|
self.try_load_cache().await;
|
|
}
|
|
return Ok(());
|
|
}
|
|
|
|
match refresh_strategy {
|
|
RefreshStrategy::Offline => {
|
|
// Only try to load from cache, never fetch
|
|
self.try_load_cache().await;
|
|
Ok(())
|
|
}
|
|
RefreshStrategy::OnlineIfUncached => {
|
|
// Try cache first, fall back to online if unavailable
|
|
if self.try_load_cache().await {
|
|
info!("models cache: using cached models for OnlineIfUncached");
|
|
return Ok(());
|
|
}
|
|
info!("models cache: cache miss, fetching remote models");
|
|
self.fetch_and_update_models().await
|
|
}
|
|
RefreshStrategy::Online => {
|
|
// Always fetch from network
|
|
self.fetch_and_update_models().await
|
|
}
|
|
}
|
|
}
|
|
|
|
async fn fetch_and_update_models(&self) -> CoreResult<()> {
|
|
let client_version = crate::client_version_to_whole();
|
|
let (models, etag) = self.endpoint_client.list_models(&client_version).await?;
|
|
self.apply_remote_models(models.clone()).await;
|
|
*self.etag.write().await = etag.clone();
|
|
self.cache_manager
|
|
.persist_cache(&models, etag, client_version)
|
|
.await;
|
|
Ok(())
|
|
}
|
|
|
|
async fn should_refresh_models(&self) -> bool {
|
|
self.endpoint_client.uses_codex_backend().await || self.endpoint_client.has_command_auth()
|
|
}
|
|
|
|
async fn get_etag(&self) -> Option<String> {
|
|
self.etag.read().await.clone()
|
|
}
|
|
|
|
/// Replace the cached remote models and rebuild the derived presets list.
|
|
async fn apply_remote_models(&self, models: Vec<ModelInfo>) {
|
|
let mut existing_models = load_remote_models_from_file().unwrap_or_default();
|
|
for model in models {
|
|
if let Some(existing_index) = existing_models
|
|
.iter()
|
|
.position(|existing| existing.slug == model.slug)
|
|
{
|
|
existing_models[existing_index] = model;
|
|
} else {
|
|
existing_models.push(model);
|
|
}
|
|
}
|
|
*self.remote_models.write().await = existing_models;
|
|
}
|
|
|
|
/// Attempt to satisfy the refresh from the cache when it matches the provider and TTL.
|
|
async fn try_load_cache(&self) -> bool {
|
|
let _timer =
|
|
codex_otel::start_global_timer("codex.remote_models.load_cache.duration_ms", &[]);
|
|
let client_version = crate::client_version_to_whole();
|
|
info!(client_version, "models cache: evaluating cache eligibility");
|
|
// TODO(celia-oai): Include provider identity in cache eligibility so switching
|
|
// providers does not reuse a fresh models_cache.json entry from another provider.
|
|
let cache = match self.cache_manager.load_fresh(&client_version).await {
|
|
Some(cache) => cache,
|
|
None => {
|
|
info!("models cache: no usable cache entry");
|
|
return false;
|
|
}
|
|
};
|
|
let models = cache.models.clone();
|
|
*self.etag.write().await = cache.etag.clone();
|
|
self.apply_remote_models(models.clone()).await;
|
|
info!(
|
|
models_count = models.len(),
|
|
etag = ?cache.etag,
|
|
"models cache: cache entry applied"
|
|
);
|
|
true
|
|
}
|
|
}
|
|
|
|
#[async_trait]
|
|
impl ModelsManager for StaticModelsManager {
|
|
async fn raw_model_catalog(&self, _refresh_strategy: RefreshStrategy) -> ModelsResponse {
|
|
ModelsResponse {
|
|
models: self.get_remote_models().await,
|
|
}
|
|
}
|
|
|
|
async fn get_remote_models(&self) -> Vec<ModelInfo> {
|
|
self.remote_models.clone()
|
|
}
|
|
|
|
fn try_get_remote_models(&self) -> Result<Vec<ModelInfo>, TryLockError> {
|
|
Ok(self.remote_models.clone())
|
|
}
|
|
|
|
fn auth_manager(&self) -> Option<&AuthManager> {
|
|
self.auth_manager.as_deref()
|
|
}
|
|
|
|
fn list_collaboration_modes(&self) -> Vec<CollaborationModeMask> {
|
|
builtin_collaboration_mode_presets()
|
|
}
|
|
|
|
async fn refresh_if_new_etag(&self, _etag: String) {}
|
|
}
|
|
|
|
fn load_remote_models_from_file() -> Result<Vec<ModelInfo>, std::io::Error> {
|
|
Ok(crate::bundled_models_response()?.models)
|
|
}
|
|
|
|
fn default_model_from_available(available: Vec<ModelPreset>) -> String {
|
|
available
|
|
.iter()
|
|
.find(|model| model.is_default)
|
|
.or_else(|| available.first())
|
|
.map(|model| model.model.clone())
|
|
.unwrap_or_default()
|
|
}
|
|
|
|
fn find_model_by_longest_prefix(model: &str, candidates: &[ModelInfo]) -> Option<ModelInfo> {
|
|
let mut best: Option<ModelInfo> = None;
|
|
for candidate in candidates {
|
|
if !model.starts_with(&candidate.slug) {
|
|
continue;
|
|
}
|
|
let is_better_match = if let Some(current) = best.as_ref() {
|
|
candidate.slug.len() > current.slug.len()
|
|
} else {
|
|
true
|
|
};
|
|
if is_better_match {
|
|
best = Some(candidate.clone());
|
|
}
|
|
}
|
|
best
|
|
}
|
|
|
|
fn find_model_by_namespaced_suffix(model: &str, candidates: &[ModelInfo]) -> Option<ModelInfo> {
|
|
// Retry metadata lookup for a single namespaced slug like `namespace/model-name`.
|
|
//
|
|
// This only strips one leading namespace segment and only when the namespace looks
|
|
// like a simple provider id to avoid broadly matching arbitrary aliases.
|
|
let (namespace, suffix) = model.split_once('/')?;
|
|
if suffix.contains('/') {
|
|
return None;
|
|
}
|
|
if namespace.is_empty()
|
|
|| !namespace
|
|
.chars()
|
|
.all(|c| c.is_ascii_alphanumeric() || c == '_' || c == '-')
|
|
{
|
|
return None;
|
|
}
|
|
find_model_by_longest_prefix(suffix, candidates)
|
|
}
|
|
|
|
pub(crate) fn construct_model_info_from_candidates(
|
|
model: &str,
|
|
candidates: &[ModelInfo],
|
|
config: &ModelsManagerConfig,
|
|
) -> ModelInfo {
|
|
// First use the normal longest-prefix match. If that misses, allow a narrowly scoped
|
|
// retry for namespaced slugs like `custom/gpt-5.3-codex`.
|
|
let remote = find_model_by_longest_prefix(model, candidates)
|
|
.or_else(|| find_model_by_namespaced_suffix(model, candidates));
|
|
let model_info = if let Some(remote) = remote {
|
|
ModelInfo {
|
|
slug: model.to_string(),
|
|
used_fallback_model_metadata: false,
|
|
..remote
|
|
}
|
|
} else {
|
|
model_info::model_info_from_slug(model)
|
|
};
|
|
model_info::with_config_overrides(model_info, config)
|
|
}
|
|
|
|
#[cfg(test)]
|
|
#[path = "manager_tests.rs"]
|
|
mod tests;
|