mirror of
https://github.com/openai/codex.git
synced 2026-04-24 14:45:27 +00:00
## Summary - update documentation, example configs, and automation defaults to reference gpt-5.1 / gpt-5.1-codex - bump the CLI and core configuration defaults, model presets, and error messaging to the new models while keeping the model-family/tool coverage for legacy slugs - refresh tests, fixtures, and TUI snapshots so they expect the upgraded defaults ## Testing - `cargo test -p codex-core config::tests::test_precedence_fixture_with_gpt5_profile` ------ [Codex Task](https://chatgpt.com/codex/tasks/task_i_6916c5b3c2b08321ace04ee38604fc6b)
287 lines
8.9 KiB
Rust
287 lines
8.9 KiB
Rust
#![cfg(not(target_os = "windows"))]
|
|
#![allow(clippy::unwrap_used)]
|
|
|
|
use std::time::Duration;
|
|
use std::time::Instant;
|
|
|
|
use codex_core::model_family::find_family_for_model;
|
|
use codex_core::protocol::AskForApproval;
|
|
use codex_core::protocol::EventMsg;
|
|
use codex_core::protocol::Op;
|
|
use codex_core::protocol::SandboxPolicy;
|
|
use codex_protocol::config_types::ReasoningSummary;
|
|
use codex_protocol::user_input::UserInput;
|
|
use core_test_support::responses::ev_assistant_message;
|
|
use core_test_support::responses::ev_completed;
|
|
use core_test_support::responses::ev_function_call;
|
|
use core_test_support::responses::mount_sse_once;
|
|
use core_test_support::responses::mount_sse_sequence;
|
|
use core_test_support::responses::sse;
|
|
use core_test_support::responses::start_mock_server;
|
|
use core_test_support::skip_if_no_network;
|
|
use core_test_support::test_codex::TestCodex;
|
|
use core_test_support::test_codex::test_codex;
|
|
use core_test_support::wait_for_event;
|
|
use serde_json::Value;
|
|
use serde_json::json;
|
|
|
|
async fn run_turn(test: &TestCodex, prompt: &str) -> anyhow::Result<()> {
|
|
let session_model = test.session_configured.model.clone();
|
|
|
|
test.codex
|
|
.submit(Op::UserTurn {
|
|
items: vec![UserInput::Text {
|
|
text: prompt.into(),
|
|
}],
|
|
final_output_json_schema: None,
|
|
cwd: test.cwd.path().to_path_buf(),
|
|
approval_policy: AskForApproval::Never,
|
|
sandbox_policy: SandboxPolicy::DangerFullAccess,
|
|
model: session_model,
|
|
effort: None,
|
|
summary: ReasoningSummary::Auto,
|
|
})
|
|
.await?;
|
|
|
|
wait_for_event(&test.codex, |ev| matches!(ev, EventMsg::TaskComplete(_))).await;
|
|
|
|
Ok(())
|
|
}
|
|
|
|
async fn run_turn_and_measure(test: &TestCodex, prompt: &str) -> anyhow::Result<Duration> {
|
|
let start = Instant::now();
|
|
run_turn(test, prompt).await?;
|
|
Ok(start.elapsed())
|
|
}
|
|
|
|
#[allow(clippy::expect_used)]
|
|
async fn build_codex_with_test_tool(server: &wiremock::MockServer) -> anyhow::Result<TestCodex> {
|
|
let mut builder = test_codex().with_config(|config| {
|
|
config.model = "test-gpt-5.1-codex".to_string();
|
|
config.model_family =
|
|
find_family_for_model("test-gpt-5.1-codex").expect("test-gpt-5.1-codex model family");
|
|
});
|
|
builder.build(server).await
|
|
}
|
|
|
|
fn assert_parallel_duration(actual: Duration) {
|
|
// Allow headroom for runtime overhead while still differentiating from serial execution.
|
|
assert!(
|
|
actual < Duration::from_millis(750),
|
|
"expected parallel execution to finish quickly, got {actual:?}"
|
|
);
|
|
}
|
|
|
|
fn assert_serial_duration(actual: Duration) {
|
|
assert!(
|
|
actual >= Duration::from_millis(500),
|
|
"expected serial execution to take longer, got {actual:?}"
|
|
);
|
|
}
|
|
|
|
#[tokio::test(flavor = "multi_thread", worker_threads = 2)]
|
|
async fn read_file_tools_run_in_parallel() -> anyhow::Result<()> {
|
|
skip_if_no_network!(Ok(()));
|
|
|
|
let server = start_mock_server().await;
|
|
let test = build_codex_with_test_tool(&server).await?;
|
|
|
|
let warmup_args = json!({
|
|
"sleep_after_ms": 10,
|
|
"barrier": {
|
|
"id": "parallel-test-sync-warmup",
|
|
"participants": 2,
|
|
"timeout_ms": 1_000,
|
|
}
|
|
})
|
|
.to_string();
|
|
|
|
let parallel_args = json!({
|
|
"sleep_after_ms": 300,
|
|
"barrier": {
|
|
"id": "parallel-test-sync",
|
|
"participants": 2,
|
|
"timeout_ms": 1_000,
|
|
}
|
|
})
|
|
.to_string();
|
|
|
|
let warmup_first = sse(vec![
|
|
json!({"type": "response.created", "response": {"id": "resp-warm-1"}}),
|
|
ev_function_call("warm-call-1", "test_sync_tool", &warmup_args),
|
|
ev_function_call("warm-call-2", "test_sync_tool", &warmup_args),
|
|
ev_completed("resp-warm-1"),
|
|
]);
|
|
let warmup_second = sse(vec![
|
|
ev_assistant_message("warm-msg-1", "warmup complete"),
|
|
ev_completed("resp-warm-2"),
|
|
]);
|
|
|
|
let first_response = sse(vec![
|
|
json!({"type": "response.created", "response": {"id": "resp-1"}}),
|
|
ev_function_call("call-1", "test_sync_tool", ¶llel_args),
|
|
ev_function_call("call-2", "test_sync_tool", ¶llel_args),
|
|
ev_completed("resp-1"),
|
|
]);
|
|
let second_response = sse(vec![
|
|
ev_assistant_message("msg-1", "done"),
|
|
ev_completed("resp-2"),
|
|
]);
|
|
mount_sse_sequence(
|
|
&server,
|
|
vec![warmup_first, warmup_second, first_response, second_response],
|
|
)
|
|
.await;
|
|
|
|
run_turn(&test, "warm up parallel tool").await?;
|
|
|
|
let duration = run_turn_and_measure(&test, "exercise sync tool").await?;
|
|
assert_parallel_duration(duration);
|
|
|
|
Ok(())
|
|
}
|
|
|
|
#[tokio::test(flavor = "multi_thread", worker_threads = 2)]
|
|
async fn non_parallel_tools_run_serially() -> anyhow::Result<()> {
|
|
skip_if_no_network!(Ok(()));
|
|
|
|
let server = start_mock_server().await;
|
|
let test = test_codex().build(&server).await?;
|
|
|
|
let shell_args = json!({
|
|
"command": ["/bin/sh", "-c", "sleep 0.3"],
|
|
"timeout_ms": 1_000,
|
|
});
|
|
let args_one = serde_json::to_string(&shell_args)?;
|
|
let args_two = serde_json::to_string(&shell_args)?;
|
|
|
|
let first_response = sse(vec![
|
|
json!({"type": "response.created", "response": {"id": "resp-1"}}),
|
|
ev_function_call("call-1", "shell", &args_one),
|
|
ev_function_call("call-2", "shell", &args_two),
|
|
ev_completed("resp-1"),
|
|
]);
|
|
let second_response = sse(vec![
|
|
ev_assistant_message("msg-1", "done"),
|
|
ev_completed("resp-2"),
|
|
]);
|
|
mount_sse_sequence(&server, vec![first_response, second_response]).await;
|
|
|
|
let duration = run_turn_and_measure(&test, "run shell twice").await?;
|
|
assert_serial_duration(duration);
|
|
|
|
Ok(())
|
|
}
|
|
|
|
#[tokio::test(flavor = "multi_thread", worker_threads = 2)]
|
|
async fn mixed_tools_fall_back_to_serial() -> anyhow::Result<()> {
|
|
skip_if_no_network!(Ok(()));
|
|
|
|
let server = start_mock_server().await;
|
|
let test = build_codex_with_test_tool(&server).await?;
|
|
|
|
let sync_args = json!({
|
|
"sleep_after_ms": 300
|
|
})
|
|
.to_string();
|
|
let shell_args = serde_json::to_string(&json!({
|
|
"command": ["/bin/sh", "-c", "sleep 0.3"],
|
|
"timeout_ms": 1_000,
|
|
}))?;
|
|
|
|
let first_response = sse(vec![
|
|
json!({"type": "response.created", "response": {"id": "resp-1"}}),
|
|
ev_function_call("call-1", "test_sync_tool", &sync_args),
|
|
ev_function_call("call-2", "shell", &shell_args),
|
|
ev_completed("resp-1"),
|
|
]);
|
|
let second_response = sse(vec![
|
|
ev_assistant_message("msg-1", "done"),
|
|
ev_completed("resp-2"),
|
|
]);
|
|
mount_sse_sequence(&server, vec![first_response, second_response]).await;
|
|
|
|
let duration = run_turn_and_measure(&test, "mix tools").await?;
|
|
assert_serial_duration(duration);
|
|
|
|
Ok(())
|
|
}
|
|
|
|
#[tokio::test(flavor = "multi_thread", worker_threads = 2)]
|
|
async fn tool_results_grouped() -> anyhow::Result<()> {
|
|
skip_if_no_network!(Ok(()));
|
|
|
|
let server = start_mock_server().await;
|
|
let test = build_codex_with_test_tool(&server).await?;
|
|
|
|
let shell_args = serde_json::to_string(&json!({
|
|
"command": ["/bin/sh", "-c", "echo 'shell output'"],
|
|
"timeout_ms": 1_000,
|
|
}))?;
|
|
|
|
mount_sse_once(
|
|
&server,
|
|
sse(vec![
|
|
json!({"type": "response.created", "response": {"id": "resp-1"}}),
|
|
ev_function_call("call-1", "shell", &shell_args),
|
|
ev_function_call("call-2", "shell", &shell_args),
|
|
ev_function_call("call-3", "shell", &shell_args),
|
|
ev_completed("resp-1"),
|
|
]),
|
|
)
|
|
.await;
|
|
let tool_output_request = mount_sse_once(
|
|
&server,
|
|
sse(vec![
|
|
ev_assistant_message("msg-1", "done"),
|
|
ev_completed("resp-2"),
|
|
]),
|
|
)
|
|
.await;
|
|
|
|
run_turn(&test, "run shell three times").await?;
|
|
|
|
let input = tool_output_request.single_request().input();
|
|
|
|
// find all function_call inputs with indexes
|
|
let function_calls = input
|
|
.iter()
|
|
.enumerate()
|
|
.filter(|(_, item)| item.get("type").and_then(Value::as_str) == Some("function_call"))
|
|
.collect::<Vec<_>>();
|
|
|
|
let function_call_outputs = input
|
|
.iter()
|
|
.enumerate()
|
|
.filter(|(_, item)| {
|
|
item.get("type").and_then(Value::as_str) == Some("function_call_output")
|
|
})
|
|
.collect::<Vec<_>>();
|
|
|
|
assert_eq!(function_calls.len(), 3);
|
|
assert_eq!(function_call_outputs.len(), 3);
|
|
|
|
for (index, _) in &function_calls {
|
|
for (output_index, _) in &function_call_outputs {
|
|
assert!(
|
|
*index < *output_index,
|
|
"all function calls must come before outputs"
|
|
);
|
|
}
|
|
}
|
|
|
|
// output should come in the order of the function calls
|
|
let zipped = function_calls
|
|
.iter()
|
|
.zip(function_call_outputs.iter())
|
|
.collect::<Vec<_>>();
|
|
for (call, output) in zipped {
|
|
assert_eq!(
|
|
call.1.get("call_id").and_then(Value::as_str),
|
|
output.1.get("call_id").and_then(Value::as_str)
|
|
);
|
|
}
|
|
|
|
Ok(())
|
|
}
|