mirror of
https://github.com/openai/codex.git
synced 2026-04-26 23:55:25 +00:00
## Why `argument-comment-lint` was green in CI even though the repo still had many uncommented literal arguments. The main gap was target coverage: the repo wrapper did not force Cargo to inspect test-only call sites, so examples like the `latest_session_lookup_params(true, ...)` tests in `codex-rs/tui_app_server/src/lib.rs` never entered the blocking CI path. This change cleans up the existing backlog, makes the default repo lint path cover all Cargo targets, and starts rolling that stricter CI enforcement out on the platform where it is currently validated. ## What changed - mechanically fixed existing `argument-comment-lint` violations across the `codex-rs` workspace, including tests, examples, and benches - updated `tools/argument-comment-lint/run-prebuilt-linter.sh` and `tools/argument-comment-lint/run.sh` so non-`--fix` runs default to `--all-targets` unless the caller explicitly narrows the target set - fixed both wrappers so forwarded cargo arguments after `--` are preserved with a single separator - documented the new default behavior in `tools/argument-comment-lint/README.md` - updated `rust-ci` so the macOS lint lane keeps the plain wrapper invocation and therefore enforces `--all-targets`, while Linux and Windows temporarily pass `-- --lib --bins` That temporary CI split keeps the stricter all-targets check where it is already cleaned up, while leaving room to finish the remaining Linux- and Windows-specific target-gated cleanup before enabling `--all-targets` on those runners. The Linux and Windows failures on the intermediate revision were caused by the wrapper forwarding bug, not by additional lint findings in those lanes. ## Validation - `bash -n tools/argument-comment-lint/run.sh` - `bash -n tools/argument-comment-lint/run-prebuilt-linter.sh` - shell-level wrapper forwarding check for `-- --lib --bins` - shell-level wrapper forwarding check for `-- --tests` - `just argument-comment-lint` - `cargo test` in `tools/argument-comment-lint` - `cargo test -p codex-terminal-detection` ## Follow-up - Clean up remaining Linux-only target-gated callsites, then switch the Linux lint lane back to the plain wrapper invocation. - Clean up remaining Windows-only target-gated callsites, then switch the Windows lint lane back to the plain wrapper invocation.
290 lines
10 KiB
Rust
290 lines
10 KiB
Rust
use anyhow::Result;
|
|
use anyhow::anyhow;
|
|
use anyhow::bail;
|
|
use app_test_support::McpProcess;
|
|
use app_test_support::create_mock_responses_server_sequence_unchecked;
|
|
use app_test_support::to_response;
|
|
use codex_app_server_protocol::ItemCompletedNotification;
|
|
use codex_app_server_protocol::ItemStartedNotification;
|
|
use codex_app_server_protocol::JSONRPCMessage;
|
|
use codex_app_server_protocol::JSONRPCResponse;
|
|
use codex_app_server_protocol::PlanDeltaNotification;
|
|
use codex_app_server_protocol::RequestId;
|
|
use codex_app_server_protocol::ThreadItem;
|
|
use codex_app_server_protocol::ThreadStartParams;
|
|
use codex_app_server_protocol::ThreadStartResponse;
|
|
use codex_app_server_protocol::TurnCompletedNotification;
|
|
use codex_app_server_protocol::TurnStartParams;
|
|
use codex_app_server_protocol::TurnStartResponse;
|
|
use codex_app_server_protocol::TurnStatus;
|
|
use codex_app_server_protocol::UserInput as V2UserInput;
|
|
use codex_features::FEATURES;
|
|
use codex_features::Feature;
|
|
use codex_protocol::config_types::CollaborationMode;
|
|
use codex_protocol::config_types::ModeKind;
|
|
use codex_protocol::config_types::Settings;
|
|
use core_test_support::responses;
|
|
use core_test_support::skip_if_no_network;
|
|
use pretty_assertions::assert_eq;
|
|
use std::collections::BTreeMap;
|
|
use std::path::Path;
|
|
use tempfile::TempDir;
|
|
use tokio::time::sleep;
|
|
use tokio::time::timeout;
|
|
use wiremock::MockServer;
|
|
|
|
const DEFAULT_READ_TIMEOUT: std::time::Duration = std::time::Duration::from_secs(10);
|
|
|
|
#[tokio::test(flavor = "multi_thread", worker_threads = 4)]
|
|
async fn plan_mode_uses_proposed_plan_block_for_plan_item() -> Result<()> {
|
|
skip_if_no_network!(Ok(()));
|
|
|
|
let plan_block = "<proposed_plan>\n# Final plan\n- first\n- second\n</proposed_plan>\n";
|
|
let full_message = format!("Preface\n{plan_block}Postscript");
|
|
let responses = vec![responses::sse(vec![
|
|
responses::ev_response_created("resp-1"),
|
|
responses::ev_message_item_added("msg-1", ""),
|
|
responses::ev_output_text_delta(&full_message),
|
|
responses::ev_assistant_message("msg-1", &full_message),
|
|
responses::ev_completed("resp-1"),
|
|
])];
|
|
let server = create_mock_responses_server_sequence_unchecked(responses).await;
|
|
|
|
let codex_home = TempDir::new()?;
|
|
create_config_toml(codex_home.path(), &server.uri())?;
|
|
|
|
let mut mcp = McpProcess::new(codex_home.path()).await?;
|
|
timeout(DEFAULT_READ_TIMEOUT, mcp.initialize()).await??;
|
|
|
|
let turn = start_plan_mode_turn(&mut mcp).await?;
|
|
let (_, completed_items, plan_deltas, turn_completed) =
|
|
collect_turn_notifications(&mut mcp).await?;
|
|
wait_for_responses_request_count(&server, /*expected_count*/ 1).await?;
|
|
|
|
assert_eq!(turn_completed.turn.id, turn.id);
|
|
assert_eq!(turn_completed.turn.status, TurnStatus::Completed);
|
|
|
|
let expected_plan = ThreadItem::Plan {
|
|
id: format!("{}-plan", turn.id),
|
|
text: "# Final plan\n- first\n- second\n".to_string(),
|
|
};
|
|
let expected_plan_id = format!("{}-plan", turn.id);
|
|
let streamed_plan = plan_deltas
|
|
.iter()
|
|
.map(|delta| delta.delta.as_str())
|
|
.collect::<String>();
|
|
assert_eq!(streamed_plan, "# Final plan\n- first\n- second\n");
|
|
assert!(
|
|
plan_deltas
|
|
.iter()
|
|
.all(|delta| delta.item_id == expected_plan_id)
|
|
);
|
|
let plan_items = completed_items
|
|
.iter()
|
|
.filter_map(|item| match item {
|
|
ThreadItem::Plan { .. } => Some(item.clone()),
|
|
_ => None,
|
|
})
|
|
.collect::<Vec<_>>();
|
|
assert_eq!(plan_items, vec![expected_plan]);
|
|
assert!(
|
|
completed_items
|
|
.iter()
|
|
.any(|item| matches!(item, ThreadItem::AgentMessage { .. })),
|
|
"agent message items should still be emitted alongside the plan item"
|
|
);
|
|
|
|
Ok(())
|
|
}
|
|
|
|
#[tokio::test(flavor = "multi_thread", worker_threads = 4)]
|
|
async fn plan_mode_without_proposed_plan_does_not_emit_plan_item() -> Result<()> {
|
|
skip_if_no_network!(Ok(()));
|
|
|
|
let responses = vec![responses::sse(vec![
|
|
responses::ev_response_created("resp-1"),
|
|
responses::ev_assistant_message("msg-1", "Done"),
|
|
responses::ev_completed("resp-1"),
|
|
])];
|
|
let server = create_mock_responses_server_sequence_unchecked(responses).await;
|
|
|
|
let codex_home = TempDir::new()?;
|
|
create_config_toml(codex_home.path(), &server.uri())?;
|
|
|
|
let mut mcp = McpProcess::new(codex_home.path()).await?;
|
|
timeout(DEFAULT_READ_TIMEOUT, mcp.initialize()).await??;
|
|
|
|
let _turn = start_plan_mode_turn(&mut mcp).await?;
|
|
let (_, completed_items, plan_deltas, _) = collect_turn_notifications(&mut mcp).await?;
|
|
wait_for_responses_request_count(&server, /*expected_count*/ 1).await?;
|
|
|
|
let has_plan_item = completed_items
|
|
.iter()
|
|
.any(|item| matches!(item, ThreadItem::Plan { .. }));
|
|
assert!(!has_plan_item);
|
|
assert!(plan_deltas.is_empty());
|
|
|
|
Ok(())
|
|
}
|
|
|
|
async fn start_plan_mode_turn(mcp: &mut McpProcess) -> Result<codex_app_server_protocol::Turn> {
|
|
let thread_req = mcp
|
|
.send_thread_start_request(ThreadStartParams {
|
|
model: Some("mock-model".to_string()),
|
|
..Default::default()
|
|
})
|
|
.await?;
|
|
let thread_resp: JSONRPCResponse = timeout(
|
|
DEFAULT_READ_TIMEOUT,
|
|
mcp.read_stream_until_response_message(RequestId::Integer(thread_req)),
|
|
)
|
|
.await??;
|
|
let thread = to_response::<ThreadStartResponse>(thread_resp)?.thread;
|
|
|
|
let collaboration_mode = CollaborationMode {
|
|
mode: ModeKind::Plan,
|
|
settings: Settings {
|
|
model: "mock-model".to_string(),
|
|
reasoning_effort: None,
|
|
developer_instructions: None,
|
|
},
|
|
};
|
|
let turn_req = mcp
|
|
.send_turn_start_request(TurnStartParams {
|
|
thread_id: thread.id,
|
|
input: vec![V2UserInput::Text {
|
|
text: "Plan this".to_string(),
|
|
text_elements: Vec::new(),
|
|
}],
|
|
collaboration_mode: Some(collaboration_mode),
|
|
..Default::default()
|
|
})
|
|
.await?;
|
|
let turn_resp: JSONRPCResponse = timeout(
|
|
DEFAULT_READ_TIMEOUT,
|
|
mcp.read_stream_until_response_message(RequestId::Integer(turn_req)),
|
|
)
|
|
.await??;
|
|
Ok(to_response::<TurnStartResponse>(turn_resp)?.turn)
|
|
}
|
|
|
|
async fn collect_turn_notifications(
|
|
mcp: &mut McpProcess,
|
|
) -> Result<(
|
|
Vec<ThreadItem>,
|
|
Vec<ThreadItem>,
|
|
Vec<PlanDeltaNotification>,
|
|
TurnCompletedNotification,
|
|
)> {
|
|
let mut started_items = Vec::new();
|
|
let mut completed_items = Vec::new();
|
|
let mut plan_deltas = Vec::new();
|
|
|
|
loop {
|
|
let message = timeout(DEFAULT_READ_TIMEOUT, mcp.read_next_message()).await??;
|
|
let JSONRPCMessage::Notification(notification) = message else {
|
|
continue;
|
|
};
|
|
match notification.method.as_str() {
|
|
"item/started" => {
|
|
let params = notification
|
|
.params
|
|
.ok_or_else(|| anyhow!("item/started notifications must include params"))?;
|
|
let payload: ItemStartedNotification = serde_json::from_value(params)?;
|
|
started_items.push(payload.item);
|
|
}
|
|
"item/completed" => {
|
|
let params = notification
|
|
.params
|
|
.ok_or_else(|| anyhow!("item/completed notifications must include params"))?;
|
|
let payload: ItemCompletedNotification = serde_json::from_value(params)?;
|
|
completed_items.push(payload.item);
|
|
}
|
|
"item/plan/delta" => {
|
|
let params = notification
|
|
.params
|
|
.ok_or_else(|| anyhow!("item/plan/delta notifications must include params"))?;
|
|
let payload: PlanDeltaNotification = serde_json::from_value(params)?;
|
|
plan_deltas.push(payload);
|
|
}
|
|
"turn/completed" => {
|
|
let params = notification
|
|
.params
|
|
.ok_or_else(|| anyhow!("turn/completed notifications must include params"))?;
|
|
let payload: TurnCompletedNotification = serde_json::from_value(params)?;
|
|
return Ok((started_items, completed_items, plan_deltas, payload));
|
|
}
|
|
_ => {}
|
|
}
|
|
}
|
|
}
|
|
|
|
async fn wait_for_responses_request_count(
|
|
server: &MockServer,
|
|
expected_count: usize,
|
|
) -> Result<()> {
|
|
timeout(DEFAULT_READ_TIMEOUT, async {
|
|
loop {
|
|
let Some(requests) = server.received_requests().await else {
|
|
bail!("wiremock did not record requests");
|
|
};
|
|
let responses_request_count = requests
|
|
.iter()
|
|
.filter(|request| {
|
|
request.method == "POST" && request.url.path().ends_with("/responses")
|
|
})
|
|
.count();
|
|
if responses_request_count == expected_count {
|
|
return Ok::<(), anyhow::Error>(());
|
|
}
|
|
if responses_request_count > expected_count {
|
|
bail!(
|
|
"expected exactly {expected_count} /responses requests, got {responses_request_count}"
|
|
);
|
|
}
|
|
sleep(std::time::Duration::from_millis(10)).await;
|
|
}
|
|
})
|
|
.await??;
|
|
Ok(())
|
|
}
|
|
|
|
fn create_config_toml(codex_home: &Path, server_uri: &str) -> std::io::Result<()> {
|
|
let features = BTreeMap::from([(Feature::CollaborationModes, true)]);
|
|
let feature_entries = features
|
|
.into_iter()
|
|
.map(|(feature, enabled)| {
|
|
let key = FEATURES
|
|
.iter()
|
|
.find(|spec| spec.id == feature)
|
|
.map(|spec| spec.key)
|
|
.unwrap_or_else(|| panic!("missing feature key for {feature:?}"));
|
|
format!("{key} = {enabled}")
|
|
})
|
|
.collect::<Vec<_>>()
|
|
.join("\n");
|
|
let config_toml = codex_home.join("config.toml");
|
|
std::fs::write(
|
|
config_toml,
|
|
format!(
|
|
r#"
|
|
model = "mock-model"
|
|
approval_policy = "never"
|
|
sandbox_mode = "read-only"
|
|
|
|
model_provider = "mock_provider"
|
|
|
|
[features]
|
|
{feature_entries}
|
|
|
|
[model_providers.mock_provider]
|
|
name = "Mock provider for test"
|
|
base_url = "{server_uri}/v1"
|
|
wire_api = "responses"
|
|
request_max_retries = 0
|
|
stream_max_retries = 0
|
|
"#
|
|
),
|
|
)
|
|
}
|