mirror of
https://github.com/openai/codex.git
synced 2026-05-16 09:12:54 +00:00
770 lines
24 KiB
Rust
770 lines
24 KiB
Rust
use super::*;
|
|
use pretty_assertions::assert_eq;
|
|
use std::io::Write;
|
|
use std::path::Path;
|
|
use std::path::PathBuf;
|
|
use tempfile::tempdir;
|
|
use wiremock::Mock;
|
|
use wiremock::MockServer;
|
|
use wiremock::ResponseTemplate;
|
|
use wiremock::matchers::method;
|
|
use wiremock::matchers::path;
|
|
use zip::ZipWriter;
|
|
use zip::write::SimpleFileOptions;
|
|
|
|
const TEST_CURATED_PLUGIN_SHA: &str = "0123456789abcdef0123456789abcdef01234567";
|
|
|
|
fn write_file(path: &Path, contents: &str) {
|
|
std::fs::create_dir_all(path.parent().expect("file should have a parent")).unwrap();
|
|
std::fs::write(path, contents).unwrap();
|
|
}
|
|
|
|
fn write_curated_plugin(root: &Path, plugin_name: &str) {
|
|
let plugin_root = root.join("plugins").join(plugin_name);
|
|
write_file(
|
|
&plugin_root.join(".codex-plugin/plugin.json"),
|
|
&format!(r#"{{"name":"{plugin_name}"}}"#),
|
|
);
|
|
}
|
|
|
|
fn write_openai_curated_marketplace(root: &Path, plugin_names: &[&str]) {
|
|
let plugins = plugin_names
|
|
.iter()
|
|
.map(|plugin_name| {
|
|
format!(
|
|
r#"{{
|
|
"name": "{plugin_name}",
|
|
"source": {{
|
|
"source": "local",
|
|
"path": "./plugins/{plugin_name}"
|
|
}}
|
|
}}"#
|
|
)
|
|
})
|
|
.collect::<Vec<_>>()
|
|
.join(",\n");
|
|
write_file(
|
|
&root.join(".agents/plugins/marketplace.json"),
|
|
&format!(
|
|
r#"{{
|
|
"name": "openai-curated",
|
|
"plugins": [
|
|
{plugins}
|
|
]
|
|
}}"#
|
|
),
|
|
);
|
|
for plugin_name in plugin_names {
|
|
write_curated_plugin(root, plugin_name);
|
|
}
|
|
}
|
|
|
|
fn write_curated_plugin_sha(codex_home: &Path) {
|
|
write_file(
|
|
&codex_home.join(".tmp/plugins.sha"),
|
|
&format!("{TEST_CURATED_PLUGIN_SHA}\n"),
|
|
);
|
|
}
|
|
|
|
fn has_plugins_clone_dirs(codex_home: &Path) -> bool {
|
|
let Ok(entries) = std::fs::read_dir(codex_home.join(".tmp")) else {
|
|
return false;
|
|
};
|
|
|
|
entries.flatten().any(|entry| {
|
|
let path = entry.path();
|
|
path.is_dir()
|
|
&& path
|
|
.file_name()
|
|
.and_then(|name| name.to_str())
|
|
.is_some_and(|name| name.starts_with("plugins-clone-"))
|
|
})
|
|
}
|
|
|
|
#[cfg(unix)]
|
|
fn write_executable_script(path: &Path, contents: &str) {
|
|
#[cfg(unix)]
|
|
use std::os::unix::fs::PermissionsExt;
|
|
|
|
std::fs::write(path, contents).expect("write script");
|
|
#[cfg(unix)]
|
|
{
|
|
let mut permissions = std::fs::metadata(path).expect("metadata").permissions();
|
|
permissions.set_mode(0o755);
|
|
std::fs::set_permissions(path, permissions).expect("chmod");
|
|
}
|
|
}
|
|
|
|
async fn mount_github_repo_and_ref(server: &MockServer, sha: &str) {
|
|
Mock::given(method("GET"))
|
|
.and(path("/repos/openai/plugins"))
|
|
.respond_with(ResponseTemplate::new(200).set_body_string(r#"{"default_branch":"main"}"#))
|
|
.mount(server)
|
|
.await;
|
|
Mock::given(method("GET"))
|
|
.and(path("/repos/openai/plugins/git/ref/heads/main"))
|
|
.respond_with(
|
|
ResponseTemplate::new(200)
|
|
.set_body_string(format!(r#"{{"object":{{"sha":"{sha}"}}}}"#)),
|
|
)
|
|
.mount(server)
|
|
.await;
|
|
}
|
|
|
|
async fn mount_github_zipball(server: &MockServer, sha: &str, bytes: Vec<u8>) {
|
|
Mock::given(method("GET"))
|
|
.and(path(format!("/repos/openai/plugins/zipball/{sha}")))
|
|
.respond_with(
|
|
ResponseTemplate::new(200)
|
|
.insert_header("content-type", "application/zip")
|
|
.set_body_bytes(bytes),
|
|
)
|
|
.mount(server)
|
|
.await;
|
|
}
|
|
|
|
async fn mount_export_archive(server: &MockServer, bytes: Vec<u8>) -> String {
|
|
let export_api_url = format!("{}/backend-api/plugins/export/curated", server.uri());
|
|
Mock::given(method("GET"))
|
|
.and(path("/backend-api/plugins/export/curated"))
|
|
.respond_with(ResponseTemplate::new(200).set_body_string(format!(
|
|
r#"{{"download_url":"{}/files/curated-plugins.zip"}}"#,
|
|
server.uri()
|
|
)))
|
|
.mount(server)
|
|
.await;
|
|
Mock::given(method("GET"))
|
|
.and(path("/files/curated-plugins.zip"))
|
|
.respond_with(
|
|
ResponseTemplate::new(200)
|
|
.insert_header("content-type", "application/zip")
|
|
.set_body_bytes(bytes),
|
|
)
|
|
.mount(server)
|
|
.await;
|
|
export_api_url
|
|
}
|
|
|
|
async fn run_sync_with_transport_overrides(
|
|
codex_home: PathBuf,
|
|
git_binary: impl Into<String>,
|
|
api_base_url: impl Into<String>,
|
|
backup_archive_api_url: impl Into<String>,
|
|
) -> Result<String, String> {
|
|
let git_binary = git_binary.into();
|
|
let api_base_url = api_base_url.into();
|
|
let backup_archive_api_url = backup_archive_api_url.into();
|
|
tokio::task::spawn_blocking(move || {
|
|
sync_openai_plugins_repo_with_transport_overrides(
|
|
codex_home.as_path(),
|
|
&git_binary,
|
|
&api_base_url,
|
|
&backup_archive_api_url,
|
|
)
|
|
})
|
|
.await
|
|
.expect("sync task should join")
|
|
}
|
|
|
|
async fn run_http_sync(
|
|
codex_home: PathBuf,
|
|
api_base_url: impl Into<String>,
|
|
) -> Result<String, String> {
|
|
let api_base_url = api_base_url.into();
|
|
tokio::task::spawn_blocking(move || {
|
|
sync_openai_plugins_repo_via_http(codex_home.as_path(), &api_base_url)
|
|
})
|
|
.await
|
|
.expect("sync task should join")
|
|
}
|
|
|
|
fn assert_curated_gmail_repo(repo_path: &Path) {
|
|
assert!(repo_path.join(".agents/plugins/marketplace.json").is_file());
|
|
assert!(
|
|
repo_path
|
|
.join("plugins/gmail/.codex-plugin/plugin.json")
|
|
.is_file()
|
|
);
|
|
}
|
|
|
|
#[test]
|
|
fn curated_plugins_repo_path_uses_codex_home_tmp_dir() {
|
|
let tmp = tempdir().expect("tempdir");
|
|
assert_eq!(
|
|
curated_plugins_repo_path(tmp.path()),
|
|
tmp.path().join(".tmp/plugins")
|
|
);
|
|
}
|
|
|
|
#[test]
|
|
fn read_curated_plugins_sha_reads_trimmed_sha_file() {
|
|
let tmp = tempdir().expect("tempdir");
|
|
std::fs::create_dir_all(tmp.path().join(".tmp")).expect("create tmp");
|
|
std::fs::write(tmp.path().join(".tmp/plugins.sha"), "abc123\n").expect("write sha");
|
|
|
|
assert_eq!(
|
|
read_curated_plugins_sha(tmp.path()).as_deref(),
|
|
Some("abc123")
|
|
);
|
|
}
|
|
|
|
#[cfg(unix)]
|
|
#[test]
|
|
fn remove_stale_curated_repo_temp_dirs_removes_only_matching_directories() {
|
|
use std::os::unix::ffi::OsStrExt;
|
|
use std::time::SystemTime;
|
|
|
|
fn set_dir_mtime(path: &Path, age: Duration) -> Result<(), Box<dyn std::error::Error>> {
|
|
let now = SystemTime::now().duration_since(SystemTime::UNIX_EPOCH)?;
|
|
let modified_at = now.saturating_sub(age);
|
|
let tv_sec = i64::try_from(modified_at.as_secs())?;
|
|
let ts = libc::timespec { tv_sec, tv_nsec: 0 };
|
|
let times = [ts, ts];
|
|
let c_path = std::ffi::CString::new(path.as_os_str().as_bytes())?;
|
|
let result = unsafe { libc::utimensat(libc::AT_FDCWD, c_path.as_ptr(), times.as_ptr(), 0) };
|
|
if result != 0 {
|
|
return Err(std::io::Error::last_os_error().into());
|
|
}
|
|
Ok(())
|
|
}
|
|
|
|
let tmp = tempdir().expect("tempdir");
|
|
let parent = tmp.path().join(".tmp");
|
|
let stale_clone_dir = parent.join("plugins-clone-stale");
|
|
let fresh_clone_dir = parent.join("plugins-clone-fresh");
|
|
let unrelated_dir = parent.join("plugins-cache");
|
|
|
|
std::fs::create_dir_all(&stale_clone_dir).expect("create stale clone dir");
|
|
std::fs::create_dir_all(&fresh_clone_dir).expect("create fresh clone dir");
|
|
std::fs::create_dir_all(&unrelated_dir).expect("create unrelated dir");
|
|
set_dir_mtime(
|
|
&stale_clone_dir,
|
|
CURATED_PLUGINS_STALE_TEMP_DIR_MAX_AGE + Duration::from_secs(60),
|
|
)
|
|
.expect("age stale clone dir");
|
|
set_dir_mtime(&fresh_clone_dir, Duration::ZERO).expect("age fresh clone dir");
|
|
|
|
remove_stale_curated_repo_temp_dirs(&parent, CURATED_PLUGINS_STALE_TEMP_DIR_MAX_AGE);
|
|
|
|
assert!(!stale_clone_dir.exists());
|
|
assert!(fresh_clone_dir.is_dir());
|
|
assert!(unrelated_dir.is_dir());
|
|
}
|
|
|
|
#[cfg(unix)]
|
|
#[test]
|
|
fn sync_openai_plugins_repo_prefers_git_when_available() {
|
|
let tmp = tempdir().expect("tempdir");
|
|
let bin_dir = tempfile::Builder::new()
|
|
.prefix("fake-git-")
|
|
.tempdir()
|
|
.expect("tempdir");
|
|
let git_path = bin_dir.path().join("git");
|
|
let sha = "0123456789abcdef0123456789abcdef01234567";
|
|
|
|
write_executable_script(
|
|
&git_path,
|
|
&format!(
|
|
r#"#!/bin/sh
|
|
if [ "$1" = "ls-remote" ]; then
|
|
printf '%s\tHEAD\n' "{sha}"
|
|
exit 0
|
|
fi
|
|
if [ "$1" = "clone" ]; then
|
|
dest="$5"
|
|
mkdir -p "$dest/.git" "$dest/.agents/plugins" "$dest/plugins/gmail/.codex-plugin"
|
|
cat > "$dest/.agents/plugins/marketplace.json" <<'EOF'
|
|
{{"name":"openai-curated","plugins":[{{"name":"gmail","source":{{"source":"local","path":"./plugins/gmail"}}}}]}}
|
|
EOF
|
|
printf '%s\n' '{{"name":"gmail"}}' > "$dest/plugins/gmail/.codex-plugin/plugin.json"
|
|
exit 0
|
|
fi
|
|
if [ "$1" = "-C" ] && [ "$3" = "rev-parse" ] && [ "$4" = "HEAD" ]; then
|
|
printf '%s\n' "{sha}"
|
|
exit 0
|
|
fi
|
|
echo "unexpected git invocation: $@" >&2
|
|
exit 1
|
|
"#
|
|
),
|
|
);
|
|
|
|
let synced_sha = sync_openai_plugins_repo_with_transport_overrides(
|
|
tmp.path(),
|
|
git_path.to_str().expect("utf8 path"),
|
|
"http://127.0.0.1:9",
|
|
"http://127.0.0.1:9/backend-api/plugins/export/curated",
|
|
)
|
|
.expect("git sync should succeed");
|
|
|
|
assert_eq!(synced_sha, sha);
|
|
let repo_path = curated_plugins_repo_path(tmp.path());
|
|
assert!(repo_path.join(".git").is_dir());
|
|
assert_curated_gmail_repo(&repo_path);
|
|
assert_eq!(read_curated_plugins_sha(tmp.path()).as_deref(), Some(sha));
|
|
}
|
|
|
|
#[cfg(unix)]
|
|
#[test]
|
|
fn sync_openai_plugins_repo_via_git_succeeds_with_local_rewritten_remote() {
|
|
let tmp = tempdir().expect("tempdir");
|
|
let repo_root = tempfile::Builder::new()
|
|
.prefix("curated-repo-success-")
|
|
.tempdir()
|
|
.expect("tempdir");
|
|
let work_repo = repo_root.path().join("work/plugins");
|
|
let remote_repo = repo_root.path().join("remotes/openai/plugins.git");
|
|
std::fs::create_dir_all(work_repo.join(".agents/plugins")).expect("create marketplace dir");
|
|
std::fs::create_dir_all(work_repo.join("plugins/gmail/.codex-plugin"))
|
|
.expect("create plugin dir");
|
|
std::fs::write(
|
|
work_repo.join(".agents/plugins/marketplace.json"),
|
|
r#"{"name":"openai-curated","plugins":[{"name":"gmail","source":{"source":"local","path":"./plugins/gmail"}}]}"#,
|
|
)
|
|
.expect("write marketplace");
|
|
std::fs::write(
|
|
work_repo.join("plugins/gmail/.codex-plugin/plugin.json"),
|
|
r#"{"name":"gmail"}"#,
|
|
)
|
|
.expect("write plugin manifest");
|
|
|
|
let init_status = Command::new("git")
|
|
.arg("-C")
|
|
.arg(&work_repo)
|
|
.arg("init")
|
|
.status()
|
|
.expect("run git init");
|
|
assert!(init_status.success());
|
|
|
|
let add_status = Command::new("git")
|
|
.arg("-C")
|
|
.arg(&work_repo)
|
|
.arg("add")
|
|
.arg(".")
|
|
.status()
|
|
.expect("run git add");
|
|
assert!(add_status.success());
|
|
|
|
let commit_status = Command::new("git")
|
|
.arg("-C")
|
|
.arg(&work_repo)
|
|
.arg("-c")
|
|
.arg("user.name=Codex Test")
|
|
.arg("-c")
|
|
.arg("user.email=codex@example.com")
|
|
.arg("commit")
|
|
.arg("-m")
|
|
.arg("init")
|
|
.status()
|
|
.expect("run git commit");
|
|
assert!(commit_status.success());
|
|
|
|
std::fs::create_dir_all(remote_repo.parent().expect("remote parent"))
|
|
.expect("create remote parent");
|
|
let clone_status = Command::new("git")
|
|
.arg("clone")
|
|
.arg("--bare")
|
|
.arg(&work_repo)
|
|
.arg(&remote_repo)
|
|
.status()
|
|
.expect("run git clone --bare");
|
|
assert!(clone_status.success());
|
|
|
|
let sha_output = Command::new("git")
|
|
.arg("-C")
|
|
.arg(&work_repo)
|
|
.arg("rev-parse")
|
|
.arg("HEAD")
|
|
.output()
|
|
.expect("run git rev-parse");
|
|
assert!(sha_output.status.success());
|
|
let sha = String::from_utf8_lossy(&sha_output.stdout)
|
|
.trim()
|
|
.to_string();
|
|
|
|
let git_config_path = repo_root.path().join("git-rewrite.conf");
|
|
std::fs::write(
|
|
&git_config_path,
|
|
format!(
|
|
"[url \"file://{}/\"]\n insteadOf = https://github.com/\n",
|
|
repo_root.path().join("remotes").display()
|
|
),
|
|
)
|
|
.expect("write git config");
|
|
|
|
let bin_dir = tempfile::Builder::new()
|
|
.prefix("git-rewrite-wrapper-")
|
|
.tempdir()
|
|
.expect("tempdir");
|
|
let git_wrapper = bin_dir.path().join("git");
|
|
write_executable_script(
|
|
&git_wrapper,
|
|
&format!(
|
|
"#!/bin/sh\nGIT_CONFIG_GLOBAL='{}' exec git \"$@\"\n",
|
|
git_config_path.display()
|
|
),
|
|
);
|
|
|
|
let synced_sha =
|
|
sync_openai_plugins_repo_via_git(tmp.path(), git_wrapper.to_str().expect("utf8 path"))
|
|
.expect("git sync should succeed");
|
|
|
|
assert_eq!(synced_sha, sha);
|
|
assert_curated_gmail_repo(&curated_plugins_repo_path(tmp.path()));
|
|
assert_eq!(
|
|
read_curated_plugins_sha(tmp.path()).as_deref(),
|
|
Some(sha.as_str())
|
|
);
|
|
assert!(!has_plugins_clone_dirs(tmp.path()));
|
|
}
|
|
|
|
#[tokio::test]
|
|
async fn sync_openai_plugins_repo_falls_back_to_http_when_git_is_unavailable() {
|
|
let tmp = tempdir().expect("tempdir");
|
|
let server = MockServer::start().await;
|
|
let sha = "0123456789abcdef0123456789abcdef01234567";
|
|
|
|
mount_github_repo_and_ref(&server, sha).await;
|
|
mount_github_zipball(&server, sha, curated_repo_zipball_bytes(sha)).await;
|
|
|
|
let synced_sha = run_sync_with_transport_overrides(
|
|
tmp.path().to_path_buf(),
|
|
"missing-git-for-test",
|
|
server.uri(),
|
|
"http://127.0.0.1:9/backend-api/plugins/export/curated",
|
|
)
|
|
.await
|
|
.expect("fallback sync should succeed");
|
|
|
|
let repo_path = curated_plugins_repo_path(tmp.path());
|
|
assert_eq!(synced_sha, sha);
|
|
assert_curated_gmail_repo(&repo_path);
|
|
assert_eq!(read_curated_plugins_sha(tmp.path()).as_deref(), Some(sha));
|
|
}
|
|
|
|
#[cfg(unix)]
|
|
#[tokio::test]
|
|
async fn sync_openai_plugins_repo_falls_back_to_http_when_git_sync_fails() {
|
|
let tmp = tempdir().expect("tempdir");
|
|
let bin_dir = tempfile::Builder::new()
|
|
.prefix("fake-git-fail-")
|
|
.tempdir()
|
|
.expect("tempdir");
|
|
let git_path = bin_dir.path().join("git");
|
|
let sha = "0123456789abcdef0123456789abcdef01234567";
|
|
|
|
write_executable_script(
|
|
&git_path,
|
|
r#"#!/bin/sh
|
|
echo "simulated git failure" >&2
|
|
exit 1
|
|
"#,
|
|
);
|
|
|
|
let server = MockServer::start().await;
|
|
mount_github_repo_and_ref(&server, sha).await;
|
|
mount_github_zipball(&server, sha, curated_repo_zipball_bytes(sha)).await;
|
|
|
|
let synced_sha = run_sync_with_transport_overrides(
|
|
tmp.path().to_path_buf(),
|
|
git_path.to_str().expect("utf8 path"),
|
|
server.uri(),
|
|
"http://127.0.0.1:9/backend-api/plugins/export/curated",
|
|
)
|
|
.await
|
|
.expect("fallback sync should succeed");
|
|
|
|
let repo_path = curated_plugins_repo_path(tmp.path());
|
|
assert_eq!(synced_sha, sha);
|
|
assert_curated_gmail_repo(&repo_path);
|
|
assert_eq!(read_curated_plugins_sha(tmp.path()).as_deref(), Some(sha));
|
|
}
|
|
|
|
#[cfg(unix)]
|
|
#[test]
|
|
fn sync_openai_plugins_repo_via_git_cleans_up_staged_dir_on_clone_failure() {
|
|
let tmp = tempdir().expect("tempdir");
|
|
let bin_dir = tempfile::Builder::new()
|
|
.prefix("fake-git-partial-fail-")
|
|
.tempdir()
|
|
.expect("tempdir");
|
|
let git_path = bin_dir.path().join("git");
|
|
let sha = "0123456789abcdef0123456789abcdef01234567";
|
|
|
|
write_executable_script(
|
|
&git_path,
|
|
&format!(
|
|
r#"#!/bin/sh
|
|
if [ "$1" = "ls-remote" ]; then
|
|
printf '%s\tHEAD\n' "{sha}"
|
|
exit 0
|
|
fi
|
|
if [ "$1" = "clone" ]; then
|
|
dest="$5"
|
|
mkdir -p "$dest/.git"
|
|
echo "fatal: early EOF" >&2
|
|
exit 128
|
|
fi
|
|
echo "unexpected git invocation: $@" >&2
|
|
exit 1
|
|
"#
|
|
),
|
|
);
|
|
|
|
let err = sync_openai_plugins_repo_via_git(tmp.path(), git_path.to_str().expect("utf8 path"))
|
|
.expect_err("git sync should fail");
|
|
|
|
assert!(err.contains("fatal: early EOF"));
|
|
assert!(!has_plugins_clone_dirs(tmp.path()));
|
|
}
|
|
|
|
#[tokio::test]
|
|
async fn sync_openai_plugins_repo_via_http_cleans_up_staged_dir_on_extract_failure() {
|
|
let tmp = tempdir().expect("tempdir");
|
|
let server = MockServer::start().await;
|
|
let sha = "0123456789abcdef0123456789abcdef01234567";
|
|
|
|
mount_github_repo_and_ref(&server, sha).await;
|
|
mount_github_zipball(&server, sha, b"not a zip archive".to_vec()).await;
|
|
|
|
let err = run_http_sync(tmp.path().to_path_buf(), server.uri())
|
|
.await
|
|
.expect_err("http sync should fail");
|
|
|
|
assert!(err.contains("failed to open curated plugins zip archive"));
|
|
assert!(!has_plugins_clone_dirs(tmp.path()));
|
|
}
|
|
|
|
#[tokio::test]
|
|
async fn sync_openai_plugins_repo_skips_archive_download_when_sha_matches() {
|
|
let tmp = tempdir().expect("tempdir");
|
|
let repo_path = curated_plugins_repo_path(tmp.path());
|
|
std::fs::create_dir_all(repo_path.join(".agents/plugins")).expect("create repo");
|
|
std::fs::write(
|
|
repo_path.join(".agents/plugins/marketplace.json"),
|
|
r#"{"name":"openai-curated","plugins":[]}"#,
|
|
)
|
|
.expect("write marketplace");
|
|
std::fs::create_dir_all(tmp.path().join(".tmp")).expect("create tmp");
|
|
let sha = "fedcba9876543210fedcba9876543210fedcba98";
|
|
std::fs::write(tmp.path().join(".tmp/plugins.sha"), format!("{sha}\n")).expect("write sha");
|
|
|
|
let server = MockServer::start().await;
|
|
mount_github_repo_and_ref(&server, sha).await;
|
|
|
|
run_sync_with_transport_overrides(
|
|
tmp.path().to_path_buf(),
|
|
"missing-git-for-test",
|
|
server.uri(),
|
|
"http://127.0.0.1:9/backend-api/plugins/export/curated",
|
|
)
|
|
.await
|
|
.expect("sync should succeed");
|
|
|
|
assert_eq!(read_curated_plugins_sha(tmp.path()).as_deref(), Some(sha));
|
|
assert!(repo_path.join(".agents/plugins/marketplace.json").is_file());
|
|
}
|
|
|
|
#[tokio::test]
|
|
async fn sync_openai_plugins_repo_falls_back_to_export_archive_when_no_snapshot_exists() {
|
|
let tmp = tempdir().expect("tempdir");
|
|
let server = MockServer::start().await;
|
|
let export_sha = "1111111111111111111111111111111111111111";
|
|
|
|
Mock::given(method("GET"))
|
|
.and(path("/repos/openai/plugins"))
|
|
.respond_with(ResponseTemplate::new(500).set_body_string("github repo lookup failed"))
|
|
.mount(&server)
|
|
.await;
|
|
let export_api_url =
|
|
mount_export_archive(&server, curated_repo_backup_archive_zip_bytes(export_sha)).await;
|
|
|
|
let synced_sha = run_sync_with_transport_overrides(
|
|
tmp.path().to_path_buf(),
|
|
"missing-git-for-test",
|
|
server.uri(),
|
|
export_api_url,
|
|
)
|
|
.await
|
|
.expect("export fallback sync should succeed");
|
|
|
|
let repo_path = curated_plugins_repo_path(tmp.path());
|
|
assert_eq!(synced_sha, export_sha);
|
|
assert_curated_gmail_repo(&repo_path);
|
|
assert_eq!(
|
|
read_curated_plugins_sha(tmp.path()).as_deref(),
|
|
Some(export_sha)
|
|
);
|
|
}
|
|
|
|
#[tokio::test]
|
|
async fn sync_openai_plugins_repo_skips_export_archive_when_snapshot_exists() {
|
|
let tmp = tempdir().expect("tempdir");
|
|
let curated_root = curated_plugins_repo_path(tmp.path());
|
|
write_openai_curated_marketplace(&curated_root, &["linear"]);
|
|
write_curated_plugin_sha(tmp.path());
|
|
|
|
let plugin_manifest_path = curated_root.join("plugins/linear/.codex-plugin/plugin.json");
|
|
let original_manifest =
|
|
std::fs::read_to_string(&plugin_manifest_path).expect("read existing plugin manifest");
|
|
|
|
let server = MockServer::start().await;
|
|
|
|
Mock::given(method("GET"))
|
|
.and(path("/repos/openai/plugins"))
|
|
.respond_with(ResponseTemplate::new(500).set_body_string("github repo lookup failed"))
|
|
.mount(&server)
|
|
.await;
|
|
let export_api_url = mount_export_archive(
|
|
&server,
|
|
curated_repo_backup_archive_zip_bytes("2222222222222222222222222222222222222222"),
|
|
)
|
|
.await;
|
|
|
|
let err = run_sync_with_transport_overrides(
|
|
tmp.path().to_path_buf(),
|
|
"missing-git-for-test",
|
|
server.uri(),
|
|
export_api_url,
|
|
)
|
|
.await
|
|
.expect_err("existing snapshot should suppress export fallback");
|
|
|
|
assert!(err.contains("export archive fallback skipped"));
|
|
assert_eq!(
|
|
std::fs::read_to_string(&plugin_manifest_path).expect("read plugin manifest after sync"),
|
|
original_manifest
|
|
);
|
|
assert_eq!(
|
|
read_curated_plugins_sha(tmp.path()).as_deref(),
|
|
Some(TEST_CURATED_PLUGIN_SHA)
|
|
);
|
|
}
|
|
|
|
#[test]
|
|
fn read_extracted_backup_archive_git_sha_reads_head_ref_from_extracted_repo() {
|
|
let tmp = tempdir().expect("tempdir");
|
|
let git_dir = tmp.path().join(".git/refs/heads");
|
|
std::fs::create_dir_all(&git_dir).expect("create git ref dir");
|
|
std::fs::write(tmp.path().join(".git/HEAD"), "ref: refs/heads/main\n").expect("write HEAD");
|
|
std::fs::write(
|
|
git_dir.join("main"),
|
|
"3333333333333333333333333333333333333333\n",
|
|
)
|
|
.expect("write main ref");
|
|
|
|
assert_eq!(
|
|
read_extracted_backup_archive_git_sha(tmp.path())
|
|
.expect("read extracted backup archive git sha"),
|
|
Some("3333333333333333333333333333333333333333".to_string())
|
|
);
|
|
}
|
|
|
|
#[test]
|
|
fn read_extracted_backup_archive_git_sha_rejects_non_refs_head_target() {
|
|
let tmp = tempdir().expect("tempdir");
|
|
std::fs::create_dir_all(tmp.path().join(".git")).expect("create git dir");
|
|
std::fs::write(tmp.path().join(".git/HEAD"), "ref: HEAD\n").expect("write HEAD");
|
|
|
|
let err = read_extracted_backup_archive_git_sha(tmp.path())
|
|
.expect_err("non-refs target should be rejected");
|
|
|
|
assert!(err.contains("must stay under refs/"));
|
|
}
|
|
|
|
#[test]
|
|
fn read_extracted_backup_archive_git_sha_rejects_path_traversal_ref() {
|
|
let tmp = tempdir().expect("tempdir");
|
|
std::fs::create_dir_all(tmp.path().join(".git")).expect("create git dir");
|
|
std::fs::write(tmp.path().join(".git/HEAD"), "ref: refs/heads/../../evil\n")
|
|
.expect("write HEAD");
|
|
|
|
let err = read_extracted_backup_archive_git_sha(tmp.path())
|
|
.expect_err("path traversal ref should be rejected");
|
|
|
|
assert!(err.contains("invalid path components"));
|
|
}
|
|
|
|
fn curated_repo_zipball_bytes(sha: &str) -> Vec<u8> {
|
|
let cursor = std::io::Cursor::new(Vec::new());
|
|
let mut writer = ZipWriter::new(cursor);
|
|
let options = SimpleFileOptions::default();
|
|
let root = format!("openai-plugins-{sha}");
|
|
writer
|
|
.start_file(format!("{root}/.agents/plugins/marketplace.json"), options)
|
|
.expect("start marketplace entry");
|
|
writer
|
|
.write_all(
|
|
br#"{
|
|
"name": "openai-curated",
|
|
"plugins": [
|
|
{
|
|
"name": "gmail",
|
|
"source": {
|
|
"source": "local",
|
|
"path": "./plugins/gmail"
|
|
}
|
|
}
|
|
]
|
|
}"#,
|
|
)
|
|
.expect("write marketplace");
|
|
writer
|
|
.start_file(
|
|
format!("{root}/plugins/gmail/.codex-plugin/plugin.json"),
|
|
options,
|
|
)
|
|
.expect("start plugin manifest entry");
|
|
writer
|
|
.write_all(br#"{"name":"gmail"}"#)
|
|
.expect("write plugin manifest");
|
|
|
|
writer.finish().expect("finish zip writer").into_inner()
|
|
}
|
|
|
|
fn curated_repo_backup_archive_zip_bytes(sha: &str) -> Vec<u8> {
|
|
let cursor = std::io::Cursor::new(Vec::new());
|
|
let mut writer = ZipWriter::new(cursor);
|
|
let options = SimpleFileOptions::default();
|
|
|
|
writer
|
|
.start_file("plugins/.git/HEAD", options)
|
|
.expect("start HEAD entry");
|
|
writer
|
|
.write_all(b"ref: refs/heads/main\n")
|
|
.expect("write HEAD");
|
|
writer
|
|
.start_file("plugins/.git/refs/heads/main", options)
|
|
.expect("start main ref entry");
|
|
writer
|
|
.write_all(format!("{sha}\n").as_bytes())
|
|
.expect("write main ref");
|
|
writer
|
|
.start_file("plugins/.agents/plugins/marketplace.json", options)
|
|
.expect("start marketplace entry");
|
|
writer
|
|
.write_all(
|
|
br#"{
|
|
"name": "openai-curated",
|
|
"plugins": [
|
|
{
|
|
"name": "gmail",
|
|
"source": {
|
|
"source": "local",
|
|
"path": "./plugins/gmail"
|
|
}
|
|
}
|
|
]
|
|
}"#,
|
|
)
|
|
.expect("write marketplace");
|
|
writer
|
|
.start_file("plugins/plugins/gmail/.codex-plugin/plugin.json", options)
|
|
.expect("start plugin manifest entry");
|
|
writer
|
|
.write_all(br#"{"name":"gmail"}"#)
|
|
.expect("write plugin manifest");
|
|
|
|
writer.finish().expect("finish zip writer").into_inner()
|
|
}
|