mirror of
https://github.com/openai/codex.git
synced 2026-04-30 01:16:54 +00:00
use rama instead of implementing our own proxy stack
This commit is contained in:
@@ -30,6 +30,7 @@ clap = { workspace = true, features = ["derive"] }
|
||||
codex-core = { workspace = true }
|
||||
globset = { workspace = true }
|
||||
hyper = { version = "0.14", features = ["full"] }
|
||||
rcgen-rama = { package = "rcgen", version = "0.14", default-features = false, features = ["pem", "x509-parser", "ring"] }
|
||||
rcgen = { version = "0.13", features = ["pem", "x509-parser"], optional = true }
|
||||
rustls = { version = "0.21", optional = true }
|
||||
rustls-native-certs = { version = "0.6", optional = true }
|
||||
@@ -41,3 +42,4 @@ tokio-rustls = { version = "0.24", optional = true }
|
||||
toml = { workspace = true }
|
||||
tracing = { workspace = true }
|
||||
tracing-subscriber = { workspace = true, features = ["fmt"] }
|
||||
rama = { version = "0.3.0-alpha.3", default-features = false, features = ["http-full", "proxy-full", "socks5", "rustls"] }
|
||||
|
||||
Reference in New Issue
Block a user