mirror of
https://github.com/openai/codex.git
synced 2026-04-23 22:24:57 +00:00
Remove the SDK alias/result layers so the wrapper surface returns canonical generated app-server models directly. - delete public type alias modules and regenerate v2_all.py against current schema - return InitializeResponse from metadata and generated Turn from run() - update docs, examples, notebook, and tests to use canonical generated models and repo-only text extraction helpers Validation: - PYTHONPATH=sdk/python/src python3 -m pytest sdk/python/tests - GH_TOKEN="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" RUN_REAL_CODEX_TESTS=1 PYTHONPATH=sdk/python/src python3 -m pytest sdk/python/tests -rs Co-authored-by: Codex <noreply@openai.com>
90 lines
2.6 KiB
Python
90 lines
2.6 KiB
Python
import sys
|
|
from pathlib import Path
|
|
|
|
_EXAMPLES_ROOT = Path(__file__).resolve().parents[1]
|
|
if str(_EXAMPLES_ROOT) not in sys.path:
|
|
sys.path.insert(0, str(_EXAMPLES_ROOT))
|
|
|
|
from _bootstrap import ensure_local_sdk_src, runtime_config
|
|
|
|
ensure_local_sdk_src()
|
|
|
|
from codex_app_server import (
|
|
Codex,
|
|
TextInput,
|
|
ThreadTokenUsageUpdatedNotification,
|
|
TurnCompletedNotification,
|
|
)
|
|
|
|
print("Codex mini CLI. Type /exit to quit.")
|
|
|
|
|
|
def _status_value(status: object | None) -> str:
|
|
return str(getattr(status, "value", status))
|
|
|
|
|
|
def _format_usage(usage: object | None) -> str:
|
|
if usage is None:
|
|
return "usage> (none)"
|
|
|
|
last = getattr(usage, "last", None)
|
|
total = getattr(usage, "total", None)
|
|
if last is None or total is None:
|
|
return f"usage> {usage}"
|
|
|
|
return (
|
|
"usage>\n"
|
|
f" last: input={last.inputTokens} output={last.outputTokens} reasoning={last.reasoningOutputTokens} total={last.totalTokens} cached={last.cachedInputTokens}\n"
|
|
f" total: input={total.inputTokens} output={total.outputTokens} reasoning={total.reasoningOutputTokens} total={total.totalTokens} cached={total.cachedInputTokens}"
|
|
)
|
|
|
|
|
|
with Codex(config=runtime_config()) as codex:
|
|
thread = codex.thread_start(model="gpt-5.4", config={"model_reasoning_effort": "high"})
|
|
print("Thread:", thread.id)
|
|
|
|
while True:
|
|
try:
|
|
user_input = input("you> ").strip()
|
|
except EOFError:
|
|
break
|
|
|
|
if not user_input:
|
|
continue
|
|
if user_input in {"/exit", "/quit"}:
|
|
break
|
|
|
|
turn = thread.turn(TextInput(user_input))
|
|
usage = None
|
|
status = None
|
|
error = None
|
|
printed_delta = False
|
|
|
|
print("assistant> ", end="", flush=True)
|
|
for event in turn.stream():
|
|
payload = event.payload
|
|
if event.method == "item/agentMessage/delta":
|
|
delta = getattr(payload, "delta", "")
|
|
if delta:
|
|
print(delta, end="", flush=True)
|
|
printed_delta = True
|
|
continue
|
|
if isinstance(payload, ThreadTokenUsageUpdatedNotification):
|
|
usage = payload.token_usage
|
|
continue
|
|
if isinstance(payload, TurnCompletedNotification):
|
|
status = payload.turn.status
|
|
error = payload.turn.error
|
|
|
|
if printed_delta:
|
|
print()
|
|
else:
|
|
print("[no text]")
|
|
|
|
status_text = _status_value(status)
|
|
print(f"assistant.status> {status_text}")
|
|
if status_text == "failed":
|
|
print("assistant.error>", error)
|
|
|
|
print(_format_usage(usage))
|