Quickstart (CPU)¶
Get Codex running locally in under five minutes — no GPU required.
Prerequisites¶
1 — Start the API¶
# Option A: directly
python -m codex.cli serve
# Option B: Docker Compose (recommended for isolation)
docker compose up api
The API listens on http://localhost:8000 by default.
2 — Run inference¶
curl -X POST http://localhost:8000/infer \
-H "Content-Type: application/json" \
-d '{"prompt": "def fibonacci(n):", "max_tokens": 64}'
Or open the demo notebook:
3 — Verify outputs¶
The response JSON contains completion, tokens_used, and model_version.
No GPU is required — the symbolic pipeline runs fully on CPU.
Next steps¶
| Goal | Guide |
|---|---|
| Full end-to-end training | End-to-End CPU tutorial |
| Deploy to production | Deploy Pipeline |
| Configure the model | Codex Setup Guide |
| Understand the architecture | Architecture |