Skip to main content

Install

Install the Codex CLI:
npm install -g @openai/codex

Usage with Ollama

Codex requires a larger context window. It is recommended to use a context window of at least 32K tokens.
To use codex with Ollama, use the --oss flag:
codex --oss

Changing Models

By default, codex will use the local gpt-oss:20b model. However, you can specify a different model with the -m flag:
codex --oss -m gpt-oss:120b

Cloud Models

codex --oss -m gpt-oss:120b-cloud

Connecting to ollama.com

Create an API key from ollama.com and export it as OLLAMA_API_KEY. To use ollama.com directly, edit your ~/.codex/config.toml file to point to ollama.com.
model = "gpt-oss:120b"
model_provider = "ollama"

[model_providers.ollama]
name = "Ollama"
base_url = "https://ollama.com/v1"
env_key = "OLLAMA_API_KEY"
Run codex in a new terminal to load the new settings.
I