A step-by-step guide to connecting Claude Code, Anthropic's terminal-based coding agent, to a locally hosted model via Red Hat AI Inference Server (a downstream vLLM build) running on OpenShift. By overriding Claude Code's default Anthropic environment variables to point at an OpenShift route, all inference stays on-premises.

Table of contents
PrerequisitesEnvironmentDeploy the Red Hat AI Inference ServerInstall and configure Claude CodeKey takeawaysSort: