A step-by-step guide to connecting Claude Code, Anthropic's terminal-based coding agent, to a locally hosted model via Red Hat AI Inference Server (a downstream vLLM build) running on OpenShift. By overriding Claude Code's default Anthropic environment variables to point at an OpenShift route, all inference stays on-premises.

5m read timeFrom developers.redhat.com
Post cover image
Table of contents
PrerequisitesEnvironmentDeploy the Red Hat AI Inference ServerInstall and configure Claude CodeKey takeaways

Sort: