llamafile allows AI developers to distribute and run LLMs with a single file. It supports multiple CPU microarchitectures and CPU architectures, as well as six operating systems. The weights for LLMs can be embedded within the llamafile for easier distribution.
Table of contents
Binary InstructionsSource Instructionszipalign documentationTechnical DetailsLicensingKnown IssuesSort: