Our containerized AI artifacts are OCI compatible allowing you to directly use them with docker, podman and kubernetes wherever you need them: whether the cloud, a datacenter, or your basement. Our artifacts are regularly rebuilt, updated, and scanned for vulnerabilities to provide, the smallest, fastest, and most secure runtime possible.Documentation Index
Fetch the complete documentation index at: https://docs.ramalama.com/llms.txt
Use this file to discover all available pages before exploring further.
Quick start
The fastest path is to deploy a model image that bundles runtime + model using docker compose. For more information about deploying in production environments check out deployment.Install dependencies
Getting started requires either Docker or Podman. We also recommend the RamaLama CLI for a streamlined experience.
- Install Podman or Docker
- (Optional) Install RamaLama CLI
Create docker-compose.yaml
Create a
docker-compose.yaml using a model image which bundles both the runtime and model together into a single runnable container.
