Available Distributionsο
Llama Stack provides several pre-configured distributions to help you get started quickly. Choose the distribution that best fits your hardware and use case.
Quick Referenceο
Distribution |
Use Case |
Hardware Requirements |
Provider |
---|---|---|---|
|
General purpose, prototyping |
Any (CPU/GPU) |
Ollama, Remote APIs |
|
High-performance inference |
GPU required |
Local GPU inference |
Remote-hosted |
Production, managed service |
None |
Partner providers |
iOS/Android SDK |
Mobile applications |
Mobile device |
On-device inference |
Choose Your Distributionο
π Getting Started (Recommended for Beginners)ο
Use distribution-starter
if you want to:
Prototype quickly without GPU requirements
Use remote inference providers (Fireworks, Together, vLLM etc.)
Run locally with Ollama for development
docker pull llama-stack/distribution-starter
Guides: Starter Distribution Guide
π₯οΈ Self-Hosted with GPUο
Use distribution-meta-reference-gpu
if you:
Have access to GPU hardware
Want maximum performance and control
Need to run inference locally
docker pull llama-stack/distribution-meta-reference-gpu
Guides: Meta Reference GPU Guide
π₯οΈ Self-Hosted with NVIDA NeMo Microservicesο
Use nvidia
if you:
Want to use Llama Stack with NVIDIA NeMo Microservices
Guides: NVIDIA Distribution Guide
βοΈ Managed Hostingο
Use remote-hosted endpoints if you:
Donβt want to manage infrastructure
Need production-ready reliability
Prefer managed services
Partners: Fireworks.ai and Together.xyz
Guides: Remote-Hosted Endpoints
π± Mobile Developmentο
Use mobile SDKs if you:
Are building iOS or Android applications
Need on-device inference capabilities
Want offline functionality
π§ Custom Solutionsο
Build your own distribution if:
None of the above fit your specific needs
You need custom configurations
You want to optimize for your specific use case
Guides: Building Custom Distributions
Detailed Documentationο
Self-Hosted Distributionsο
Remote-Hosted Solutionsο
Mobile SDKsο
Decision Flowο
graph TD
A[What's your use case?] --> B{Need mobile app?}
B -->|Yes| C[Use Mobile SDKs]
B -->|No| D{Have GPU hardware?}
D -->|Yes| E[Use Meta Reference GPU]
D -->|No| F{Want managed hosting?}
F -->|Yes| G[Use Remote-Hosted]
F -->|No| H[Use Starter Distribution]
Next Stepsο
Choose your distribution from the options above
Follow the setup guide for your selected distribution
Configure your providers with API keys or local models
Start building with Llama Stack!
For help choosing or troubleshooting, check our Getting Started Guide or Community Support.