Why Truss?
- Write once, run anywhere: Package and test model code, weights, and dependencies with a model server that behaves the same in development and production.
- Fast developer loop: Implement your model with fast feedback from a live reload server, and skip Docker and Kubernetes configuration with a batteries-included model serving environment.
- Support for all Python frameworks: From
transformers
anddiffusers
toPyTorch
andTensorFlow
toTensorRT
andTriton
, Truss supports models created and served with any framework.
- 🦙 Llama 2 7B (13B) (70B)
- 🎨 Stable Diffusion XL
- 🗣 Whisper
Deploy your first model
Quickstart
Package, deploy, and invoke an ML model in production all in less than five minutes.
Truss tutorial
Learn model deployment step-by-step from “Hello, World!” to streaming output from an open-source LLM.