Zero python dependency! Take full advantage of the GPUs. Write once, run anywhere. Get started with
Llama 2 series of models on your own device in 5 minutes.
Create an HTTP microservice for image classification. It runs YOLO and Mediapipe models at native GPU speed.
A cloud-native and edge-native WebAssembly Runtime
A WebAssembly runtime for dapr microservices.
Building Rust functions with WebAssembly
The easiest & fastest way to run customized and fine-tuned LLMs locally or on the edge.