Models

llama.cpp

A fast, portable inference stack for running open-weight language models across local machines, servers, and edge devices.

Category
Models
License
MIT
Updated
2026-04-10
Tags
inference, local-ai, open-weights