Skip to content
musegpt
Architecture
Providers
All Docs
GitHub
/
Esc
Home
/
Tags
/
gpu
gpu
2 documents categorized under gpu.
Page 1 of 1
llama.cpp
High-performance local inference engine with OpenAI-compatible server mode.
Provider
Backend
vLLM
High-throughput LLM serving engine with OpenAI-compatible API.
Provider
Backend
Compiled with SchemaFlux