Local LLMs – made easy
Run powerful large-language models entirely on your player’s machine—no servers, no cloud fees, no compliance headaches.
by Nobodywho
Price History +
NobodyWho wraps the battle-tested llama.cpp runtime in a clean Unity-ready API, giving you:
- Drop-in C# bindings – call chat, completion, or embedding functions from any script or add them as components directly in you scene.
- GPU-accelerated inference – Vulkan (Windows/Linux) & Metal (macOS) back-ends for blazing-fast token generation.
- Zero MLOps – every token is computed client-side, so you ship once and scale automatically.
- Tiny footprint – lean native library - you bring your own model.
- Open-source core – inspect, fork, and extend on GitHub.
Get started in minutes with the quick-start sample scene, or dive deeper in our docs.
Focus on the experience you’re building—NobodyWho handles the LLM heavy lifting.
Please help us by giving us a star on our Github