Local LLMs – made easy


Run powerful large-language models entirely on your player’s machine—no servers, no cloud fees, no compliance headaches.


by Nobodywho


Price History +

NobodyWho wraps the battle-tested llama.cpp runtime in a clean Unity-ready API, giving you:

  • Drop-in C# bindings – call chat, completion, or embedding functions from any script or add them as components directly in you scene.
  • GPU-accelerated inference – Vulkan (Windows/Linux) & Metal (macOS) back-ends for blazing-fast token generation.
  • Zero MLOps – every token is computed client-side, so you ship once and scale automatically.
  • Tiny footprint – lean native library - you bring your own model.
  • Open-source core – inspect, fork, and extend on GitHub.

Get started in minutes with the quick-start sample scene, or dive deeper in our docs.


Focus on the experience you’re building—NobodyWho handles the LLM heavy lifting.


Please help us by giving us a star on our Github