Your game talks to a plugin. The plugin talks to a runtime. The runtime uses the player's GPU.
The Tryll SDK ships first for Unreal Engine. Get early access before we open the waitlist.
We talked to hundreds of studios. They all want AI in their games. Three things stop them.
Cloud AI charges per token. At scale, a single game can rack up millions in API costs per month. Success becomes unsustainable.
With Tryll, tokens are free. Runs on player GPUs.
GDPR, the AI Act, export laws — sending player conversations to third-party servers creates a legal and compliance nightmare most studios won't touch.
With Tryll, personal data never leaves the device.
If the API goes down, your game breaks. If pricing changes, your margins vanish. Studios won't bet their product on someone else's uptime.
With Tryll, everything works fully offline.
Platform capabilities that ship with every Tryll-powered game
Connect your game's documentation and lore to an AI that answers player questions in real-time. Reduces support load, eliminates alt-tabbing.
NPCs that remember conversations, form opinions, and evolve over time. Build deep companion systems without thousands of dialogue branches.
Let AI trigger game mechanics — spawn enemies, modify environments, distribute rewards. Ship dynamic gameplay through API calls, not hardcoded logic.
Generate quests, lore, and world events based on player context. Infinite content without static databases.
Start free. Pay when your game succeeds.
On-device AI for your game
at $15 game price — scales with price
Book a DemoWant to see local AI in your game before integrating?
Download Tryll Assistant — a free AI overlay that runs on any game.
See how LLM inference works on player hardware, no integration needed.
Everything you need to know about the Tryll platform
8GB of VRAM is enough to run both a model and a game. However, more VRAM is better - graphics-intensive games at high settings may consume most of the memory, which can slow down the model. For the best experience, we recommend 12GB+ VRAM. Tryll automatically selects the most efficient model and quantization for the user's system to ensure smooth gameplay.
According to the Steam Hardware Survey (see VRAM section), about 2/3 of players already have 8GB+ VRAM cards, which are good enough to run local AI:
As open-source models get smaller and faster, and gamers continue upgrading hardware, this percentage grows - unlocking scalable deployment without cloud reliance.
Local AI advantages:
Cloud-based AI adds legal and UX risks:
Tryll's local-first approach is not just cheaper - it's safer, faster, and scalable to millions of players.
This is the core problem Tryll solves — like how DirectX abstracts GPU differences for graphics. Tryll's Model Manager automatically:
You write one integration. The platform handles every hardware configuration. No setup required from players or developers.
Tryll is designed so that AI is always additive to your game. AI inference runs separately from core gameplay, so if anything goes wrong, the game falls back to default behavior. We're building toward full process isolation, but even today the architecture ensures AI issues don't take down your game.
We're building Unreal and Unity plugins first, with an SDK API to follow. The platform is engine-agnostic at its core — the engine-specific plugins are convenience layers on top of the same underlying runtime. If you're working with a custom engine, reach out and we'll work with you on integration.
Still have questions?
Contact us on DiscordTalk to our team about integrating Tryll into your game.
Contact us at:
team@tryllengine.com