Jinkxy

New Member
Dec 18, 2018
8
4
Having alot of problems with that new koboldcpp local model in this version.
Responses take alot of time ( around 30s ) and alot of the times instead of a response i get a bullet list of character traits and suggestions what my answer should look like.

Switched back to the Gemma-2-Ataraxy-v4d-9B.i1-Q4_K_M.gguf from the previous local game version and it worked fine.
 

francisdrake

Member
Feb 21, 2019
107
101
Is this game set up for Vulkan/AMD on Linux? Tried doing local kobold instead of web tunneling to try and get faster responses but the log is all messed up and it just loads forever. Maybe I did the setup wrong but I Launched Kobold with my model, confirmed everything was working in the web ui, launched multiic, went to settings and picked local and looked up my model from the menu on the left, set a context value, then started the game. Then when I get to the bedroom description the log looks like this, tries to run CPU only with 4k context instead of using my GPU.

1749216233740.png
 

reidanota

Active Member
Nov 1, 2021
634
530
Local AI is working on my end, with the provided model. Sometimes the AI chat seems awkward if you go a little off-topic or into double-entendre ;) and the game pulls you back to the story, but the chat moments still add a little flavour, I'm enjoying it.
 

Ogre

Member
Mar 29, 2018
355
904
Game uses a 14B model, I'm not sure this is it but my RTX 3090 sometimes chokes on 12B models or even 7B when the prompt gets too long. Don't now how this game handles context and prompting but when I last played, a good few months back, it worked for a while but after a few interactions the llm stopped responding. I'm downloading now to check it, maybe you can set a different model in settings and use something lighter, sacrificing complexity in responses to a faster and perhaps more streamlined experience?
I also have a 3090. But I do not understand how to set a different model in settings. I seem to have only one option to choose. It's called "deep cognito something or other."

Is there a step by step set of instructions anywhere for how to get this started? This entire thread appears to be written by people who already understand how to do everything in AI. I'm a total noob and I don't know where to begin.

To quote someone a couple of posts earlier, I don't know how to correctly do anything like this guy did: " Launched Kobold with my model, confirmed everything was working in the web ui, launched multiic, went to settings and picked local and looked up my model from the menu on the left, set a context value, then started the game." That's not me. I just started the game and hoped for the best.
 
Last edited:

reidanota

Active Member
Nov 1, 2021
634
530
I also have a 3090. But I do not understand how to set a different model in settings. I seem to have only one option to choose. It's called "deep cognito something or other."

Is there a step by step set of instructions anywhere for how to get this started? This entire thread appears to be written by people who already understand how to do everything in AI. I'm a total noob and I don't know where to begin.

To quote someone a couple of posts earlier, I don't know how to correctly do anything like this guy did: " Launched Kobold with my model, confirmed everything was working in the web ui, launched multiic, went to settings and picked local and looked up my model from the menu on the left, set a context value, then started the game." That's not me. I just started the game and hoped for the best.
For local AI text generation the game uses koboldcpp, which is a back-end application that runs GGUF quantized models. You can download models from huggingface.com, such as


(these are not the latest, just two I had around - you can do a little research and find others that may suit you)
The default model is 14b, which is pretty high, and Q4_K_M which is a compromise between speed and quality. The higher the "b" and "q" numbers, the more complex the model will be, potentially giving higher quality and more nuanced responses but eating up more resources.

You should copy the model (.gguf file) that you download to the resources/koboldcpp folder in your game, and you'll be able to choose a different model at game start.

image_2025-06-06_223555656.png

This is supposed to be all you need to run the local model using the setup that the game provides. If you want fine tune things, you'll need to learn how to launch koboldcpp separately and configure it in the splash screen; then you keep it running in the background and copy the address they give you in the shell window (usually localhost) to the API address in the game's configuration.

I'm not an expert and also rely on wikis and installation instructions, if you play Skyrim and want to try LLM powered dialogue with NPCs you can check out this mod - - they have detailed instructions and videos on how to launch and configure koboldcpp (and a bunch of other tools) that would hopefully give you a better idea of what you're doing.

But, again, this game works out of the box, just click on the model in that pre-menu window and then let a few seconds for the model to load, avoid starting the game while it shows "loading model" on the top left

1749246194567.png
 
3.90 star(s) 16 Votes