Jinkxy

New Member
Dec 18, 2018
10
4
Having alot of problems with that new koboldcpp local model in this version.
Responses take alot of time ( around 30s ) and alot of the times instead of a response i get a bullet list of character traits and suggestions what my answer should look like.

Switched back to the Gemma-2-Ataraxy-v4d-9B.i1-Q4_K_M.gguf from the previous local game version and it worked fine.
 

francisdrake

Member
Feb 21, 2019
107
101
Is this game set up for Vulkan/AMD on Linux? Tried doing local kobold instead of web tunneling to try and get faster responses but the log is all messed up and it just loads forever. Maybe I did the setup wrong but I Launched Kobold with my model, confirmed everything was working in the web ui, launched multiic, went to settings and picked local and looked up my model from the menu on the left, set a context value, then started the game. Then when I get to the bedroom description the log looks like this, tries to run CPU only with 4k context instead of using my GPU.

1749216233740.png
 

reidanota

Active Member
Nov 1, 2021
636
540
Local AI is working on my end, with the provided model. Sometimes the AI chat seems awkward if you go a little off-topic or into double-entendre ;) and the game pulls you back to the story, but the chat moments still add a little flavour, I'm enjoying it.
 

Ogre

Member
Mar 29, 2018
356
904
Game uses a 14B model, I'm not sure this is it but my RTX 3090 sometimes chokes on 12B models or even 7B when the prompt gets too long. Don't now how this game handles context and prompting but when I last played, a good few months back, it worked for a while but after a few interactions the llm stopped responding. I'm downloading now to check it, maybe you can set a different model in settings and use something lighter, sacrificing complexity in responses to a faster and perhaps more streamlined experience?
I also have a 3090. But I do not understand how to set a different model in settings. I seem to have only one option to choose. It's called "deep cognito something or other."

Is there a step by step set of instructions anywhere for how to get this started? This entire thread appears to be written by people who already understand how to do everything in AI. I'm a total noob and I don't know where to begin.

To quote someone a couple of posts earlier, I don't know how to correctly do anything like this guy did: " Launched Kobold with my model, confirmed everything was working in the web ui, launched multiic, went to settings and picked local and looked up my model from the menu on the left, set a context value, then started the game." That's not me. I just started the game and hoped for the best.
 
Last edited:

reidanota

Active Member
Nov 1, 2021
636
540
I also have a 3090. But I do not understand how to set a different model in settings. I seem to have only one option to choose. It's called "deep cognito something or other."

Is there a step by step set of instructions anywhere for how to get this started? This entire thread appears to be written by people who already understand how to do everything in AI. I'm a total noob and I don't know where to begin.

To quote someone a couple of posts earlier, I don't know how to correctly do anything like this guy did: " Launched Kobold with my model, confirmed everything was working in the web ui, launched multiic, went to settings and picked local and looked up my model from the menu on the left, set a context value, then started the game." That's not me. I just started the game and hoped for the best.
For local AI text generation the game uses koboldcpp, which is a back-end application that runs GGUF quantized models. You can download models from huggingface.com, such as


(these are not the latest, just two I had around - you can do a little research and find others that may suit you)
The default model is 14b, which is pretty high, and Q4_K_M which is a compromise between speed and quality. The higher the "b" and "q" numbers, the more complex the model will be, potentially giving higher quality and more nuanced responses but eating up more resources.

You should copy the model (.gguf file) that you download to the resources/koboldcpp folder in your game, and you'll be able to choose a different model at game start.

image_2025-06-06_223555656.png

This is supposed to be all you need to run the local model using the setup that the game provides. If you want fine tune things, you'll need to learn how to launch koboldcpp separately and configure it in the splash screen; then you keep it running in the background and copy the address they give you in the shell window (usually localhost) to the API address in the game's configuration.

I'm not an expert and also rely on wikis and installation instructions, if you play Skyrim and want to try LLM powered dialogue with NPCs you can check out this mod - - they have detailed instructions and videos on how to launch and configure koboldcpp (and a bunch of other tools) that would hopefully give you a better idea of what you're doing.

But, again, this game works out of the box, just click on the model in that pre-menu window and then let a few seconds for the model to load, avoid starting the game while it shows "loading model" on the top left

1749246194567.png
 
  • Like
Reactions: Ogre

Jinkxy

New Member
Dec 18, 2018
10
4
Thanks for that reidanota, been trying alot but i cant get it to run with my own model. When i start koboldcpp on my own it instantly closes the cmd window as soon as i open multiic. And the multiic cmd window shows that he is always looking for his own koboldcpp.exe in his game path doesnt matter what i choose in the settings.

I can use different models by copying them into the game folder like you said but whatever i try i cant get it to run with my own koboldcpp at localhost:5001.
 

reidanota

Active Member
Nov 1, 2021
636
540
Thanks for that reidanota, been trying alot but i cant get it to run with my own model. When i start koboldcpp on my own it instantly closes the cmd window as soon as i open multiic. And the multiic cmd window shows that he is always looking for his own koboldcpp.exe in his game path doesnt matter what i choose in the settings.

I can use different models by copying them into the game folder like you said but whatever i try i cant get it to run with my own koboldcpp at localhost:5001.
I was able to make it work using remote tunnel

1749289116491.png
I opened multiic first, chose "use models via api" on the splash screen. Only then I loaded koboldcpp, chose the model and ticked "remote tunnel"
1749289227823.png
Why I tried that? No idea whatsoever, just think I saw that written womewhere. The console window gives you the address

1749289538158.png

I'm sure there's missing information here, I don't really know what I'm doing but the game seemed to interact with the LLM, but miss some functionality. For example, while using the local model from the game menu, the game's progression took over from the AI chat, often interrupting the conversation but keeping the story flowing. With a remote API configured like this, it felt like you had no limitations in your chat interaction, but also there wasn't an obvious way to end the conversation and move on with the story. The responses also made much less sense.

Perhaps someone can explain why, and give us better instructions. For now, using koboldcpp locally from the game works best for me overall
 

Jinkxy

New Member
Dec 18, 2018
10
4
Thanks, the tunnel was what made it work for me too.
But after trying it i will continue using the in game setting for now. Using the same text model the tunnel variant was alot slower for me.
 
Sep 21, 2019
104
196
Does that mean I can use any gguf model that I want to with the game?
y, u can use any model. on windows it’s easier - u just need to put the gguf file in the resources/koboldcpp/ folder inside the game.
the only thing is: sometimes, certain newer models might hallucinate with koboldcpp, but usually this can be fixed by updating kobold
 

Jogy1959

Member
Jan 2, 2018
156
69
Sorry but I don't run Apps or Games, that want to have the ownership over my server. I will look for the game, when it is really local.
 
3.90 star(s) 16 Votes