CREATE YOUR AI CUM SLUT ON CANDY.AI TRY FOR FREE
x
Nov 28, 2019
106
32
The game is brilliant, thank you very much to the author)
Game question: how to become dominant or at least control in sex... it's a pity that there is no dialog with AI during sex, between positions, would be in general very cool)
I had this problem as well, since even with cheats the "dominance" statistic in the relationship doesn't seem to control who is the "Alpha" in sex. I've also found that allowing the NPC to choose actions multiple times will end up in them becoming the alpha. It may be controlled by the dominance stat in fetishes instead of the relationship stats, but I haven't sufficiently experimented to answer.
 

Dnds

Out to Play
Donor
Jul 5, 2017
319
406
I managed to install Ollama and make it work (including in game), but how I do make it load the model recommended by the dev for playing?
 
  • Like
Reactions: Ssato243

pasha823

New Member
Nov 6, 2019
9
3
Question for AI connoisseurs: is it possible to make the AI give an answer with translation or duplicate it in the browser in which kobold is open?
 

kkai

Member
Dec 21, 2016
181
315
I managed to install Ollama and make it work (including in game), but how I do make it load the model recommended by the dev for playing?
Ollama is a pain in the ass and can't use quant files like GGUFs direct from huggingface (you have to create what are called modelfiles from them, tutorials abound for this stupid requirement, or pull the models directly from ollama's library). KoboldCPP is a much superior option, in my opinion.
 

kkai

Member
Dec 21, 2016
181
315
Thanks! I got the Q8 but seems like my 8GB Vram is struggling lol.. Do you have recommended settings for more of a spicy output or do I gotta lead them in that direction? Never used a local ai like this before...
Q8s are overkill. Generally Q6 is considered near perfect. I only do Q8 when I have VRAM to waste (I have 24GB so I use Q8s for 12Bs.)
Don't be afraid of quants down to, say, Q4 with a 8-12B, but lower than that and it's going to get noticeable. I use Q4_K_S IIRC for 32Bs which can just fit in 24GB, though the larger the parameter count (B = billions of parameters) of the model, typically the better it can handle quantization.
 

forgotp

Member
Nov 4, 2019
418
434
Q8s are overkill. Generally Q6 is considered near perfect. I only do Q8 when I have VRAM to waste (I have 24GB so I use Q8s for 12Bs.)
Don't be afraid of quants down to, say, Q4 with a 8-12B, but lower than that and it's going to get noticeable. I use Q4_K_S IIRC for 32Bs which can just fit in 24GB, though the larger the parameter count (B = billions of parameters) of the model, typically the better it can handle quantization.
I see, thanks. I'll try switching to Q6 and then 4 if its still slow :WeSmart:
 

Dnds

Out to Play
Donor
Jul 5, 2017
319
406
Ollama is a pain in the ass and can't use quant files like GGUFs direct from huggingface (you have to create what are called modelfiles from them, tutorials abound for this stupid requirement, or pull the models directly from ollama's library). KoboldCPP is a much superior option, in my opinion.
I tried Kobold but it has a similar problem I'm running into right now, can't figure out how to save models somewhere else since it installs on C directory and mine doesn't have that much memory left. Trying to figure out how to change directories atm.
 

kkai

Member
Dec 21, 2016
181
315
I tried Kobold but it has a similar problem I'm running into right now, can't figure out how to save models somewhere else since it installs on C directory and mine doesn't have that much memory left. Trying to figure out how to change directories atm.
KoboldCPP is just a single executable with no installer. You can put it anywhere you want and point it to a GGUF anywhere it's kept. Be sure you aren't using KoboldAI which is a related project, but not the one you want.
 
Last edited:
  • Like
Reactions: Dnds

Dnds

Out to Play
Donor
Jul 5, 2017
319
406
KoboldCPP is just a single executable with no installer. You can put it anywhere you want and point it to a GGUF anywhere it's kept. Be sure you aren't using KoboldAI which is a related project, but not the one you want.
No, I didn't know that, I'm very glad you mentioned it, I was using the wrong one as you have guessed. Currently testing the CPP but I think I got it right this time. Thank you for sticking around and helping people troubleshoot stuff, users like you are the real MVP.

Edit: It worked, nice and simple. Its a big laggy but acceptable.
 
Last edited:
  • Like
Reactions: kkai

Strec

Active Member
Feb 20, 2018
609
391
You will be able to blackmail somebody, but never force it. You know... The rules, Patreon... I don't wanna touch it...
So to resume the game will never have non-consensual content and will always remain a carebearing dating sim, sorry for such a game the could have been a great one...

One more game developpement killed by patreon system :rolleyes:
 

♂Brout&Minou♀

Engaged Member
Jul 30, 2020
2,511
1,833
It seems that this game still has some way to go before it's fully playable, but it's already made me laugh a lot. When you ask the headmaster about the job, you can ask him about his menstruation cycle... It's just that mine was a guy. Apart from that, I don't know, maybe I've missed something, but during the ‘construction’ of the characters... Eg: I'm a guy and I'm married. I add mum, dads and a sister when I add people, they all seem to live under the same roof as my mc. Then I want to add my wife's family, so say yes to ‘add customization family’, so I create my wife's family, but I can't link them in ‘relationship’ if they don't live with us? I mean... Crap! If I want to add friends, will they also have to live under the same roof?
 

Dnds

Out to Play
Donor
Jul 5, 2017
319
406
Its taking a long time to load dialogue responses, like 2 minutes. I tried the 6 and 8 models. Is there anything I can do on Kobold to make it be faster? Otherwise I might have to resort to the one online, assuming its faster.
 

Spillthebeans

Active Member
Jun 21, 2017
764
1,294
Apart from that, I don't know, maybe I've missed something, but during the ‘construction’ of the characters... Eg: I'm a guy and I'm married. I add mum, dads and a sister when I add people, they all seem to live under the same roof as my mc. Then I want to add my wife's family, so say yes to ‘add customization family’, so I create my wife's family, but I can't link them in ‘relationship’ if they don't live with us? I mean... Crap! If I want to add friends, will they also have to live under the same roof?
I haven't played the latest version but you can kind of get around that somewhat by cheating maximum dominance between you and the person you want to leave the house, then tell them to leave, they'll then move out into their own apartment(s) with a relationship hit but you can cheat that back. I'm not sure of what happens when you have a couple leave individually, likely they would move into their own separate apartments. As far as friends go, apparently in this newest version you can create custom, separate households, so you could create whatever friends you want and cheat an initial relationship after character creation.
 
  • Like
Reactions: ♂Brout&Minou♀

kkai

Member
Dec 21, 2016
181
315
Its taking a long time to load dialogue responses, like 2 minutes. I tried the 6 and 8 models. Is there anything I can do on Kobold to make it be faster? Otherwise I might have to resort to the one online, assuming its faster.
First thing to do is make sure the model and context are fitting into your VRAM. If you're using windows you can just Task Manager->Performance->GPU. When you load the model in KCPP you'll see your dedicated GPU memory usage spike. Make sure it doesn't fill enough to start using Shared GPU Memory. If it is, you need to use a smaller model or quant. You can also play with the layers value. If you absolutely cannot fit it in VRAM, you can split layers between GPU and CPU. This is still slow, but not as slow as it using shared memory.

The problem I see is, at the moment, Influencing doesn't seem to handle streaming from the backend. I get about 13 T/s (tokens per second) which is faster than I read so it would be fine if it streamed (displayed tokens/words as they are sent from the backend), but instead it waits for the full response before displaying it. It's possible it's because the dev wants the game to do some post-processing on it or something but it's a pretty big problem when it's too slow for even someone with a decent GPU to tolerate. Waiting 20 seconds or so for a full response is just too long for me to want to bother.

Even people who only get 5 T/s won't find it too unbearable if it streams. Hopefully this isn't a limitation of Ren'Py because it's pretty serious.
 

Leonxwerty

New Member
Donor
Nov 14, 2017
11
6
I've been using this ever since I got back to the AI scene via SillyTavern:


If you ever need a way to find compatibility with your current GPU VRAM available for it, using that link may help a bit to finding compatible models for your GPU.

Running EXL2 from the local text generation has been superb, fast, and quite logical.
 
  • Like
Reactions: Cheroky
4.10 star(s) 52 Votes