Nearly as strong? It beats 13B, 20B engines in most areas.They just released the Clio model which is nearly as strong as Llama 7b even though the model is 3b. Currently, it is only available to Opus but in two weeks, it will be released to everyone.
They released a graph saying that it was not as strong as Llama 7b but was nearly as strong. Even though it was 3b. According to the chart, it was also stronger than a gpt-neox 20B model and nearly as strong as a gpt 175B model even though it was a 3b model. They used 6 terabytes of training data text for clio.Nearly as strong? It beats 13B, 20B engines in most areas.
It depends in usage. I don't think Llama 7B or even 30B is as good as Clio in storytelling.They released a graph saying that it was not as strong as Llama 7b but was nearly as strong. Even though it was 3b. According to the chart, it was also stronger than a gpt-neox 20B model and nearly as strong as a gpt 175B model even though it was a 3b model. They used 6 terabytes of training data text for clio.
If by "unlimited input" you mean "let me dump this 300 page book in memory and generate from there", this isn't really possible from a technical level. All AI models work within the confines of a set amount of "tokens", which are a set assortment of words, punctuation marks, letters, or a combination of all three, which serves as its memory. When it hits the limit of it, it will inevitably have to throw some of it out to make way for new information. For most models you'll find, this limit will range from anywhere from 2048 to 8192 tokens, with a few (OpenAI's flagship models and a select assortment of experimental indie ones) reaching into 16K tokens. In practical terms, NovelAI gives an estimate of about 2.9 characters per token, though this may vary depending on the actual contents of the text.Does anybody know AI story generator tool similar to ToolBaz but with unlimited input?
Thanks for the explanation, but NovelAI is different than ToolBaz.If by "unlimited input" you mean "let me dump this 300 page book in memory and generate from there", this isn't really possible from a technical level. All AI models work within the confines of a set amount of "tokens", which are a set assortment of words, punctuation marks, letters, or a combination of all three, which serves as its memory. When it hits the limit of it, it will inevitably have to throw some of it out to make way for new information. For most models you'll find, this limit will range from anywhere from 2048 to 8192 tokens, with a few (OpenAI's flagship models and a select assortment of experimental indie ones) reaching into 16K tokens. In practical terms, NovelAI gives an estimate of about 2.9 characters per token, though this may vary depending on the actual contents of the text.
You can get around this with NovelAI, however, by using the story's memory and lorebook entries. The memory will allow you to put the most important information to always be in memory at all times, while the lorebook will allow for relevant information to be brought back into memory whenever specific keywords are hit by the text. You can then focus your generation efforts on a specific scene at a time, and then carry over relevant story information or cool stuff generated by the AI into memory/the relevant lorebook entry.
Thanks for the answer, what sort of program are you using for the self hosting?The entire concept of all the current AI stuff is that a trained neural network is handed the last X words and asked to decide the 1 single next word, repeated eg. 200 times to produce 200 words of output (except instead of words it's tokens - word fragments).
Behind the scenes, that'll be the same for ToolBaz, NovelAI, TalkDirty.ai and everything else out there.
If you have the RAM, you could try self-hosting the mpt-7b-storywriter model with 65536+ input tokens. Its writing didn't impress me, however. As a compromise, there are the WizardLM-13B-SuperHOT-8k models with 8192 input tokens and with much nicer output quality. I tried asking it to write a story and the result was okay, see attached screenshot (it took about 22 seconds to generate on a 3090)
Or simply wait it out. GPT-4 supports 8192 input tokens (and supposedly a variant can do 32768 input tokens), maybe ToolBaz or a competitor with the same service will switch to GPT-4 in the not too far future.
What is this again?
The current model Clio is already the BEST storytelling model ever. But it is only 3B size. While the model mentioned on the screenshot is bigger, fixed and more powerful. And will be released soon.What is this again?
Use Kobold Lite.Yeah yeah you can read this wall post, above. -SNIP-
Was this news talking about Kayra model? or it was something else?The current model Clio is already the BEST storytelling model ever. But it is only 3B size. While the model mentioned on the screenshot is bigger, fixed and more powerful. And will be released soon.
Kayra is a better version of Clio. News about Clio were old...Was this news talking about Kayra model? or it was something else?
It's really good. Leagues better than the previous models. It actually remembers thing without having to store it in the lorebook first (though that helps).Kayra is a better version of Clio. News about Clio were old...