- Dec 28, 2018
- 1,716
- 3,308
So lewd allowed.
Recommended :
- Install where you have some free disk space because models are quite huge.
- Nvidia card with 8Go of VRAM.
- This
Click start_windows.bat.
Select your type of GPU, when asked to download a model select "Do not download a model".
Wait for it to finish completely (can take some time).
Once fully installed you should see this line :
Open your browser, type the URL and bookmark it.
(If you struggle with installation
Model VRAM usage :
Bigger the better, but choose a model that suits your card accordingly.
For exemple, if you have a 8Go VRAM card, choose a 7 billions model (7B).
(Watch
Installing a model :
In the model tab, go to the huggingface model page and paste the model name and click download.
Once downloaded, refrest models list and select it.
Model can take a long time to load (30B one take around 10min on a 3090)
In the same tab change wbits & group values to the model you use :
EDIT: If your model & Oobabooga/text-generation-webui are up-to-date and contain a quantize_config.json file, you may skip this.
Uncensored models (4bits quantized version) :
Vicuna
WizardLM
SuperCOT/StoryTelling
Pygmalion (you need to
Recommended :
- Install where you have some free disk space because models are quite huge.
- Nvidia card with 8Go of VRAM.
- This
You must be registered to see the links
to keep thing fresh.
You must be registered to see the links
and download/unzip one-click installer.Click start_windows.bat.
Select your type of GPU, when asked to download a model select "Do not download a model".
Wait for it to finish completely (can take some time).
Once fully installed you should see this line :
Open your browser, type the URL and bookmark it.
(If you struggle with installation
You must be registered to see the links
)Model VRAM usage :
Bigger the better, but choose a model that suits your card accordingly.
For exemple, if you have a 8Go VRAM card, choose a 7 billions model (7B).
(Watch
You must be registered to see the links
to reduce VRAM usage at the cost of time it takes to generate text).Installing a model :
In the model tab, go to the huggingface model page and paste the model name and click download.
Once downloaded, refrest models list and select it.
Model can take a long time to load (30B one take around 10min on a 3090)
In the same tab change wbits & group values to the model you use :
EDIT: If your model & Oobabooga/text-generation-webui are up-to-date and contain a quantize_config.json file, you may skip this.
Uncensored models (4bits quantized version) :
Vicuna
You must be registered to see the links
(4 wbits, 128 groupsize)
You must be registered to see the links
(4 wbits, 128 groupsize)
You must be registered to see the links
(4 wbits, None groupsize)WizardLM
You must be registered to see the links
(4 wbits, 128 groupsize)
You must be registered to see the links
(4 wbits, None groupsize)SuperCOT/StoryTelling
You must be registered to see the links
(4 wbits, None groupsize)Pygmalion (you need to
You must be registered to see the links
)
You must be registered to see the links
(4 wbits, 128 groupsize)
You must be registered to see the links
(4 wbits, 128 groupsize)
Last edited: