You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Thanks for taking the time to fill out this bug report!
9
+
- type: textarea
10
+
id: bug-description
11
+
attributes:
12
+
label: Describe the bug
13
+
description: A clear and concise description of what the bug is.
14
+
placeholder: Bug description
15
+
validations:
16
+
required: true
17
+
- type: checkboxes
18
+
attributes:
19
+
label: Is there an existing issue for this?
20
+
description: Please search to see if an issue already exists for the issue you encountered.
21
+
options:
22
+
- label: I have searched the existing issues
23
+
required: true
24
+
- type: textarea
25
+
id: reproduction
26
+
attributes:
27
+
label: Reproduction
28
+
description: Please provide the steps necessary to reproduce your issue.
29
+
placeholder: Reproduction
30
+
validations:
31
+
required: true
32
+
- type: textarea
33
+
id: screenshot
34
+
attributes:
35
+
label: Screenshot
36
+
description: "If possible, please include screenshot(s) so that we can understand what the issue is."
37
+
- type: textarea
38
+
id: logs
39
+
attributes:
40
+
label: Logs
41
+
description: "Please include the full stacktrace of the errors you get in the command-line (if any)."
42
+
render: shell
43
+
validations:
44
+
required: true
45
+
- type: textarea
46
+
id: system-info
47
+
attributes:
48
+
label: System Info
49
+
description: "Please share your system info with us: operating system, GPU brand, and GPU model. If you are using a Google Colab notebook, mention that instead."
close-issue-message: "This issue has been closed due to inactivity for 30 days. If you believe it is still relevant, you can reopen it (if you are the author) or leave a comment below."
See also: [Installation instructions for human beings](https://github.com/oobabooga/text-generation-webui/wiki/Installation-instructions-for-human-beings).
63
+
> **Note**
64
+
> 1. If you are on Windows, it may be easier to run the commands above in a WSL environment. The performance may also be better.
65
+
> 2. For a more detailed, user-contributed guide, see: [Installation instructions for human beings](https://github.com/oobabooga/text-generation-webui/wiki/Installation-instructions-for-human-beings).
64
66
65
67
## Installation option 2: one-click installers
66
68
@@ -140,8 +142,9 @@ Optionally, you can use the following command-line flags:
140
142
|`--cai-chat`| Launch the web UI in chat mode with a style similar to Character.AI's. If the file `img_bot.png` or `img_bot.jpg` exists in the same folder as server.py, this image will be used as the bot's profile picture. Similarly, `img_me.png` or `img_me.jpg` will be used as your profile picture. |
141
143
|`--cpu`| Use the CPU to generate text.|
142
144
|`--load-in-8bit`| Load the model with 8-bit precision.|
143
-
|`--load-in-4bit`| Load the model with 4-bit precision. Currently only works with LLaMA.|
144
-
|`--gptq-bits GPTQ_BITS`| Load a pre-quantized model with specified precision. 2, 3, 4 and 8 (bit) are supported. Currently only works with LLaMA. |
145
+
|`--load-in-4bit`| DEPRECATED: use `--gptq-bits 4` instead. |
146
+
|`--gptq-bits GPTQ_BITS`| Load a pre-quantized model with specified precision. 2, 3, 4 and 8 (bit) are supported. Currently only works with LLaMA and OPT. |
147
+
|`--gptq-model-type MODEL_TYPE`| Model type of pre-quantized model. Currently only LLaMa and OPT are supported. |
145
148
|`--bf16`| Load the model with bfloat16 precision. Requires NVIDIA Ampere GPU. |
146
149
|`--auto-devices`| Automatically split the model across the available GPU(s) and CPU.|
147
150
|`--disk`| If the model is too large for your GPU(s) and CPU combined, send the remaining layers to the disk. |
0 commit comments