Updating model files
Browse files
README.md
CHANGED
@@ -2,6 +2,17 @@
|
|
2 |
license: other
|
3 |
inference: false
|
4 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
5 |
|
6 |
# gpt4-x-vicuna-13B-GPTQ
|
7 |
|
@@ -14,7 +25,7 @@ It is the result of quantising to 4bit using [GPTQ-for-LLaMa](https://github.com
|
|
14 |
* [4bit GPTQ models for GPU inference](https://huggingface.co/TheBloke/gpt4-x-vicuna-13B-GPTQ).
|
15 |
* [4bit and 5bit GGML models for CPU inference](https://huggingface.co/TheBloke/gpt4-x-vicuna-13B-GGML).
|
16 |
* [float16 HF model for unquantised and 8bit GPU inference](https://huggingface.co/TheBloke/gpt4-x-vicuna-13B-HF).
|
17 |
-
|
18 |
## How to easily download and use this model in text-generation-webui
|
19 |
|
20 |
Open the text-generation-webui UI as normal.
|
@@ -50,6 +61,17 @@ It was created without the `--act-order` parameter. It may have slightly lower i
|
|
50 |
CUDA_VISIBLE_DEVICES=0 python3 llama.py GPT4All-13B-snoozy c4 --wbits 4 --true-sequential --groupsize 128 --save_safetensors GPT4-x-Vicuna-13B-GPTQ-4bit-128g.compat.act-order.safetensors
|
51 |
```
|
52 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
53 |
# Original model card
|
54 |
|
55 |
As a base model used https://huggingface.co/eachadea/vicuna-13b-1.1
|
@@ -70,4 +92,4 @@ Wizard LM by https://github.com/nlpxucan
|
|
70 |
|
71 |
Nous Research Instruct Dataset by https://huggingface.co/karan4d and https://huggingface.co/huemin
|
72 |
|
73 |
-
Compute provided by our project sponsor https://redmond.ai/
|
|
|
2 |
license: other
|
3 |
inference: false
|
4 |
---
|
5 |
+
<div style="width: 100%;">
|
6 |
+
<img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;">
|
7 |
+
</div>
|
8 |
+
<div style="display: flex; justify-content: space-between; width: 100%;">
|
9 |
+
<div style="display: flex; flex-direction: column; align-items: flex-start;">
|
10 |
+
<p><a href="https://discord.gg/UBgz4VXf">Chat & support: my new Discord server</a></p>
|
11 |
+
</div>
|
12 |
+
<div style="display: flex; flex-direction: column; align-items: flex-end;">
|
13 |
+
<p><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? Patreon coming soon!</a></p>
|
14 |
+
</div>
|
15 |
+
</div>
|
16 |
|
17 |
# gpt4-x-vicuna-13B-GPTQ
|
18 |
|
|
|
25 |
* [4bit GPTQ models for GPU inference](https://huggingface.co/TheBloke/gpt4-x-vicuna-13B-GPTQ).
|
26 |
* [4bit and 5bit GGML models for CPU inference](https://huggingface.co/TheBloke/gpt4-x-vicuna-13B-GGML).
|
27 |
* [float16 HF model for unquantised and 8bit GPU inference](https://huggingface.co/TheBloke/gpt4-x-vicuna-13B-HF).
|
28 |
+
|
29 |
## How to easily download and use this model in text-generation-webui
|
30 |
|
31 |
Open the text-generation-webui UI as normal.
|
|
|
61 |
CUDA_VISIBLE_DEVICES=0 python3 llama.py GPT4All-13B-snoozy c4 --wbits 4 --true-sequential --groupsize 128 --save_safetensors GPT4-x-Vicuna-13B-GPTQ-4bit-128g.compat.act-order.safetensors
|
62 |
```
|
63 |
|
64 |
+
## Want to support my work?
|
65 |
+
|
66 |
+
I've had a lot of people ask if they can contribute. I love providing models and helping people, but it is starting to rack up pretty big cloud computing bills.
|
67 |
+
|
68 |
+
So if you're able and willing to contribute, it'd be most gratefully received and will help me to keep providing models, and work on various AI projects.
|
69 |
+
|
70 |
+
Donaters will get priority support on any and all AI/LLM/model questions, and I'll gladly quantise any model you'd like to try.
|
71 |
+
|
72 |
+
* Patreon: coming soon! (just awaiting approval)
|
73 |
+
* Ko-Fi: https://ko-fi.com/TheBlokeAI
|
74 |
+
* Discord: https://discord.gg/UBgz4VXf
|
75 |
# Original model card
|
76 |
|
77 |
As a base model used https://huggingface.co/eachadea/vicuna-13b-1.1
|
|
|
92 |
|
93 |
Nous Research Instruct Dataset by https://huggingface.co/karan4d and https://huggingface.co/huemin
|
94 |
|
95 |
+
Compute provided by our project sponsor https://redmond.ai/
|