TheBloke commited on
Commit
ff47f42
·
1 Parent(s): 9cb8c4e

Updating model files

Browse files
Files changed (1) hide show
  1. README.md +24 -2
README.md CHANGED
@@ -2,6 +2,17 @@
2
  license: other
3
  inference: false
4
  ---
 
 
 
 
 
 
 
 
 
 
 
5
 
6
  # gpt4-x-vicuna-13B-GPTQ
7
 
@@ -14,7 +25,7 @@ It is the result of quantising to 4bit using [GPTQ-for-LLaMa](https://github.com
14
  * [4bit GPTQ models for GPU inference](https://huggingface.co/TheBloke/gpt4-x-vicuna-13B-GPTQ).
15
  * [4bit and 5bit GGML models for CPU inference](https://huggingface.co/TheBloke/gpt4-x-vicuna-13B-GGML).
16
  * [float16 HF model for unquantised and 8bit GPU inference](https://huggingface.co/TheBloke/gpt4-x-vicuna-13B-HF).
17
-
18
  ## How to easily download and use this model in text-generation-webui
19
 
20
  Open the text-generation-webui UI as normal.
@@ -50,6 +61,17 @@ It was created without the `--act-order` parameter. It may have slightly lower i
50
  CUDA_VISIBLE_DEVICES=0 python3 llama.py GPT4All-13B-snoozy c4 --wbits 4 --true-sequential --groupsize 128 --save_safetensors GPT4-x-Vicuna-13B-GPTQ-4bit-128g.compat.act-order.safetensors
51
  ```
52
 
 
 
 
 
 
 
 
 
 
 
 
53
  # Original model card
54
 
55
  As a base model used https://huggingface.co/eachadea/vicuna-13b-1.1
@@ -70,4 +92,4 @@ Wizard LM by https://github.com/nlpxucan
70
 
71
  Nous Research Instruct Dataset by https://huggingface.co/karan4d and https://huggingface.co/huemin
72
 
73
- Compute provided by our project sponsor https://redmond.ai/
 
2
  license: other
3
  inference: false
4
  ---
5
+ <div style="width: 100%;">
6
+ <img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;">
7
+ </div>
8
+ <div style="display: flex; justify-content: space-between; width: 100%;">
9
+ <div style="display: flex; flex-direction: column; align-items: flex-start;">
10
+ <p><a href="https://discord.gg/UBgz4VXf">Chat & support: my new Discord server</a></p>
11
+ </div>
12
+ <div style="display: flex; flex-direction: column; align-items: flex-end;">
13
+ <p><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? Patreon coming soon!</a></p>
14
+ </div>
15
+ </div>
16
 
17
  # gpt4-x-vicuna-13B-GPTQ
18
 
 
25
  * [4bit GPTQ models for GPU inference](https://huggingface.co/TheBloke/gpt4-x-vicuna-13B-GPTQ).
26
  * [4bit and 5bit GGML models for CPU inference](https://huggingface.co/TheBloke/gpt4-x-vicuna-13B-GGML).
27
  * [float16 HF model for unquantised and 8bit GPU inference](https://huggingface.co/TheBloke/gpt4-x-vicuna-13B-HF).
28
+
29
  ## How to easily download and use this model in text-generation-webui
30
 
31
  Open the text-generation-webui UI as normal.
 
61
  CUDA_VISIBLE_DEVICES=0 python3 llama.py GPT4All-13B-snoozy c4 --wbits 4 --true-sequential --groupsize 128 --save_safetensors GPT4-x-Vicuna-13B-GPTQ-4bit-128g.compat.act-order.safetensors
62
  ```
63
 
64
+ ## Want to support my work?
65
+
66
+ I've had a lot of people ask if they can contribute. I love providing models and helping people, but it is starting to rack up pretty big cloud computing bills.
67
+
68
+ So if you're able and willing to contribute, it'd be most gratefully received and will help me to keep providing models, and work on various AI projects.
69
+
70
+ Donaters will get priority support on any and all AI/LLM/model questions, and I'll gladly quantise any model you'd like to try.
71
+
72
+ * Patreon: coming soon! (just awaiting approval)
73
+ * Ko-Fi: https://ko-fi.com/TheBlokeAI
74
+ * Discord: https://discord.gg/UBgz4VXf
75
  # Original model card
76
 
77
  As a base model used https://huggingface.co/eachadea/vicuna-13b-1.1
 
92
 
93
  Nous Research Instruct Dataset by https://huggingface.co/karan4d and https://huggingface.co/huemin
94
 
95
+ Compute provided by our project sponsor https://redmond.ai/