Commit
·
0ae9efd
1
Parent(s):
a2103c2
Update README.md
Browse files
README.md
CHANGED
@@ -5,6 +5,14 @@ tags:
|
|
5 |
model-index:
|
6 |
- name: Cerebras-GPT-Alpaca-590m
|
7 |
results: []
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
8 |
---
|
9 |
|
10 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
@@ -12,21 +20,70 @@ should probably proofread and complete it, then remove this comment. -->
|
|
12 |
|
13 |
# Cerebras-GPT-Alpaca-590m
|
14 |
|
15 |
-
This model is a fine-tuned version of [cerebras/Cerebras-GPT-590M](https://huggingface.co/cerebras/Cerebras-GPT-590M) on
|
16 |
|
17 |
## Model description
|
18 |
|
19 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
20 |
|
21 |
## Intended uses & limitations
|
22 |
|
23 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
24 |
|
25 |
-
## Training and evaluation data
|
26 |
|
27 |
-
|
28 |
|
29 |
-
|
30 |
|
31 |
### Training hyperparameters
|
32 |
|
|
|
5 |
model-index:
|
6 |
- name: Cerebras-GPT-Alpaca-590m
|
7 |
results: []
|
8 |
+
widget:
|
9 |
+
- text: |-
|
10 |
+
### Instruction:
|
11 |
+
Write a poem about pancakes.
|
12 |
+
|
13 |
+
### Response:
|
14 |
+
datasets:
|
15 |
+
- tatsu-lab/alpaca
|
16 |
---
|
17 |
|
18 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
|
|
20 |
|
21 |
# Cerebras-GPT-Alpaca-590m
|
22 |
|
23 |
+
This model is a fine-tuned version of [cerebras/Cerebras-GPT-590M](https://huggingface.co/cerebras/Cerebras-GPT-590M) on the [tatsu-lab/alpaca](https://huggingface.co/datasets/tatsu-lab/alpaca) dataset. This is a test to check if the concept of Databrick's [databricks/dolly-v1-6b](https://huggingface.co/databricks/dolly-v1-6b) can be applied to smaller models.
|
24 |
|
25 |
## Model description
|
26 |
|
27 |
+
This is an LLM trained to follow instructions, similar to ChatGPT, except much much smaller.
|
28 |
+
|
29 |
+
## Sample Use
|
30 |
+
|
31 |
+
```python
|
32 |
+
from transformers import pipeline
|
33 |
+
|
34 |
+
generator = pipeline('text-generation', model='linkanjarad/Cerebras-GPT-Alpaca-590m')
|
35 |
+
|
36 |
+
prompt = '''### Instruction:
|
37 |
+
Who is the first person on the moon?
|
38 |
+
|
39 |
+
### Response:'''
|
40 |
+
|
41 |
+
output = generator(prompt, do_sample=True, min_length=10, max_length=50, temperature=0.5)
|
42 |
+
print(output[0]['generated_text'])
|
43 |
+
```
|
44 |
|
45 |
## Intended uses & limitations
|
46 |
|
47 |
+
The model works best when given input in the format of the dataset it was trained on. Here is an example of the format of the [tatsu-lab/alpaca](https://huggingface.co/datasets/tatsu-lab/alpaca) dataset:
|
48 |
+
|
49 |
+
```
|
50 |
+
### Instruction:
|
51 |
+
What are the three primary colors?
|
52 |
+
|
53 |
+
### Response:
|
54 |
+
The three primary colors are red, blue, and yellow.
|
55 |
+
```
|
56 |
+
|
57 |
+
Here are a few sample model outputs:
|
58 |
+
|
59 |
+
```
|
60 |
+
### Instruction:
|
61 |
+
Generate a new and original business plan for a product or service related to AI.
|
62 |
+
|
63 |
+
### Response:
|
64 |
+
The plan for the new product or service should include a comprehensive plan for AI-powered AI systems that will enable businesses to automate a wide variety of tasks such as making decisions, analyzing data, and creating product designs. The plan should include a detailed review of an innovative solution, and should also incorporate customer feedback and industry-specific requirements. The plan should also include a detailed review of the business case for an AI product or service. Additionally, the plan should include a detailed review of the company's competitive positioning, customer segmentation and customer feedback. Additionally, the plan should include a detailed review of product and pricing, as well as customer feedback and user research. Finally, the plan should involve both internal and external stakeholders, as well as the customer feedback, to ensure that stakeholders are satisfied with the plan. This plan should help ensure that the right solutions are available to
|
65 |
+
```
|
66 |
+
|
67 |
+
```
|
68 |
+
### Instruction:
|
69 |
+
Write a very short story about hamsters and whales.
|
70 |
+
|
71 |
+
### Response:
|
72 |
+
Once upon a time, there was a small hamster who lived on a beach in a remote village in the Pacific Ocean. One day, he decided to take a hike to the nearby town of San Francisco. He had heard about the whale and decided to learn more about it. He spent the day learning about the whale and its behavior, and spent the day visiting the local coral reef. He eventually arrived at the shore, where he saw the whales in their graceful and majestic forms. The journey would not be the same without the whale. He was astounded by what he had encountered the first time. He saw so much beauty and kindness, and it was a pleasure to see the beauty and kindness of the whale. He would never forget the day of the trip and knew that this gentle creature was his only source of sustenance. He would never forget the whales, who were so abundant,
|
73 |
+
```
|
74 |
+
|
75 |
+
```
|
76 |
+
### Instruction:
|
77 |
+
Explain why it's important to evaluate evidence and not just accept it at face value.
|
78 |
+
|
79 |
+
### Response:
|
80 |
+
Evaluating evidence, regardless of its type, can be a valuable way of understanding the issue. It helps us develop a better understanding of the issue, as well as help us make better decisions about our future. It also helps us to understand that our actions, decisions, and actions are often not just about the truth, but about the process of our lives. Evaluating evidence can help us stay up to date on current events and trends, while avoiding the distraction of unnecessary and unnecessary decisions. In other words, it helps us to learn from the past and make better decisions. As a result, it's important to evaluate evidence and not just accept it at face value. It's important to learn how to evaluate evidence through proper judgment, and to always evaluate evidence in a manner that is consistent with the evidence. Finally, it's important to look at the
|
81 |
+
```
|
82 |
|
|
|
83 |
|
84 |
+
## Training Procedure
|
85 |
|
86 |
+
The model was trained with Huggingface's 🤗 Trainer for about 0.5 epochs.
|
87 |
|
88 |
### Training hyperparameters
|
89 |
|