linkanjarad commited on
Commit
0ae9efd
·
1 Parent(s): a2103c2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +63 -6
README.md CHANGED
@@ -5,6 +5,14 @@ tags:
5
  model-index:
6
  - name: Cerebras-GPT-Alpaca-590m
7
  results: []
 
 
 
 
 
 
 
 
8
  ---
9
 
10
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -12,21 +20,70 @@ should probably proofread and complete it, then remove this comment. -->
12
 
13
  # Cerebras-GPT-Alpaca-590m
14
 
15
- This model is a fine-tuned version of [cerebras/Cerebras-GPT-590M](https://huggingface.co/cerebras/Cerebras-GPT-590M) on an unknown dataset.
16
 
17
  ## Model description
18
 
19
- More information needed
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
20
 
21
  ## Intended uses & limitations
22
 
23
- More information needed
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
24
 
25
- ## Training and evaluation data
26
 
27
- More information needed
28
 
29
- ## Training procedure
30
 
31
  ### Training hyperparameters
32
 
 
5
  model-index:
6
  - name: Cerebras-GPT-Alpaca-590m
7
  results: []
8
+ widget:
9
+ - text: |-
10
+ ### Instruction:
11
+ Write a poem about pancakes.
12
+
13
+ ### Response:
14
+ datasets:
15
+ - tatsu-lab/alpaca
16
  ---
17
 
18
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
20
 
21
  # Cerebras-GPT-Alpaca-590m
22
 
23
+ This model is a fine-tuned version of [cerebras/Cerebras-GPT-590M](https://huggingface.co/cerebras/Cerebras-GPT-590M) on the [tatsu-lab/alpaca](https://huggingface.co/datasets/tatsu-lab/alpaca) dataset. This is a test to check if the concept of Databrick's [databricks/dolly-v1-6b](https://huggingface.co/databricks/dolly-v1-6b) can be applied to smaller models.
24
 
25
  ## Model description
26
 
27
+ This is an LLM trained to follow instructions, similar to ChatGPT, except much much smaller.
28
+
29
+ ## Sample Use
30
+
31
+ ```python
32
+ from transformers import pipeline
33
+
34
+ generator = pipeline('text-generation', model='linkanjarad/Cerebras-GPT-Alpaca-590m')
35
+
36
+ prompt = '''### Instruction:
37
+ Who is the first person on the moon?
38
+
39
+ ### Response:'''
40
+
41
+ output = generator(prompt, do_sample=True, min_length=10, max_length=50, temperature=0.5)
42
+ print(output[0]['generated_text'])
43
+ ```
44
 
45
  ## Intended uses & limitations
46
 
47
+ The model works best when given input in the format of the dataset it was trained on. Here is an example of the format of the [tatsu-lab/alpaca](https://huggingface.co/datasets/tatsu-lab/alpaca) dataset:
48
+
49
+ ```
50
+ ### Instruction:
51
+ What are the three primary colors?
52
+
53
+ ### Response:
54
+ The three primary colors are red, blue, and yellow.
55
+ ```
56
+
57
+ Here are a few sample model outputs:
58
+
59
+ ```
60
+ ### Instruction:
61
+ Generate a new and original business plan for a product or service related to AI.
62
+
63
+ ### Response:
64
+ The plan for the new product or service should include a comprehensive plan for AI-powered AI systems that will enable businesses to automate a wide variety of tasks such as making decisions, analyzing data, and creating product designs. The plan should include a detailed review of an innovative solution, and should also incorporate customer feedback and industry-specific requirements. The plan should also include a detailed review of the business case for an AI product or service. Additionally, the plan should include a detailed review of the company's competitive positioning, customer segmentation and customer feedback. Additionally, the plan should include a detailed review of product and pricing, as well as customer feedback and user research. Finally, the plan should involve both internal and external stakeholders, as well as the customer feedback, to ensure that stakeholders are satisfied with the plan. This plan should help ensure that the right solutions are available to
65
+ ```
66
+
67
+ ```
68
+ ### Instruction:
69
+ Write a very short story about hamsters and whales.
70
+
71
+ ### Response:
72
+ Once upon a time, there was a small hamster who lived on a beach in a remote village in the Pacific Ocean. One day, he decided to take a hike to the nearby town of San Francisco. He had heard about the whale and decided to learn more about it. He spent the day learning about the whale and its behavior, and spent the day visiting the local coral reef. He eventually arrived at the shore, where he saw the whales in their graceful and majestic forms. The journey would not be the same without the whale. He was astounded by what he had encountered the first time. He saw so much beauty and kindness, and it was a pleasure to see the beauty and kindness of the whale. He would never forget the day of the trip and knew that this gentle creature was his only source of sustenance. He would never forget the whales, who were so abundant,
73
+ ```
74
+
75
+ ```
76
+ ### Instruction:
77
+ Explain why it's important to evaluate evidence and not just accept it at face value.
78
+
79
+ ### Response:
80
+ Evaluating evidence, regardless of its type, can be a valuable way of understanding the issue. It helps us develop a better understanding of the issue, as well as help us make better decisions about our future. It also helps us to understand that our actions, decisions, and actions are often not just about the truth, but about the process of our lives. Evaluating evidence can help us stay up to date on current events and trends, while avoiding the distraction of unnecessary and unnecessary decisions. In other words, it helps us to learn from the past and make better decisions. As a result, it's important to evaluate evidence and not just accept it at face value. It's important to learn how to evaluate evidence through proper judgment, and to always evaluate evidence in a manner that is consistent with the evidence. Finally, it's important to look at the
81
+ ```
82
 
 
83
 
84
+ ## Training Procedure
85
 
86
+ The model was trained with Huggingface's 🤗 Trainer for about 0.5 epochs.
87
 
88
  ### Training hyperparameters
89