aiyets commited on
Commit
c18fd9f
·
verified ·
1 Parent(s): 3431099

Model save

Browse files
Files changed (1) hide show
  1. README.md +39 -24
README.md CHANGED
@@ -1,44 +1,59 @@
1
  ---
2
- language:
3
- - ko
4
- license: gemma
5
- datasets:
6
- - kuotient/orca-math-korean-dpo-pairs
7
- - aiyets/argilla_dpo-mix-7k-ko
8
  base_model: google/gemma-2-9b-it
 
 
 
 
 
 
 
 
 
9
  ---
10
- # Model Card for Model ID
11
 
12
- <!-- Provide a quick summary of what the model is/does. -->
 
13
 
14
- This modelcard aims to be a base template for new models. It has been generated using [this raw template](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/templates/modelcard_template.md?plain=1).
15
 
16
- ## Model Details
17
 
18
- ### Model Description
19
 
20
- <!-- Provide a longer summary of what this model is. -->
21
 
22
- ## Training Details
23
 
24
- ### Training Data
25
 
26
- <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
27
 
28
- [More Information Needed]
29
 
30
- ### Training Procedure
31
 
32
- <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
33
 
34
- #### Training Hyperparameters
 
 
 
 
 
 
 
 
 
 
35
 
36
- - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
37
 
38
- ## Evaluation
39
 
40
- <!-- This section describes the evaluation protocols and provides the results. -->
41
 
42
- #### Hardware
43
 
44
- [More Information Needed]
 
 
 
 
 
1
  ---
 
 
 
 
 
 
2
  base_model: google/gemma-2-9b-it
3
+ library_name: peft
4
+ license: gemma
5
+ tags:
6
+ - trl
7
+ - dpo
8
+ - generated_from_trainer
9
+ model-index:
10
+ - name: gemma-2-9b-it-dpo-dual_001
11
+ results: []
12
  ---
 
13
 
14
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
+ should probably proofread and complete it, then remove this comment. -->
16
 
17
+ # gemma-2-9b-it-dpo-dual_001
18
 
19
+ This model is a fine-tuned version of [google/gemma-2-9b-it](https://huggingface.co/google/gemma-2-9b-it) on the None dataset.
20
 
21
+ ## Model description
22
 
23
+ More information needed
24
 
25
+ ## Intended uses & limitations
26
 
27
+ More information needed
28
 
29
+ ## Training and evaluation data
30
 
31
+ More information needed
32
 
33
+ ## Training procedure
34
 
35
+ ### Training hyperparameters
36
 
37
+ The following hyperparameters were used during training:
38
+ - learning_rate: 5e-07
39
+ - train_batch_size: 1
40
+ - eval_batch_size: 8
41
+ - seed: 42
42
+ - gradient_accumulation_steps: 4
43
+ - total_train_batch_size: 4
44
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
+ - lr_scheduler_type: cosine
46
+ - lr_scheduler_warmup_ratio: 0.1
47
+ - num_epochs: 1
48
 
49
+ ### Training results
50
 
 
51
 
 
52
 
53
+ ### Framework versions
54
 
55
+ - PEFT 0.13.1
56
+ - Transformers 4.45.0
57
+ - Pytorch 2.4.0+cu121
58
+ - Datasets 3.0.0
59
+ - Tokenizers 0.20.0