laolaorkk commited on
Commit
08b051f
·
verified ·
1 Parent(s): 0a1e3c4

Model save

Browse files
README.md CHANGED
@@ -1,15 +1,12 @@
1
  ---
2
  base_model: deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
3
  library_name: transformers
4
- model_name: Qwen2.5-1.5B-Open-R1-GRPO
5
  tags:
6
  - generated_from_trainer
7
- - trl
8
- - grpo
9
  licence: license
10
  ---
11
 
12
- # Model Card for Qwen2.5-1.5B-Open-R1-GRPO
13
 
14
  This model is a fine-tuned version of [deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
@@ -27,7 +24,7 @@ print(output["generated_text"])
27
 
28
  ## Training procedure
29
 
30
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/ruikangzhao-danmarks-tekniske-universitet-dtu/huggingface/runs/brfye0bx)
31
 
32
 
33
  This model was trained with GRPO, a method introduced in [DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models](https://huggingface.co/papers/2402.03300).
 
1
  ---
2
  base_model: deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
3
  library_name: transformers
 
4
  tags:
5
  - generated_from_trainer
 
 
6
  licence: license
7
  ---
8
 
9
+ # Model Card for None
10
 
11
  This model is a fine-tuned version of [deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B).
12
  It has been trained using [TRL](https://github.com/huggingface/trl).
 
24
 
25
  ## Training procedure
26
 
27
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/ruikangzhao-danmarks-tekniske-universitet-dtu/huggingface/runs/u8581ung)
28
 
29
 
30
  This model was trained with GRPO, a method introduced in [DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models](https://huggingface.co/papers/2402.03300).
all_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "total_flos": 0.0,
3
- "train_loss": 0.00015604885993525386,
4
- "train_runtime": 10592.7921,
5
- "train_samples": 100,
6
- "train_samples_per_second": 0.009,
7
- "train_steps_per_second": 0.001
8
  }
 
1
  {
2
  "total_flos": 0.0,
3
+ "train_loss": 1.862645149230957e-09,
4
+ "train_runtime": 387.7333,
5
+ "train_samples": 3,
6
+ "train_samples_per_second": 0.008,
7
+ "train_steps_per_second": 0.003
8
  }
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:aaef5d1ced970e06eac33309db3aa15291869bdd1e1cbaa6ef004a53c031e96a
3
  size 4996670464
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02688ef1124d12440fd239b434dddc8ee9f818de68b1a651e3999555db4595b3
3
  size 4996670464
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:473a9c09782385eaa0e3cf068bef93b025c5711a8f3b3e6f12dc40f15224dd56
3
  size 2111719976
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4d0eae97532a467f289e0be214543efb098996f4878c8bf54b251b4d15ac9e6
3
  size 2111719976
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "total_flos": 0.0,
3
- "train_loss": 0.00015604885993525386,
4
- "train_runtime": 10592.7921,
5
- "train_samples": 100,
6
- "train_samples_per_second": 0.009,
7
- "train_steps_per_second": 0.001
8
  }
 
1
  {
2
  "total_flos": 0.0,
3
+ "train_loss": 1.862645149230957e-09,
4
+ "train_runtime": 387.7333,
5
+ "train_samples": 3,
6
+ "train_samples_per_second": 0.008,
7
+ "train_steps_per_second": 0.003
8
  }
trainer_state.json CHANGED
@@ -1,31 +1,31 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.96,
5
  "eval_steps": 100,
6
- "global_step": 6,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
- "completion_length": 791.9205729166666,
13
- "epoch": 0.96,
14
- "kl": 0.0039011780130143356,
15
- "reward": 0.3671875,
16
- "reward_std": 0.20240005695571503,
17
- "rewards/accuracy_reward": 0.3671875,
18
  "rewards/format_reward": 0.0,
19
- "step": 6,
20
  "total_flos": 0.0,
21
- "train_loss": 0.00015604885993525386,
22
- "train_runtime": 10592.7921,
23
- "train_samples_per_second": 0.009,
24
- "train_steps_per_second": 0.001
25
  }
26
  ],
27
  "logging_steps": 10,
28
- "max_steps": 6,
29
  "num_input_tokens_seen": 0,
30
  "num_train_epochs": 1,
31
  "save_steps": 500,
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 1.0,
5
  "eval_steps": 100,
6
+ "global_step": 1,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "completion_length": 955.5833333333334,
13
+ "epoch": 1.0,
14
+ "kl": 0.0,
15
+ "reward": 0.25,
16
+ "reward_std": 0.15430335203806558,
17
+ "rewards/accuracy_reward": 0.25,
18
  "rewards/format_reward": 0.0,
19
+ "step": 1,
20
  "total_flos": 0.0,
21
+ "train_loss": 1.862645149230957e-09,
22
+ "train_runtime": 387.7333,
23
+ "train_samples_per_second": 0.008,
24
+ "train_steps_per_second": 0.003
25
  }
26
  ],
27
  "logging_steps": 10,
28
+ "max_steps": 1,
29
  "num_input_tokens_seen": 0,
30
  "num_train_epochs": 1,
31
  "save_steps": 500,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c8bd5b42bb729c1abe0b9035a79589d55fe19e38ed393671496f55d11e783fdc
3
  size 5752
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f0ba137453abddbfd819c5f1476a21083d44a541e4ecd3b4e7af905fe6f8160
3
  size 5752