SFT Final Models
Collection
Models that were trained on clembench v0.9 - v1.6
•
4 items
•
Updated
This model is a fine-tuned version of unsloth/meta-llama-3.1-8b-instruct-bnb-4bit on the None dataset. It achieves the following results on the evaluation set:
More information needed
More information needed
More information needed
The following hyperparameters were used during training:
Training Loss | Epoch | Step | Validation Loss |
---|---|---|---|
0.2648 | 0.0565 | 100 | 0.3758 |
0.1823 | 0.1130 | 200 | 0.2916 |
0.1909 | 0.1695 | 300 | 0.2712 |
0.1562 | 0.2260 | 400 | 0.2602 |
0.169 | 0.2825 | 500 | 0.2468 |
0.1553 | 0.3390 | 600 | 0.2520 |
0.1015 | 0.3955 | 700 | 0.2467 |
0.139 | 0.4520 | 800 | 0.2501 |
0.1013 | 0.5085 | 900 | 0.2399 |
0.1229 | 0.5650 | 1000 | 0.2389 |
0.1457 | 0.6215 | 1100 | 0.2386 |