Ministral-3B-lora-BF16 / Ministral-3B-BF16-lora-BF16-Evaluation_Results.json
HabibAhmed's picture
Upload folder using huggingface_hub
83afb28 verified
{
"eval_loss:": 1.3282925000908221,
"perplexity:": 3.7745927651138,
"performance_metrics:": {
"accuracy:": 1.0,
"precision:": 1.0,
"recall:": 1.0,
"f1:": 1.0,
"bleu:": 0.953271324235709,
"rouge:": {
"rouge1": 0.9746492971444021,
"rouge2": 0.9744754227651914,
"rougeL": 0.9746492971444021
},
"semantic_similarity_avg:": 0.9969900250434875
},
"mauve:": 0.8685080885635668,
"inference_performance:": {
"min_latency_ms": 40.31872749328613,
"max_latency_ms": 1942.1961307525635,
"lower_quartile_ms": 41.228532791137695,
"median_latency_ms": 41.55397415161133,
"upper_quartile_ms": 1494.4334030151367,
"avg_latency_ms": 546.7280667864966,
"min_memory_gb": 0.13282537460327148,
"max_memory_gb": 0.13331365585327148,
"lower_quartile_gb": 0.13282537460327148,
"median_memory_gb": 0.13331365585327148,
"upper_quartile_gb": 0.13331365585327148,
"avg_memory_gb": 0.13310294797918998,
"model_load_memory_gb": 9.460663318634033,
"avg_inference_memory_gb": 0.13310294797918998
}
}