HabibAhmed commited on
Commit
fa91824
·
verified ·
1 Parent(s): 0e6ded7

Upload folder using huggingface_hub

Browse files
.ipynb_checkpoints/Phi2-2B-FP4-lora-FP16-Evaluation_Results-checkpoint.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eval_loss": 0.624643579111105,
3
+ "perplexity": 1.8675801942004093,
4
+ "performance_metrics": {
5
+ "accuracy:": 1.0,
6
+ "precision:": 1.0,
7
+ "recall:": 1.0,
8
+ "f1:": 1.0,
9
+ "bleu:": 0.9592784829712196,
10
+ "rouge:": {
11
+ "rouge1": 0.9779364422053355,
12
+ "rouge2": 0.9777944006210844,
13
+ "rougeL": 0.9779364422053355
14
+ },
15
+ "semantic_similarity_avg:": 0.9979658126831055
16
+ },
17
+ "mauve": 0.8616094522887235,
18
+ "inference_performance": {
19
+ "min_latency_ms": 64.36681747436523,
20
+ "max_latency_ms": 2727.415084838867,
21
+ "lower_quartile_ms": 66.3864016532898,
22
+ "median_latency_ms": 67.23260879516602,
23
+ "upper_quartile_ms": 2232.484459877014,
24
+ "avg_latency_ms": 822.4574397498361,
25
+ "min_memory_mb": 0.0,
26
+ "max_memory_mb": 0.00439453125,
27
+ "lower_quartile_mb": 0.0,
28
+ "median_memory_mb": 0.0,
29
+ "upper_quartile_mb": 0.0,
30
+ "avg_memory_mb": 2.9335989652870494e-06
31
+ }
32
+ }
Phi2-2B-FP4-lora-FP16-Evaluation_Results.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eval_loss:": 0.624643579111105,
3
+ "perplexity:": 1.8675801942004093,
4
+ "performance_metrics:": {
5
+ "accuracy:": 1.0,
6
+ "precision:": 1.0,
7
+ "recall:": 1.0,
8
+ "f1:": 1.0,
9
+ "bleu:": 0.959783834510939,
10
+ "rouge:": {
11
+ "rouge1": 0.9782729194127011,
12
+ "rouge2": 0.9781230898662496,
13
+ "rougeL": 0.9782729194127011
14
+ },
15
+ "semantic_similarity_avg:": 0.9978792071342468
16
+ },
17
+ "mauve:": 0.8616094522887235,
18
+ "inference_performance:": {
19
+ "min_latency_ms": 70.59812545776367,
20
+ "max_latency_ms": 3547.1789836883545,
21
+ "lower_quartile_ms": 71.33513689041138,
22
+ "median_latency_ms": 87.7445936203003,
23
+ "upper_quartile_ms": 2200.851857662201,
24
+ "avg_latency_ms": 899.7708788223674,
25
+ "min_memory_gb": 0.22810125350952148,
26
+ "max_memory_gb": 0.22858953475952148,
27
+ "lower_quartile_gb": 0.22810125350952148,
28
+ "median_memory_gb": 0.22858953475952148,
29
+ "upper_quartile_gb": 0.22858953475952148,
30
+ "avg_memory_gb": 0.22840341420294605,
31
+ "model_load_memory_gb": 4.8278045654296875,
32
+ "avg_inference_memory_gb": 0.22840341420294605
33
+ }
34
+ }
Phi2-2B-FP4-lora-FP16-Inference_Curve.png ADDED
Phi2-2B-FP4-lora-FP16-Latency_Histogram.png ADDED
Phi2-2B-FP4-lora-FP16-Memory_Histogram.png ADDED
Phi2-2B-FP4-lora-FP16-Memory_Usage_Curve.png ADDED