HabibAhmed commited on
Commit
c66a73b
·
verified ·
1 Parent(s): 3319a5e

Upload folder using huggingface_hub

Browse files
.ipynb_checkpoints/Falcon3-1B-INT8-lora-BF16-Evaluation_Results-checkpoint.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eval_loss:": 1.9421286801192088,
3
+ "perplexity:": 6.973579702697131,
4
+ "performance_metrics:": {
5
+ "accuracy:": 0.985981308411215,
6
+ "precision:": 1.0,
7
+ "recall:": 1.0,
8
+ "f1:": 1.0,
9
+ "bleu:": 0.8613288102630609,
10
+ "rouge:": {
11
+ "rouge1": 0.9182306633191507,
12
+ "rouge2": 0.9157780948282156,
13
+ "rougeL": 0.9178061579709774
14
+ },
15
+ "semantic_similarity_avg:": 0.9822680950164795
16
+ },
17
+ "mauve:": 0.6879283933970437,
18
+ "inference_performance:": {
19
+ "min_latency_ms": 80.66940307617188,
20
+ "max_latency_ms": 3611.3381385803223,
21
+ "lower_quartile_ms": 2922.58220911026,
22
+ "median_latency_ms": 3024.101734161377,
23
+ "upper_quartile_ms": 3077.1530270576477,
24
+ "avg_latency_ms": 2588.715038885262,
25
+ "min_memory_gb": 0.07576847076416016,
26
+ "max_memory_gb": 0.09750175476074219,
27
+ "lower_quartile_gb": 0.0917661190032959,
28
+ "median_memory_gb": 0.09322285652160645,
29
+ "upper_quartile_gb": 0.09434831142425537,
30
+ "avg_memory_gb": 0.09272574136986116,
31
+ "model_load_memory_gb": 5.110301971435547,
32
+ "avg_inference_memory_gb": 0.09272574136986116
33
+ }
34
+ }
Falcon3-1B-INT8-lora-BF16-Evaluation_Results.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eval_loss:": 1.9421286801192088,
3
+ "perplexity:": 6.973579702697131,
4
+ "performance_metrics:": {
5
+ "accuracy:": 0.985981308411215,
6
+ "precision:": 1.0,
7
+ "recall:": 1.0,
8
+ "f1:": 1.0,
9
+ "bleu:": 0.8613288102630609,
10
+ "rouge:": {
11
+ "rouge1": 0.9182306633191507,
12
+ "rouge2": 0.9157780948282156,
13
+ "rougeL": 0.9178061579709774
14
+ },
15
+ "semantic_similarity_avg:": 0.9822680950164795
16
+ },
17
+ "mauve:": 0.6879283933970437,
18
+ "inference_performance:": {
19
+ "min_latency_ms": 80.66940307617188,
20
+ "max_latency_ms": 3611.3381385803223,
21
+ "lower_quartile_ms": 2922.58220911026,
22
+ "median_latency_ms": 3024.101734161377,
23
+ "upper_quartile_ms": 3077.1530270576477,
24
+ "avg_latency_ms": 2588.715038885262,
25
+ "min_memory_gb": 0.07576847076416016,
26
+ "max_memory_gb": 0.09750175476074219,
27
+ "lower_quartile_gb": 0.0917661190032959,
28
+ "median_memory_gb": 0.09322285652160645,
29
+ "upper_quartile_gb": 0.09434831142425537,
30
+ "avg_memory_gb": 0.09272574136986116,
31
+ "model_load_memory_gb": 5.110301971435547,
32
+ "avg_inference_memory_gb": 0.09272574136986116
33
+ }
34
+ }
Falcon3-1B-INT8-lora-BF16-Inference_Curve.png ADDED
Falcon3-1B-INT8-lora-BF16-Latency_Histogram.png ADDED
Falcon3-1B-INT8-lora-BF16-Memory_Histogram.png ADDED
Falcon3-1B-INT8-lora-BF16-Memory_Usage_Curve.png ADDED