HabibAhmed commited on
Commit
c65e5a8
·
verified ·
1 Parent(s): bca5e2f

Upload folder using huggingface_hub

Browse files
.ipynb_checkpoints/Falcon3-1B-NF4-lora-FP16-Evaluation_Results-checkpoint.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eval_loss": 2.5688072489562117,
3
+ "perplexity": 13.050249455581184,
4
+ "performance_metrics": {
5
+ "accuracy:": 0.9919893190921228,
6
+ "precision:": 1.0,
7
+ "recall:": 1.0,
8
+ "f1:": 1.0,
9
+ "bleu:": 0.8599438812778457,
10
+ "rouge:": {
11
+ "rouge1": 0.9209388932567808,
12
+ "rouge2": 0.9199417562331094,
13
+ "rougeL": 0.9208951190441694
14
+ },
15
+ "semantic_similarity_avg:": 0.9860197901725769
16
+ },
17
+ "mauve": 0.8745679633596274,
18
+ "inference_performance": {
19
+ "min_latency_ms": 41.147708892822266,
20
+ "max_latency_ms": 2038.865566253662,
21
+ "lower_quartile_ms": 1533.2009196281433,
22
+ "median_latency_ms": 1568.9027309417725,
23
+ "upper_quartile_ms": 1620.8247542381287,
24
+ "avg_latency_ms": 1498.4863044741317,
25
+ "min_memory_mb": 0.0,
26
+ "max_memory_mb": 0.00439453125,
27
+ "lower_quartile_mb": 0.0,
28
+ "median_memory_mb": 0.0,
29
+ "upper_quartile_mb": 0.0,
30
+ "avg_memory_mb": 2.9335989652870494e-06
31
+ }
32
+ }
.ipynb_checkpoints/Falcon3-1B-NF4-lora-FP16-Latency_Histogram-checkpoint.png ADDED
.ipynb_checkpoints/Falcon3-1B-NF4-lora-FP16-Memory_Histogram-checkpoint.png ADDED
.ipynb_checkpoints/Falcon3-1B-NF4-lora-FP16-Memory_Usage_Curve-checkpoint.png ADDED
Falcon3-1B-NF4-lora-FP16-Evaluation_Results.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eval_loss:": 2.746601107561381,
3
+ "perplexity:": 15.589554514190564,
4
+ "performance_metrics:": {
5
+ "accuracy:": 0.9939919893190922,
6
+ "precision:": 1.0,
7
+ "recall:": 1.0,
8
+ "f1:": 1.0,
9
+ "bleu:": 0.8631491900537053,
10
+ "rouge:": {
11
+ "rouge1": 0.9227783862741349,
12
+ "rouge2": 0.9216706649447948,
13
+ "rougeL": 0.9227142207671568
14
+ },
15
+ "semantic_similarity_avg:": 0.9864102005958557
16
+ },
17
+ "mauve:": 0.6804962052299759,
18
+ "inference_performance:": {
19
+ "min_latency_ms": 46.35119438171387,
20
+ "max_latency_ms": 2166.898727416992,
21
+ "lower_quartile_ms": 1524.4737267494202,
22
+ "median_latency_ms": 1553.6243915557861,
23
+ "upper_quartile_ms": 1574.8480558395386,
24
+ "avg_latency_ms": 1516.4662039327686,
25
+ "min_memory_gb": 0.09034490585327148,
26
+ "max_memory_gb": 0.09083318710327148,
27
+ "lower_quartile_gb": 0.09034490585327148,
28
+ "median_memory_gb": 0.09083318710327148,
29
+ "upper_quartile_gb": 0.09083318710327148,
30
+ "avg_memory_gb": 0.09066173454152249,
31
+ "model_load_memory_gb": 1.6676692962646484,
32
+ "avg_inference_memory_gb": 0.09066173454152249
33
+ }
34
+ }
Falcon3-1B-NF4-lora-FP16-Inference_Curve.png ADDED
Falcon3-1B-NF4-lora-FP16-Latency_Histogram.png ADDED
Falcon3-1B-NF4-lora-FP16-Memory_Histogram.png ADDED
Falcon3-1B-NF4-lora-FP16-Memory_Usage_Curve.png ADDED