fsagorsarker/bangla-bert-base
Browse files- sagorsarker/bangla-bert-base/checkpoint-4500/optimizer.pt +3 -0
- sagorsarker/bangla-bert-base/checkpoint-4500/pytorch_model.bin +3 -0
- sagorsarker/bangla-bert-base/checkpoint-4500/rng_state.pth +3 -0
- sagorsarker/bangla-bert-base/checkpoint-4500/scaler.pt +3 -0
- sagorsarker/bangla-bert-base/checkpoint-4500/scheduler.pt +3 -0
- sagorsarker/bangla-bert-base/checkpoint-4500/special_tokens_map.json +1 -0
- sagorsarker/bangla-bert-base/checkpoint-4500/tokenizer_config.json +1 -0
- sagorsarker/bangla-bert-base/checkpoint-4500/trainer_state.json +136 -0
- sagorsarker/bangla-bert-base/checkpoint-4500/training_args.bin +3 -0
- sagorsarker/bangla-bert-base/checkpoint-4500/vocab.txt +0 -0
- sagorsarker/bangla-bert-base/runs/Jan13_14-45-35_22ece1fdce42/1642085147.7494185/events.out.tfevents.1642085147.22ece1fdce42.7156.1 +3 -0
- sagorsarker/bangla-bert-base/runs/Jan13_14-45-35_22ece1fdce42/events.out.tfevents.1642085147.22ece1fdce42.7156.0 +3 -0
sagorsarker/bangla-bert-base/checkpoint-4500/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0e3fc9e7ac5844ce8531b48c8344b2fc0377461d73f935bca66bb3974d620c51
|
3 |
+
size 1310575333
|
sagorsarker/bangla-bert-base/checkpoint-4500/pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7cf2623c5633772eeb6812b01d27d375887f19ec315a3cacf99ebe5cd91debc4
|
3 |
+
size 657676845
|
sagorsarker/bangla-bert-base/checkpoint-4500/rng_state.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cc5c1efc1bb823974a380eefa21fef1c92f1a4a0de8dd3c95b3899c696def0ee
|
3 |
+
size 14503
|
sagorsarker/bangla-bert-base/checkpoint-4500/scaler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8b3a01745698088bd6b17367f8985f3608b477f953bdd7fc29ca0e690cad736e
|
3 |
+
size 559
|
sagorsarker/bangla-bert-base/checkpoint-4500/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9c5fd1086620b85558b5b18e6210e97dbe97069519f3ee419d192e55b598ca6a
|
3 |
+
size 623
|
sagorsarker/bangla-bert-base/checkpoint-4500/special_tokens_map.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
|
sagorsarker/bangla-bert-base/checkpoint-4500/tokenizer_config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"do_lower_case": true, "do_basic_tokenize": true, "never_split": null, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "special_tokens_map_file": null, "tokenizer_file": null, "name_or_path": "sagorsarker/bangla-bert-base", "tokenizer_class": "BertTokenizer"}
|
sagorsarker/bangla-bert-base/checkpoint-4500/trainer_state.json
ADDED
@@ -0,0 +1,136 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"best_metric": 0.6700201207243461,
|
3 |
+
"best_model_checkpoint": "./clickbait_identification/sagorsarker/bangla-bert-base/checkpoint-4500",
|
4 |
+
"epoch": 2.9742233972240584,
|
5 |
+
"global_step": 4500,
|
6 |
+
"is_hyper_param_search": false,
|
7 |
+
"is_local_process_zero": true,
|
8 |
+
"is_world_process_zero": true,
|
9 |
+
"log_history": [
|
10 |
+
{
|
11 |
+
"epoch": 0.33,
|
12 |
+
"eval_accuracy": 0.7102526002971769,
|
13 |
+
"eval_f1": 0.4213649851632047,
|
14 |
+
"eval_loss": 0.6094053983688354,
|
15 |
+
"eval_precision": 0.6794258373205742,
|
16 |
+
"eval_recall": 0.3053763440860215,
|
17 |
+
"eval_runtime": 33.7164,
|
18 |
+
"eval_samples_per_second": 39.921,
|
19 |
+
"eval_steps_per_second": 5.012,
|
20 |
+
"step": 500
|
21 |
+
},
|
22 |
+
{
|
23 |
+
"epoch": 0.66,
|
24 |
+
"eval_accuracy": 0.7533432392273403,
|
25 |
+
"eval_f1": 0.5891089108910892,
|
26 |
+
"eval_loss": 0.4936341941356659,
|
27 |
+
"eval_precision": 0.6938775510204082,
|
28 |
+
"eval_recall": 0.5118279569892473,
|
29 |
+
"eval_runtime": 29.617,
|
30 |
+
"eval_samples_per_second": 45.447,
|
31 |
+
"eval_steps_per_second": 5.706,
|
32 |
+
"step": 1000
|
33 |
+
},
|
34 |
+
{
|
35 |
+
"epoch": 0.99,
|
36 |
+
"eval_accuracy": 0.736998514115899,
|
37 |
+
"eval_f1": 0.45370370370370366,
|
38 |
+
"eval_loss": 0.5591641664505005,
|
39 |
+
"eval_precision": 0.8032786885245902,
|
40 |
+
"eval_recall": 0.3161290322580645,
|
41 |
+
"eval_runtime": 29.6259,
|
42 |
+
"eval_samples_per_second": 45.433,
|
43 |
+
"eval_steps_per_second": 5.704,
|
44 |
+
"step": 1500
|
45 |
+
},
|
46 |
+
{
|
47 |
+
"epoch": 1.0,
|
48 |
+
"learning_rate": 1.900066093853272e-05,
|
49 |
+
"loss": 0.5818,
|
50 |
+
"step": 1513
|
51 |
+
},
|
52 |
+
{
|
53 |
+
"epoch": 1.32,
|
54 |
+
"eval_accuracy": 0.673848439821694,
|
55 |
+
"eval_f1": 0.6344712739383848,
|
56 |
+
"eval_loss": 0.623744010925293,
|
57 |
+
"eval_precision": 0.5176630434782609,
|
58 |
+
"eval_recall": 0.8193548387096774,
|
59 |
+
"eval_runtime": 29.6166,
|
60 |
+
"eval_samples_per_second": 45.448,
|
61 |
+
"eval_steps_per_second": 5.706,
|
62 |
+
"step": 2000
|
63 |
+
},
|
64 |
+
{
|
65 |
+
"epoch": 1.65,
|
66 |
+
"eval_accuracy": 0.7682020802377415,
|
67 |
+
"eval_f1": 0.6320754716981133,
|
68 |
+
"eval_loss": 0.6347715258598328,
|
69 |
+
"eval_precision": 0.6997389033942559,
|
70 |
+
"eval_recall": 0.5763440860215053,
|
71 |
+
"eval_runtime": 30.3538,
|
72 |
+
"eval_samples_per_second": 44.344,
|
73 |
+
"eval_steps_per_second": 5.568,
|
74 |
+
"step": 2500
|
75 |
+
},
|
76 |
+
{
|
77 |
+
"epoch": 1.98,
|
78 |
+
"eval_accuracy": 0.6820208023774146,
|
79 |
+
"eval_f1": 0.6491803278688525,
|
80 |
+
"eval_loss": 0.8944938778877258,
|
81 |
+
"eval_precision": 0.5245033112582781,
|
82 |
+
"eval_recall": 0.8516129032258064,
|
83 |
+
"eval_runtime": 29.7011,
|
84 |
+
"eval_samples_per_second": 45.318,
|
85 |
+
"eval_steps_per_second": 5.69,
|
86 |
+
"step": 3000
|
87 |
+
},
|
88 |
+
{
|
89 |
+
"epoch": 2.0,
|
90 |
+
"learning_rate": 1.8001321877065433e-05,
|
91 |
+
"loss": 0.486,
|
92 |
+
"step": 3026
|
93 |
+
},
|
94 |
+
{
|
95 |
+
"epoch": 2.31,
|
96 |
+
"eval_accuracy": 0.7592867756315007,
|
97 |
+
"eval_f1": 0.6326530612244898,
|
98 |
+
"eval_loss": 0.5959650278091431,
|
99 |
+
"eval_precision": 0.6690647482014388,
|
100 |
+
"eval_recall": 0.6,
|
101 |
+
"eval_runtime": 29.6182,
|
102 |
+
"eval_samples_per_second": 45.445,
|
103 |
+
"eval_steps_per_second": 5.706,
|
104 |
+
"step": 3500
|
105 |
+
},
|
106 |
+
{
|
107 |
+
"epoch": 2.64,
|
108 |
+
"eval_accuracy": 0.7459138187221397,
|
109 |
+
"eval_f1": 0.6131221719457014,
|
110 |
+
"eval_loss": 0.729725182056427,
|
111 |
+
"eval_precision": 0.6467780429594272,
|
112 |
+
"eval_recall": 0.5827956989247312,
|
113 |
+
"eval_runtime": 29.6047,
|
114 |
+
"eval_samples_per_second": 45.466,
|
115 |
+
"eval_steps_per_second": 5.709,
|
116 |
+
"step": 4000
|
117 |
+
},
|
118 |
+
{
|
119 |
+
"epoch": 2.97,
|
120 |
+
"eval_accuracy": 0.7563150074294205,
|
121 |
+
"eval_f1": 0.6700201207243461,
|
122 |
+
"eval_loss": 0.6430042386054993,
|
123 |
+
"eval_precision": 0.6294896030245747,
|
124 |
+
"eval_recall": 0.7161290322580646,
|
125 |
+
"eval_runtime": 29.6897,
|
126 |
+
"eval_samples_per_second": 45.336,
|
127 |
+
"eval_steps_per_second": 5.692,
|
128 |
+
"step": 4500
|
129 |
+
}
|
130 |
+
],
|
131 |
+
"max_steps": 30260,
|
132 |
+
"num_train_epochs": 20,
|
133 |
+
"total_flos": 0.0,
|
134 |
+
"trial_name": null,
|
135 |
+
"trial_params": null
|
136 |
+
}
|
sagorsarker/bangla-bert-base/checkpoint-4500/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b9a6d68675bed1173a6675b1852aa0d3310315b37538cd0020fd6eb46cc03090
|
3 |
+
size 3055
|
sagorsarker/bangla-bert-base/checkpoint-4500/vocab.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
sagorsarker/bangla-bert-base/runs/Jan13_14-45-35_22ece1fdce42/1642085147.7494185/events.out.tfevents.1642085147.22ece1fdce42.7156.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3704cfe6ade6a4c403126c6af29c3ca05400bfb3a42d73196b99c1a9ae7daf30
|
3 |
+
size 4845
|
sagorsarker/bangla-bert-base/runs/Jan13_14-45-35_22ece1fdce42/events.out.tfevents.1642085147.22ece1fdce42.7156.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e8813a62f53192cf5909cd78485cb04ab17893990c2ceda351ebec1ab36490f4
|
3 |
+
size 9308
|