Commit
·
02b20f6
1
Parent(s):
e15e7f4
f[neuralspace-reverie/indic-transformers-bn-bert, neuralspace-reverie/indic-transformers-bn-roberta, distilbert-base-multilingual-cased, neuralspace-reverie/indic-transformers-bn-distilbert, monsoon-nlp/bangla-electra, csebuetnlp/banglabert, xlm-roberta-large, neuralspace-reverie/indic-transformers-bn-xlmroberta]
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- csebuetnlp/banglabert/all_results.json +12 -0
- csebuetnlp/banglabert/checkpoint-6000/optimizer.pt +3 -0
- csebuetnlp/banglabert/checkpoint-6000/pytorch_model.bin +3 -0
- {neuralspace-reverie/indic-transformers-bn-bert/checkpoint-9000 → csebuetnlp/banglabert/checkpoint-6000}/rng_state.pth +1 -1
- neuralspace-reverie/indic-transformers-bn-bert/runs/Jan13_22-23-10_e3061b854d02/1642112608.4571896/events.out.tfevents.1642112608.e3061b854d02.83.1 → csebuetnlp/banglabert/checkpoint-6000/scheduler.pt +2 -2
- {neuralspace-reverie/indic-transformers-bn-bert/checkpoint-9000 → csebuetnlp/banglabert/checkpoint-6000}/special_tokens_map.json +0 -0
- csebuetnlp/banglabert/checkpoint-6000/tokenizer.json +0 -0
- csebuetnlp/banglabert/checkpoint-6000/tokenizer_config.json +1 -0
- csebuetnlp/banglabert/checkpoint-6000/trainer_state.json +94 -0
- {neuralspace-reverie/indic-transformers-bn-bert/checkpoint-9000 → csebuetnlp/banglabert/checkpoint-6000}/training_args.bin +2 -2
- csebuetnlp/banglabert/checkpoint-6000/vocab.txt +0 -0
- csebuetnlp/banglabert/eval_results.json +12 -0
- csebuetnlp/banglabert/pytorch_model.bin +3 -0
- neuralspace-reverie/indic-transformers-bn-bert/runs/Jan13_22-23-10_e3061b854d02/events.out.tfevents.1642112608.e3061b854d02.83.0 → csebuetnlp/banglabert/runs/Jan14_10-42-02_4ac0fd88951a/1642156936.1436265/events.out.tfevents.1642156936.4ac0fd88951a.2300.11 +2 -2
- csebuetnlp/banglabert/runs/Jan14_10-42-02_4ac0fd88951a/events.out.tfevents.1642156936.4ac0fd88951a.2300.10 +3 -0
- csebuetnlp/banglabert/special_tokens_map.json +1 -0
- csebuetnlp/banglabert/tokenizer.json +0 -0
- csebuetnlp/banglabert/tokenizer_config.json +1 -0
- csebuetnlp/banglabert/training_args.bin +3 -0
- csebuetnlp/banglabert/vocab.txt +0 -0
- distilbert-base-multilingual-cased/all_results.json +12 -0
- distilbert-base-multilingual-cased/checkpoint-13000/optimizer.pt +3 -0
- distilbert-base-multilingual-cased/checkpoint-13000/pytorch_model.bin +3 -0
- distilbert-base-multilingual-cased/checkpoint-13000/rng_state.pth +3 -0
- distilbert-base-multilingual-cased/checkpoint-13000/scheduler.pt +3 -0
- distilbert-base-multilingual-cased/checkpoint-13000/special_tokens_map.json +1 -0
- distilbert-base-multilingual-cased/checkpoint-13000/tokenizer.json +0 -0
- distilbert-base-multilingual-cased/checkpoint-13000/tokenizer_config.json +1 -0
- distilbert-base-multilingual-cased/checkpoint-13000/trainer_state.json +196 -0
- distilbert-base-multilingual-cased/checkpoint-13000/training_args.bin +3 -0
- distilbert-base-multilingual-cased/checkpoint-13000/vocab.txt +0 -0
- distilbert-base-multilingual-cased/eval_results.json +12 -0
- distilbert-base-multilingual-cased/pytorch_model.bin +3 -0
- distilbert-base-multilingual-cased/runs/Jan14_09-24-05_4ac0fd88951a/1642152252.538074/events.out.tfevents.1642152252.4ac0fd88951a.2300.5 +3 -0
- distilbert-base-multilingual-cased/runs/Jan14_09-24-05_4ac0fd88951a/events.out.tfevents.1642152252.4ac0fd88951a.2300.4 +3 -0
- distilbert-base-multilingual-cased/special_tokens_map.json +1 -0
- distilbert-base-multilingual-cased/tokenizer.json +0 -0
- distilbert-base-multilingual-cased/tokenizer_config.json +1 -0
- distilbert-base-multilingual-cased/training_args.bin +3 -0
- distilbert-base-multilingual-cased/vocab.txt +0 -0
- final_result.csv +9 -0
- monsoon-nlp/bangla-electra/all_results.json +12 -0
- monsoon-nlp/bangla-electra/checkpoint-9000/optimizer.pt +3 -0
- monsoon-nlp/bangla-electra/checkpoint-9000/pytorch_model.bin +3 -0
- monsoon-nlp/bangla-electra/checkpoint-9000/rng_state.pth +3 -0
- {neuralspace-reverie/indic-transformers-bn-bert → monsoon-nlp/bangla-electra}/checkpoint-9000/scheduler.pt +0 -0
- monsoon-nlp/bangla-electra/checkpoint-9000/special_tokens_map.json +1 -0
- monsoon-nlp/bangla-electra/checkpoint-9000/tokenizer.json +0 -0
- monsoon-nlp/bangla-electra/checkpoint-9000/tokenizer_config.json +1 -0
- monsoon-nlp/bangla-electra/checkpoint-9000/trainer_state.json +136 -0
csebuetnlp/banglabert/all_results.json
ADDED
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"epoch": 0.33,
|
3 |
+
"eval_samples": 1346,
|
4 |
+
"test_accuracy": 0.7838038632986627,
|
5 |
+
"test_f1": 0.7063572149344096,
|
6 |
+
"test_loss": 0.6481547951698303,
|
7 |
+
"test_precision": 0.6653992395437263,
|
8 |
+
"test_recall": 0.7526881720430108,
|
9 |
+
"test_runtime": 25.7216,
|
10 |
+
"test_samples_per_second": 52.33,
|
11 |
+
"test_steps_per_second": 13.102
|
12 |
+
}
|
csebuetnlp/banglabert/checkpoint-6000/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4258268f20e9b4adc92b50d325444ca5fe0321345b4d5ca0e5427d72de9581e2
|
3 |
+
size 880341733
|
csebuetnlp/banglabert/checkpoint-6000/pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4a3ddeaed8431e6cfc5e1bcee875048299cfd136dee17d09155643129db7368f
|
3 |
+
size 440196977
|
{neuralspace-reverie/indic-transformers-bn-bert/checkpoint-9000 → csebuetnlp/banglabert/checkpoint-6000}/rng_state.pth
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14503
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:aa114000851e2597418582b2934c1866379c0f0a17955b4b0adb8118903acaa1
|
3 |
size 14503
|
neuralspace-reverie/indic-transformers-bn-bert/runs/Jan13_22-23-10_e3061b854d02/1642112608.4571896/events.out.tfevents.1642112608.e3061b854d02.83.1 → csebuetnlp/banglabert/checkpoint-6000/scheduler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5aed322968a4c02e3f947148f035a35d2206eb8fcc1753d1bf64eb8bfb95083c
|
3 |
+
size 623
|
{neuralspace-reverie/indic-transformers-bn-bert/checkpoint-9000 → csebuetnlp/banglabert/checkpoint-6000}/special_tokens_map.json
RENAMED
File without changes
|
csebuetnlp/banglabert/checkpoint-6000/tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
csebuetnlp/banglabert/checkpoint-6000/tokenizer_config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"do_lower_case": false, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": false, "strip_accents": null, "special_tokens_map_file": null, "full_tokenizer_file": null, "name_or_path": "csebuetnlp/banglabert", "do_basic_tokenize": true, "never_split": null, "tokenizer_class": "ElectraTokenizer"}
|
csebuetnlp/banglabert/checkpoint-6000/trainer_state.json
ADDED
@@ -0,0 +1,94 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"best_metric": 0.7063572149344096,
|
3 |
+
"best_model_checkpoint": "./clickbait_identification/csebuetnlp/banglabert/checkpoint-6000",
|
4 |
+
"epoch": 1.9828155981493722,
|
5 |
+
"global_step": 6000,
|
6 |
+
"is_hyper_param_search": false,
|
7 |
+
"is_local_process_zero": true,
|
8 |
+
"is_world_process_zero": true,
|
9 |
+
"log_history": [
|
10 |
+
{
|
11 |
+
"epoch": 0.33,
|
12 |
+
"eval_accuracy": 0.7013372956909361,
|
13 |
+
"eval_f1": 0.32094594594594594,
|
14 |
+
"eval_loss": 0.5926413536071777,
|
15 |
+
"eval_precision": 0.7480314960629921,
|
16 |
+
"eval_recall": 0.20430107526881722,
|
17 |
+
"eval_runtime": 25.7194,
|
18 |
+
"eval_samples_per_second": 52.334,
|
19 |
+
"eval_steps_per_second": 13.103,
|
20 |
+
"step": 1000
|
21 |
+
},
|
22 |
+
{
|
23 |
+
"epoch": 0.66,
|
24 |
+
"eval_accuracy": 0.7332838038632986,
|
25 |
+
"eval_f1": 0.4665676077265974,
|
26 |
+
"eval_loss": 0.514832079410553,
|
27 |
+
"eval_precision": 0.7548076923076923,
|
28 |
+
"eval_recall": 0.33763440860215055,
|
29 |
+
"eval_runtime": 25.719,
|
30 |
+
"eval_samples_per_second": 52.335,
|
31 |
+
"eval_steps_per_second": 13.103,
|
32 |
+
"step": 2000
|
33 |
+
},
|
34 |
+
{
|
35 |
+
"epoch": 0.99,
|
36 |
+
"eval_accuracy": 0.7540861812778603,
|
37 |
+
"eval_f1": 0.6814244465832531,
|
38 |
+
"eval_loss": 0.49710002541542053,
|
39 |
+
"eval_precision": 0.6167247386759582,
|
40 |
+
"eval_recall": 0.7612903225806451,
|
41 |
+
"eval_runtime": 25.7279,
|
42 |
+
"eval_samples_per_second": 52.317,
|
43 |
+
"eval_steps_per_second": 13.099,
|
44 |
+
"step": 3000
|
45 |
+
},
|
46 |
+
{
|
47 |
+
"epoch": 1.0,
|
48 |
+
"learning_rate": 1.9e-05,
|
49 |
+
"loss": 0.58,
|
50 |
+
"step": 3026
|
51 |
+
},
|
52 |
+
{
|
53 |
+
"epoch": 1.32,
|
54 |
+
"eval_accuracy": 0.7674591381872214,
|
55 |
+
"eval_f1": 0.6916256157635469,
|
56 |
+
"eval_loss": 0.628989577293396,
|
57 |
+
"eval_precision": 0.6381818181818182,
|
58 |
+
"eval_recall": 0.7548387096774194,
|
59 |
+
"eval_runtime": 25.7155,
|
60 |
+
"eval_samples_per_second": 52.342,
|
61 |
+
"eval_steps_per_second": 13.105,
|
62 |
+
"step": 4000
|
63 |
+
},
|
64 |
+
{
|
65 |
+
"epoch": 1.65,
|
66 |
+
"eval_accuracy": 0.7867756315007429,
|
67 |
+
"eval_f1": 0.6849615806805708,
|
68 |
+
"eval_loss": 0.6614874005317688,
|
69 |
+
"eval_precision": 0.6995515695067265,
|
70 |
+
"eval_recall": 0.6709677419354839,
|
71 |
+
"eval_runtime": 26.0668,
|
72 |
+
"eval_samples_per_second": 51.636,
|
73 |
+
"eval_steps_per_second": 12.928,
|
74 |
+
"step": 5000
|
75 |
+
},
|
76 |
+
{
|
77 |
+
"epoch": 1.98,
|
78 |
+
"eval_accuracy": 0.7838038632986627,
|
79 |
+
"eval_f1": 0.7063572149344096,
|
80 |
+
"eval_loss": 0.6481547951698303,
|
81 |
+
"eval_precision": 0.6653992395437263,
|
82 |
+
"eval_recall": 0.7526881720430108,
|
83 |
+
"eval_runtime": 25.9916,
|
84 |
+
"eval_samples_per_second": 51.786,
|
85 |
+
"eval_steps_per_second": 12.966,
|
86 |
+
"step": 6000
|
87 |
+
}
|
88 |
+
],
|
89 |
+
"max_steps": 60520,
|
90 |
+
"num_train_epochs": 20,
|
91 |
+
"total_flos": 0.0,
|
92 |
+
"trial_name": null,
|
93 |
+
"trial_params": null
|
94 |
+
}
|
{neuralspace-reverie/indic-transformers-bn-bert/checkpoint-9000 → csebuetnlp/banglabert/checkpoint-6000}/training_args.bin
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f2e5952205fe295833f75995454e9b8718a4b278409827724e27ec10f8788ff1
|
3 |
+
size 2991
|
csebuetnlp/banglabert/checkpoint-6000/vocab.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
csebuetnlp/banglabert/eval_results.json
ADDED
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"epoch": 0.33,
|
3 |
+
"eval_samples": 1346,
|
4 |
+
"test_accuracy": 0.7838038632986627,
|
5 |
+
"test_f1": 0.7063572149344096,
|
6 |
+
"test_loss": 0.6481547951698303,
|
7 |
+
"test_precision": 0.6653992395437263,
|
8 |
+
"test_recall": 0.7526881720430108,
|
9 |
+
"test_runtime": 25.7216,
|
10 |
+
"test_samples_per_second": 52.33,
|
11 |
+
"test_steps_per_second": 13.102
|
12 |
+
}
|
csebuetnlp/banglabert/pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4a3ddeaed8431e6cfc5e1bcee875048299cfd136dee17d09155643129db7368f
|
3 |
+
size 440196977
|
neuralspace-reverie/indic-transformers-bn-bert/runs/Jan13_22-23-10_e3061b854d02/events.out.tfevents.1642112608.e3061b854d02.83.0 → csebuetnlp/banglabert/runs/Jan14_10-42-02_4ac0fd88951a/1642156936.1436265/events.out.tfevents.1642156936.4ac0fd88951a.2300.11
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8e0a8c741725d06cfc25ed9dc39923be0d12db4000e929af0494538dc43972ac
|
3 |
+
size 4830
|
csebuetnlp/banglabert/runs/Jan14_10-42-02_4ac0fd88951a/events.out.tfevents.1642156936.4ac0fd88951a.2300.10
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:816f0b2be9728594b1d5331f395bfa66a60d2b3081d736d2d0ef004986fba1bf
|
3 |
+
size 7723
|
csebuetnlp/banglabert/special_tokens_map.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
|
csebuetnlp/banglabert/tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
csebuetnlp/banglabert/tokenizer_config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"do_lower_case": false, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": false, "strip_accents": null, "special_tokens_map_file": null, "full_tokenizer_file": null, "name_or_path": "csebuetnlp/banglabert", "do_basic_tokenize": true, "never_split": null, "tokenizer_class": "ElectraTokenizer"}
|
csebuetnlp/banglabert/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f2e5952205fe295833f75995454e9b8718a4b278409827724e27ec10f8788ff1
|
3 |
+
size 2991
|
csebuetnlp/banglabert/vocab.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
distilbert-base-multilingual-cased/all_results.json
ADDED
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"epoch": 0.33,
|
3 |
+
"eval_samples": 1346,
|
4 |
+
"test_accuracy": 0.7578008915304606,
|
5 |
+
"test_f1": 0.6632231404958677,
|
6 |
+
"test_loss": 0.7381333708763123,
|
7 |
+
"test_precision": 0.6381709741550696,
|
8 |
+
"test_recall": 0.6903225806451613,
|
9 |
+
"test_runtime": 15.8433,
|
10 |
+
"test_samples_per_second": 84.957,
|
11 |
+
"test_steps_per_second": 21.271
|
12 |
+
}
|
distilbert-base-multilingual-cased/checkpoint-13000/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:46fad17c165fb4906369b88a570c6ccd372d38d309a685861be52b97b275f620
|
3 |
+
size 1077943721
|
distilbert-base-multilingual-cased/checkpoint-13000/pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:025a283d17baff317ddb014ac0edf7d928cf174f5fd7adc0e92a0d36fd8d0bd8
|
3 |
+
size 538980917
|
distilbert-base-multilingual-cased/checkpoint-13000/rng_state.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8603747db40b433ca7b897885b152f245f1673f1691cecdc7eb027c0cc359691
|
3 |
+
size 14503
|
distilbert-base-multilingual-cased/checkpoint-13000/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7cebc06887b1ea5796f15d5311dd791a8f4bf296812f347ff6f02ef2b6525a8b
|
3 |
+
size 623
|
distilbert-base-multilingual-cased/checkpoint-13000/special_tokens_map.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
|
distilbert-base-multilingual-cased/checkpoint-13000/tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
distilbert-base-multilingual-cased/checkpoint-13000/tokenizer_config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"do_lower_case": false, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "distilbert-base-multilingual-cased", "tokenizer_class": "DistilBertTokenizer"}
|
distilbert-base-multilingual-cased/checkpoint-13000/trainer_state.json
ADDED
@@ -0,0 +1,196 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"best_metric": 0.6632231404958677,
|
3 |
+
"best_model_checkpoint": "./clickbait_identification/distilbert-base-multilingual-cased/checkpoint-13000",
|
4 |
+
"epoch": 4.296100462656973,
|
5 |
+
"global_step": 13000,
|
6 |
+
"is_hyper_param_search": false,
|
7 |
+
"is_local_process_zero": true,
|
8 |
+
"is_world_process_zero": true,
|
9 |
+
"log_history": [
|
10 |
+
{
|
11 |
+
"epoch": 0.33,
|
12 |
+
"eval_accuracy": 0.6909361069836553,
|
13 |
+
"eval_f1": 0.3396825396825397,
|
14 |
+
"eval_loss": 0.598863959312439,
|
15 |
+
"eval_precision": 0.6484848484848484,
|
16 |
+
"eval_recall": 0.23010752688172043,
|
17 |
+
"eval_runtime": 15.8265,
|
18 |
+
"eval_samples_per_second": 85.047,
|
19 |
+
"eval_steps_per_second": 21.293,
|
20 |
+
"step": 1000
|
21 |
+
},
|
22 |
+
{
|
23 |
+
"epoch": 0.66,
|
24 |
+
"eval_accuracy": 0.7139673105497771,
|
25 |
+
"eval_f1": 0.48598130841121495,
|
26 |
+
"eval_loss": 0.5456291437149048,
|
27 |
+
"eval_precision": 0.6408450704225352,
|
28 |
+
"eval_recall": 0.3913978494623656,
|
29 |
+
"eval_runtime": 15.7981,
|
30 |
+
"eval_samples_per_second": 85.2,
|
31 |
+
"eval_steps_per_second": 21.332,
|
32 |
+
"step": 2000
|
33 |
+
},
|
34 |
+
{
|
35 |
+
"epoch": 0.99,
|
36 |
+
"eval_accuracy": 0.736998514115899,
|
37 |
+
"eval_f1": 0.5378590078328981,
|
38 |
+
"eval_loss": 0.601733386516571,
|
39 |
+
"eval_precision": 0.6843853820598007,
|
40 |
+
"eval_recall": 0.443010752688172,
|
41 |
+
"eval_runtime": 16.1605,
|
42 |
+
"eval_samples_per_second": 83.29,
|
43 |
+
"eval_steps_per_second": 20.853,
|
44 |
+
"step": 3000
|
45 |
+
},
|
46 |
+
{
|
47 |
+
"epoch": 1.0,
|
48 |
+
"learning_rate": 1.9e-05,
|
49 |
+
"loss": 0.6042,
|
50 |
+
"step": 3026
|
51 |
+
},
|
52 |
+
{
|
53 |
+
"epoch": 1.32,
|
54 |
+
"eval_accuracy": 0.7288261515601783,
|
55 |
+
"eval_f1": 0.6178010471204188,
|
56 |
+
"eval_loss": 0.5779110789299011,
|
57 |
+
"eval_precision": 0.6020408163265306,
|
58 |
+
"eval_recall": 0.6344086021505376,
|
59 |
+
"eval_runtime": 15.9992,
|
60 |
+
"eval_samples_per_second": 84.129,
|
61 |
+
"eval_steps_per_second": 21.064,
|
62 |
+
"step": 4000
|
63 |
+
},
|
64 |
+
{
|
65 |
+
"epoch": 1.65,
|
66 |
+
"eval_accuracy": 0.7288261515601783,
|
67 |
+
"eval_f1": 0.5866364665911664,
|
68 |
+
"eval_loss": 0.5593090057373047,
|
69 |
+
"eval_precision": 0.6196172248803827,
|
70 |
+
"eval_recall": 0.556989247311828,
|
71 |
+
"eval_runtime": 16.0298,
|
72 |
+
"eval_samples_per_second": 83.968,
|
73 |
+
"eval_steps_per_second": 21.023,
|
74 |
+
"step": 5000
|
75 |
+
},
|
76 |
+
{
|
77 |
+
"epoch": 1.98,
|
78 |
+
"eval_accuracy": 0.7347696879643388,
|
79 |
+
"eval_f1": 0.6375634517766496,
|
80 |
+
"eval_loss": 0.6610430479049683,
|
81 |
+
"eval_precision": 0.6038461538461538,
|
82 |
+
"eval_recall": 0.6752688172043011,
|
83 |
+
"eval_runtime": 15.807,
|
84 |
+
"eval_samples_per_second": 85.152,
|
85 |
+
"eval_steps_per_second": 21.32,
|
86 |
+
"step": 6000
|
87 |
+
},
|
88 |
+
{
|
89 |
+
"epoch": 2.0,
|
90 |
+
"learning_rate": 1.8e-05,
|
91 |
+
"loss": 0.5693,
|
92 |
+
"step": 6052
|
93 |
+
},
|
94 |
+
{
|
95 |
+
"epoch": 2.31,
|
96 |
+
"eval_accuracy": 0.7421991084695394,
|
97 |
+
"eval_f1": 0.5510996119016817,
|
98 |
+
"eval_loss": 0.7018317580223083,
|
99 |
+
"eval_precision": 0.6915584415584416,
|
100 |
+
"eval_recall": 0.45806451612903226,
|
101 |
+
"eval_runtime": 15.8126,
|
102 |
+
"eval_samples_per_second": 85.122,
|
103 |
+
"eval_steps_per_second": 21.312,
|
104 |
+
"step": 7000
|
105 |
+
},
|
106 |
+
{
|
107 |
+
"epoch": 2.64,
|
108 |
+
"eval_accuracy": 0.7488855869242199,
|
109 |
+
"eval_f1": 0.6434599156118143,
|
110 |
+
"eval_loss": 0.6016787886619568,
|
111 |
+
"eval_precision": 0.6314699792960663,
|
112 |
+
"eval_recall": 0.6559139784946236,
|
113 |
+
"eval_runtime": 16.1537,
|
114 |
+
"eval_samples_per_second": 83.324,
|
115 |
+
"eval_steps_per_second": 20.862,
|
116 |
+
"step": 8000
|
117 |
+
},
|
118 |
+
{
|
119 |
+
"epoch": 2.97,
|
120 |
+
"eval_accuracy": 0.7473997028231798,
|
121 |
+
"eval_f1": 0.6473029045643153,
|
122 |
+
"eval_loss": 0.5445184707641602,
|
123 |
+
"eval_precision": 0.625250501002004,
|
124 |
+
"eval_recall": 0.6709677419354839,
|
125 |
+
"eval_runtime": 15.9994,
|
126 |
+
"eval_samples_per_second": 84.128,
|
127 |
+
"eval_steps_per_second": 21.063,
|
128 |
+
"step": 9000
|
129 |
+
},
|
130 |
+
{
|
131 |
+
"epoch": 3.0,
|
132 |
+
"learning_rate": 1.7e-05,
|
133 |
+
"loss": 0.5564,
|
134 |
+
"step": 9078
|
135 |
+
},
|
136 |
+
{
|
137 |
+
"epoch": 3.3,
|
138 |
+
"eval_accuracy": 0.7243684992570579,
|
139 |
+
"eval_f1": 0.6522961574507966,
|
140 |
+
"eval_loss": 0.6023839116096497,
|
141 |
+
"eval_precision": 0.5780730897009967,
|
142 |
+
"eval_recall": 0.7483870967741936,
|
143 |
+
"eval_runtime": 15.9931,
|
144 |
+
"eval_samples_per_second": 84.161,
|
145 |
+
"eval_steps_per_second": 21.072,
|
146 |
+
"step": 10000
|
147 |
+
},
|
148 |
+
{
|
149 |
+
"epoch": 3.64,
|
150 |
+
"eval_accuracy": 0.7147102526002972,
|
151 |
+
"eval_f1": 0.6489945155393052,
|
152 |
+
"eval_loss": 0.6206891536712646,
|
153 |
+
"eval_precision": 0.5643879173290938,
|
154 |
+
"eval_recall": 0.7634408602150538,
|
155 |
+
"eval_runtime": 15.7964,
|
156 |
+
"eval_samples_per_second": 85.21,
|
157 |
+
"eval_steps_per_second": 21.334,
|
158 |
+
"step": 11000
|
159 |
+
},
|
160 |
+
{
|
161 |
+
"epoch": 3.97,
|
162 |
+
"eval_accuracy": 0.7473997028231798,
|
163 |
+
"eval_f1": 0.6620278330019881,
|
164 |
+
"eval_loss": 0.6622892618179321,
|
165 |
+
"eval_precision": 0.6155268022181146,
|
166 |
+
"eval_recall": 0.7161290322580646,
|
167 |
+
"eval_runtime": 15.8234,
|
168 |
+
"eval_samples_per_second": 85.064,
|
169 |
+
"eval_steps_per_second": 21.298,
|
170 |
+
"step": 12000
|
171 |
+
},
|
172 |
+
{
|
173 |
+
"epoch": 4.0,
|
174 |
+
"learning_rate": 1.6000000000000003e-05,
|
175 |
+
"loss": 0.5472,
|
176 |
+
"step": 12104
|
177 |
+
},
|
178 |
+
{
|
179 |
+
"epoch": 4.3,
|
180 |
+
"eval_accuracy": 0.7578008915304606,
|
181 |
+
"eval_f1": 0.6632231404958677,
|
182 |
+
"eval_loss": 0.7381333708763123,
|
183 |
+
"eval_precision": 0.6381709741550696,
|
184 |
+
"eval_recall": 0.6903225806451613,
|
185 |
+
"eval_runtime": 16.1759,
|
186 |
+
"eval_samples_per_second": 83.21,
|
187 |
+
"eval_steps_per_second": 20.833,
|
188 |
+
"step": 13000
|
189 |
+
}
|
190 |
+
],
|
191 |
+
"max_steps": 60520,
|
192 |
+
"num_train_epochs": 20,
|
193 |
+
"total_flos": 0.0,
|
194 |
+
"trial_name": null,
|
195 |
+
"trial_params": null
|
196 |
+
}
|
distilbert-base-multilingual-cased/checkpoint-13000/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:47d830c2782ebe8b10368751d6867c1f878d2df28a7ef85e6f2b2eb6962848f9
|
3 |
+
size 3055
|
distilbert-base-multilingual-cased/checkpoint-13000/vocab.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
distilbert-base-multilingual-cased/eval_results.json
ADDED
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"epoch": 0.33,
|
3 |
+
"eval_samples": 1346,
|
4 |
+
"test_accuracy": 0.7578008915304606,
|
5 |
+
"test_f1": 0.6632231404958677,
|
6 |
+
"test_loss": 0.7381333708763123,
|
7 |
+
"test_precision": 0.6381709741550696,
|
8 |
+
"test_recall": 0.6903225806451613,
|
9 |
+
"test_runtime": 15.8433,
|
10 |
+
"test_samples_per_second": 84.957,
|
11 |
+
"test_steps_per_second": 21.271
|
12 |
+
}
|
distilbert-base-multilingual-cased/pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:025a283d17baff317ddb014ac0edf7d928cf174f5fd7adc0e92a0d36fd8d0bd8
|
3 |
+
size 538980917
|
distilbert-base-multilingual-cased/runs/Jan14_09-24-05_4ac0fd88951a/1642152252.538074/events.out.tfevents.1642152252.4ac0fd88951a.2300.5
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a3b21e093ded94909557bae23bfadddaeca3e3ccd10735ddce98eecbfb949996
|
3 |
+
size 4869
|
distilbert-base-multilingual-cased/runs/Jan14_09-24-05_4ac0fd88951a/events.out.tfevents.1642152252.4ac0fd88951a.2300.4
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:baaf0f9243e82c55cfc2d40c06965f419cea15217789e6b265073b39a38c759f
|
3 |
+
size 11537
|
distilbert-base-multilingual-cased/special_tokens_map.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
|
distilbert-base-multilingual-cased/tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
distilbert-base-multilingual-cased/tokenizer_config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"do_lower_case": false, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "distilbert-base-multilingual-cased", "tokenizer_class": "DistilBertTokenizer"}
|
distilbert-base-multilingual-cased/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:47d830c2782ebe8b10368751d6867c1f878d2df28a7ef85e6f2b2eb6962848f9
|
3 |
+
size 3055
|
distilbert-base-multilingual-cased/vocab.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
final_result.csv
ADDED
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
,train_epoch,test_accuracy,test_f1,test_precision,test_recall,test_loss
|
2 |
+
neuralspace-reverie/indic-transformers-bn-bert,0.33,0.7689450222882616,0.661588683351469,0.6696035242290749,0.6537634408602151,0.6427938938140869
|
3 |
+
neuralspace-reverie/indic-transformers-bn-roberta,0.33,0.7466567607726597,0.6827906976744187,0.601639344262295,0.789247311827957,0.5567601323127747
|
4 |
+
distilbert-base-multilingual-cased,0.33,0.7578008915304606,0.6632231404958677,0.6381709741550696,0.6903225806451613,0.7381333708763123
|
5 |
+
neuralspace-reverie/indic-transformers-bn-distilbert,0.33,0.7979197622585439,0.7213114754098361,0.6888454011741683,0.7569892473118279,0.7468417882919312
|
6 |
+
monsoon-nlp/bangla-electra,0.33,0.7414561664190193,0.6692015209125476,0.5996592844974447,0.7569892473118279,0.547999382019043
|
7 |
+
csebuetnlp/banglabert,0.33,0.7838038632986627,0.7063572149344096,0.6653992395437263,0.7526881720430108,0.6481547951698303
|
8 |
+
xlm-roberta-large,0.17,0.6545319465081724,0.0,0.0,0.0,0.6548592448234558
|
9 |
+
neuralspace-reverie/indic-transformers-bn-xlmroberta,0.17,0.7771173848439822,0.6781115879828327,0.6766595289079229,0.6795698924731183,1.1499747037887573
|
monsoon-nlp/bangla-electra/all_results.json
ADDED
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"epoch": 0.33,
|
3 |
+
"eval_samples": 1346,
|
4 |
+
"test_accuracy": 0.7414561664190193,
|
5 |
+
"test_f1": 0.6692015209125476,
|
6 |
+
"test_loss": 0.547999382019043,
|
7 |
+
"test_precision": 0.5996592844974447,
|
8 |
+
"test_recall": 0.7569892473118279,
|
9 |
+
"test_runtime": 6.8808,
|
10 |
+
"test_samples_per_second": 195.617,
|
11 |
+
"test_steps_per_second": 48.977
|
12 |
+
}
|
monsoon-nlp/bangla-electra/checkpoint-9000/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f160ff966f41851aed2bc2d11cbfb60bb5705631344c6f748c890325c48d59b7
|
3 |
+
size 107345693
|
monsoon-nlp/bangla-electra/checkpoint-9000/pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a08f51ae88a5ae3d20797e32400fe98bbb38ac1d17348fd0966bd36055f7a862
|
3 |
+
size 53699053
|
monsoon-nlp/bangla-electra/checkpoint-9000/rng_state.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fbf949d1440a036887b89511c3f2943839dda58cd7391e3d3032f221b696bc5f
|
3 |
+
size 14503
|
{neuralspace-reverie/indic-transformers-bn-bert → monsoon-nlp/bangla-electra}/checkpoint-9000/scheduler.pt
RENAMED
File without changes
|
monsoon-nlp/bangla-electra/checkpoint-9000/special_tokens_map.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
|
monsoon-nlp/bangla-electra/checkpoint-9000/tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
monsoon-nlp/bangla-electra/checkpoint-9000/tokenizer_config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"do_lower_case": false, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": false, "special_tokens_map_file": "/root/.cache/huggingface/transformers/03b59f5325d31d4b9c4899993115ee1540291ed877e0d2ff3fd4dff814ada400.dd8bd9bfd3664b530ea4e645105f557769387b3da9f79bdb55ed556bdd80611d", "name_or_path": "monsoon-nlp/bangla-electra", "do_basic_tokenize": true, "never_split": null, "tokenizer_class": "ElectraTokenizer"}
|
monsoon-nlp/bangla-electra/checkpoint-9000/trainer_state.json
ADDED
@@ -0,0 +1,136 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"best_metric": 0.6692015209125476,
|
3 |
+
"best_model_checkpoint": "./clickbait_identification/monsoon-nlp/bangla-electra/checkpoint-9000",
|
4 |
+
"epoch": 2.9742233972240584,
|
5 |
+
"global_step": 9000,
|
6 |
+
"is_hyper_param_search": false,
|
7 |
+
"is_local_process_zero": true,
|
8 |
+
"is_world_process_zero": true,
|
9 |
+
"log_history": [
|
10 |
+
{
|
11 |
+
"epoch": 0.33,
|
12 |
+
"eval_accuracy": 0.7013372956909361,
|
13 |
+
"eval_f1": 0.3116438356164384,
|
14 |
+
"eval_loss": 0.5911309123039246,
|
15 |
+
"eval_precision": 0.7647058823529411,
|
16 |
+
"eval_recall": 0.1956989247311828,
|
17 |
+
"eval_runtime": 6.9517,
|
18 |
+
"eval_samples_per_second": 193.622,
|
19 |
+
"eval_steps_per_second": 48.477,
|
20 |
+
"step": 1000
|
21 |
+
},
|
22 |
+
{
|
23 |
+
"epoch": 0.66,
|
24 |
+
"eval_accuracy": 0.711738484398217,
|
25 |
+
"eval_f1": 0.4947916666666667,
|
26 |
+
"eval_loss": 0.5567843317985535,
|
27 |
+
"eval_precision": 0.6270627062706271,
|
28 |
+
"eval_recall": 0.40860215053763443,
|
29 |
+
"eval_runtime": 6.8804,
|
30 |
+
"eval_samples_per_second": 195.628,
|
31 |
+
"eval_steps_per_second": 48.98,
|
32 |
+
"step": 2000
|
33 |
+
},
|
34 |
+
{
|
35 |
+
"epoch": 0.99,
|
36 |
+
"eval_accuracy": 0.6731054977711739,
|
37 |
+
"eval_f1": 0.608540925266904,
|
38 |
+
"eval_loss": 0.5933745503425598,
|
39 |
+
"eval_precision": 0.5189681335356601,
|
40 |
+
"eval_recall": 0.7354838709677419,
|
41 |
+
"eval_runtime": 6.8721,
|
42 |
+
"eval_samples_per_second": 195.863,
|
43 |
+
"eval_steps_per_second": 49.039,
|
44 |
+
"step": 3000
|
45 |
+
},
|
46 |
+
{
|
47 |
+
"epoch": 1.0,
|
48 |
+
"learning_rate": 1.9e-05,
|
49 |
+
"loss": 0.5987,
|
50 |
+
"step": 3026
|
51 |
+
},
|
52 |
+
{
|
53 |
+
"epoch": 1.32,
|
54 |
+
"eval_accuracy": 0.7325408618127786,
|
55 |
+
"eval_f1": 0.6086956521739132,
|
56 |
+
"eval_loss": 0.5164241790771484,
|
57 |
+
"eval_precision": 0.6153846153846154,
|
58 |
+
"eval_recall": 0.6021505376344086,
|
59 |
+
"eval_runtime": 6.8817,
|
60 |
+
"eval_samples_per_second": 195.59,
|
61 |
+
"eval_steps_per_second": 48.97,
|
62 |
+
"step": 4000
|
63 |
+
},
|
64 |
+
{
|
65 |
+
"epoch": 1.65,
|
66 |
+
"eval_accuracy": 0.7169390787518574,
|
67 |
+
"eval_f1": 0.6514181152790485,
|
68 |
+
"eval_loss": 0.5508525371551514,
|
69 |
+
"eval_precision": 0.5668789808917197,
|
70 |
+
"eval_recall": 0.7655913978494624,
|
71 |
+
"eval_runtime": 6.8787,
|
72 |
+
"eval_samples_per_second": 195.677,
|
73 |
+
"eval_steps_per_second": 48.992,
|
74 |
+
"step": 5000
|
75 |
+
},
|
76 |
+
{
|
77 |
+
"epoch": 1.98,
|
78 |
+
"eval_accuracy": 0.736998514115899,
|
79 |
+
"eval_f1": 0.6522593320235757,
|
80 |
+
"eval_loss": 0.5713074803352356,
|
81 |
+
"eval_precision": 0.6003616636528029,
|
82 |
+
"eval_recall": 0.7139784946236559,
|
83 |
+
"eval_runtime": 6.8865,
|
84 |
+
"eval_samples_per_second": 195.456,
|
85 |
+
"eval_steps_per_second": 48.937,
|
86 |
+
"step": 6000
|
87 |
+
},
|
88 |
+
{
|
89 |
+
"epoch": 2.0,
|
90 |
+
"learning_rate": 1.8e-05,
|
91 |
+
"loss": 0.5535,
|
92 |
+
"step": 6052
|
93 |
+
},
|
94 |
+
{
|
95 |
+
"epoch": 2.31,
|
96 |
+
"eval_accuracy": 0.7518573551263001,
|
97 |
+
"eval_f1": 0.6563786008230453,
|
98 |
+
"eval_loss": 0.5634607076644897,
|
99 |
+
"eval_precision": 0.6291913214990138,
|
100 |
+
"eval_recall": 0.6860215053763441,
|
101 |
+
"eval_runtime": 6.8807,
|
102 |
+
"eval_samples_per_second": 195.619,
|
103 |
+
"eval_steps_per_second": 48.977,
|
104 |
+
"step": 7000
|
105 |
+
},
|
106 |
+
{
|
107 |
+
"epoch": 2.64,
|
108 |
+
"eval_accuracy": 0.7273402674591382,
|
109 |
+
"eval_f1": 0.6654512306289881,
|
110 |
+
"eval_loss": 0.6494086980819702,
|
111 |
+
"eval_precision": 0.5775316455696202,
|
112 |
+
"eval_recall": 0.7849462365591398,
|
113 |
+
"eval_runtime": 6.878,
|
114 |
+
"eval_samples_per_second": 195.696,
|
115 |
+
"eval_steps_per_second": 48.997,
|
116 |
+
"step": 8000
|
117 |
+
},
|
118 |
+
{
|
119 |
+
"epoch": 2.97,
|
120 |
+
"eval_accuracy": 0.7414561664190193,
|
121 |
+
"eval_f1": 0.6692015209125476,
|
122 |
+
"eval_loss": 0.547999382019043,
|
123 |
+
"eval_precision": 0.5996592844974447,
|
124 |
+
"eval_recall": 0.7569892473118279,
|
125 |
+
"eval_runtime": 6.8738,
|
126 |
+
"eval_samples_per_second": 195.815,
|
127 |
+
"eval_steps_per_second": 49.027,
|
128 |
+
"step": 9000
|
129 |
+
}
|
130 |
+
],
|
131 |
+
"max_steps": 60520,
|
132 |
+
"num_train_epochs": 20,
|
133 |
+
"total_flos": 0.0,
|
134 |
+
"trial_name": null,
|
135 |
+
"trial_params": null
|
136 |
+
}
|