MotaharMahtabAnkon commited on
Commit
02b20f6
·
1 Parent(s): e15e7f4

f[neuralspace-reverie/indic-transformers-bn-bert, neuralspace-reverie/indic-transformers-bn-roberta, distilbert-base-multilingual-cased, neuralspace-reverie/indic-transformers-bn-distilbert, monsoon-nlp/bangla-electra, csebuetnlp/banglabert, xlm-roberta-large, neuralspace-reverie/indic-transformers-bn-xlmroberta]

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. csebuetnlp/banglabert/all_results.json +12 -0
  2. csebuetnlp/banglabert/checkpoint-6000/optimizer.pt +3 -0
  3. csebuetnlp/banglabert/checkpoint-6000/pytorch_model.bin +3 -0
  4. {neuralspace-reverie/indic-transformers-bn-bert/checkpoint-9000 → csebuetnlp/banglabert/checkpoint-6000}/rng_state.pth +1 -1
  5. neuralspace-reverie/indic-transformers-bn-bert/runs/Jan13_22-23-10_e3061b854d02/1642112608.4571896/events.out.tfevents.1642112608.e3061b854d02.83.1 → csebuetnlp/banglabert/checkpoint-6000/scheduler.pt +2 -2
  6. {neuralspace-reverie/indic-transformers-bn-bert/checkpoint-9000 → csebuetnlp/banglabert/checkpoint-6000}/special_tokens_map.json +0 -0
  7. csebuetnlp/banglabert/checkpoint-6000/tokenizer.json +0 -0
  8. csebuetnlp/banglabert/checkpoint-6000/tokenizer_config.json +1 -0
  9. csebuetnlp/banglabert/checkpoint-6000/trainer_state.json +94 -0
  10. {neuralspace-reverie/indic-transformers-bn-bert/checkpoint-9000 → csebuetnlp/banglabert/checkpoint-6000}/training_args.bin +2 -2
  11. csebuetnlp/banglabert/checkpoint-6000/vocab.txt +0 -0
  12. csebuetnlp/banglabert/eval_results.json +12 -0
  13. csebuetnlp/banglabert/pytorch_model.bin +3 -0
  14. neuralspace-reverie/indic-transformers-bn-bert/runs/Jan13_22-23-10_e3061b854d02/events.out.tfevents.1642112608.e3061b854d02.83.0 → csebuetnlp/banglabert/runs/Jan14_10-42-02_4ac0fd88951a/1642156936.1436265/events.out.tfevents.1642156936.4ac0fd88951a.2300.11 +2 -2
  15. csebuetnlp/banglabert/runs/Jan14_10-42-02_4ac0fd88951a/events.out.tfevents.1642156936.4ac0fd88951a.2300.10 +3 -0
  16. csebuetnlp/banglabert/special_tokens_map.json +1 -0
  17. csebuetnlp/banglabert/tokenizer.json +0 -0
  18. csebuetnlp/banglabert/tokenizer_config.json +1 -0
  19. csebuetnlp/banglabert/training_args.bin +3 -0
  20. csebuetnlp/banglabert/vocab.txt +0 -0
  21. distilbert-base-multilingual-cased/all_results.json +12 -0
  22. distilbert-base-multilingual-cased/checkpoint-13000/optimizer.pt +3 -0
  23. distilbert-base-multilingual-cased/checkpoint-13000/pytorch_model.bin +3 -0
  24. distilbert-base-multilingual-cased/checkpoint-13000/rng_state.pth +3 -0
  25. distilbert-base-multilingual-cased/checkpoint-13000/scheduler.pt +3 -0
  26. distilbert-base-multilingual-cased/checkpoint-13000/special_tokens_map.json +1 -0
  27. distilbert-base-multilingual-cased/checkpoint-13000/tokenizer.json +0 -0
  28. distilbert-base-multilingual-cased/checkpoint-13000/tokenizer_config.json +1 -0
  29. distilbert-base-multilingual-cased/checkpoint-13000/trainer_state.json +196 -0
  30. distilbert-base-multilingual-cased/checkpoint-13000/training_args.bin +3 -0
  31. distilbert-base-multilingual-cased/checkpoint-13000/vocab.txt +0 -0
  32. distilbert-base-multilingual-cased/eval_results.json +12 -0
  33. distilbert-base-multilingual-cased/pytorch_model.bin +3 -0
  34. distilbert-base-multilingual-cased/runs/Jan14_09-24-05_4ac0fd88951a/1642152252.538074/events.out.tfevents.1642152252.4ac0fd88951a.2300.5 +3 -0
  35. distilbert-base-multilingual-cased/runs/Jan14_09-24-05_4ac0fd88951a/events.out.tfevents.1642152252.4ac0fd88951a.2300.4 +3 -0
  36. distilbert-base-multilingual-cased/special_tokens_map.json +1 -0
  37. distilbert-base-multilingual-cased/tokenizer.json +0 -0
  38. distilbert-base-multilingual-cased/tokenizer_config.json +1 -0
  39. distilbert-base-multilingual-cased/training_args.bin +3 -0
  40. distilbert-base-multilingual-cased/vocab.txt +0 -0
  41. final_result.csv +9 -0
  42. monsoon-nlp/bangla-electra/all_results.json +12 -0
  43. monsoon-nlp/bangla-electra/checkpoint-9000/optimizer.pt +3 -0
  44. monsoon-nlp/bangla-electra/checkpoint-9000/pytorch_model.bin +3 -0
  45. monsoon-nlp/bangla-electra/checkpoint-9000/rng_state.pth +3 -0
  46. {neuralspace-reverie/indic-transformers-bn-bert → monsoon-nlp/bangla-electra}/checkpoint-9000/scheduler.pt +0 -0
  47. monsoon-nlp/bangla-electra/checkpoint-9000/special_tokens_map.json +1 -0
  48. monsoon-nlp/bangla-electra/checkpoint-9000/tokenizer.json +0 -0
  49. monsoon-nlp/bangla-electra/checkpoint-9000/tokenizer_config.json +1 -0
  50. monsoon-nlp/bangla-electra/checkpoint-9000/trainer_state.json +136 -0
csebuetnlp/banglabert/all_results.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 0.33,
3
+ "eval_samples": 1346,
4
+ "test_accuracy": 0.7838038632986627,
5
+ "test_f1": 0.7063572149344096,
6
+ "test_loss": 0.6481547951698303,
7
+ "test_precision": 0.6653992395437263,
8
+ "test_recall": 0.7526881720430108,
9
+ "test_runtime": 25.7216,
10
+ "test_samples_per_second": 52.33,
11
+ "test_steps_per_second": 13.102
12
+ }
csebuetnlp/banglabert/checkpoint-6000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4258268f20e9b4adc92b50d325444ca5fe0321345b4d5ca0e5427d72de9581e2
3
+ size 880341733
csebuetnlp/banglabert/checkpoint-6000/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a3ddeaed8431e6cfc5e1bcee875048299cfd136dee17d09155643129db7368f
3
+ size 440196977
{neuralspace-reverie/indic-transformers-bn-bert/checkpoint-9000 → csebuetnlp/banglabert/checkpoint-6000}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:68209be3febbde2507e5520fcf993ad8a03c411182d39984b191c3736c196da8
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aa114000851e2597418582b2934c1866379c0f0a17955b4b0adb8118903acaa1
3
  size 14503
neuralspace-reverie/indic-transformers-bn-bert/runs/Jan13_22-23-10_e3061b854d02/1642112608.4571896/events.out.tfevents.1642112608.e3061b854d02.83.1 → csebuetnlp/banglabert/checkpoint-6000/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c783b070cc22f5a430f6bda538dcb1a62dbd7d3dd581cea2ea3fa9cec8f71ed7
3
- size 4386
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5aed322968a4c02e3f947148f035a35d2206eb8fcc1753d1bf64eb8bfb95083c
3
+ size 623
{neuralspace-reverie/indic-transformers-bn-bert/checkpoint-9000 → csebuetnlp/banglabert/checkpoint-6000}/special_tokens_map.json RENAMED
File without changes
csebuetnlp/banglabert/checkpoint-6000/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
csebuetnlp/banglabert/checkpoint-6000/tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"do_lower_case": false, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": false, "strip_accents": null, "special_tokens_map_file": null, "full_tokenizer_file": null, "name_or_path": "csebuetnlp/banglabert", "do_basic_tokenize": true, "never_split": null, "tokenizer_class": "ElectraTokenizer"}
csebuetnlp/banglabert/checkpoint-6000/trainer_state.json ADDED
@@ -0,0 +1,94 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.7063572149344096,
3
+ "best_model_checkpoint": "./clickbait_identification/csebuetnlp/banglabert/checkpoint-6000",
4
+ "epoch": 1.9828155981493722,
5
+ "global_step": 6000,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.33,
12
+ "eval_accuracy": 0.7013372956909361,
13
+ "eval_f1": 0.32094594594594594,
14
+ "eval_loss": 0.5926413536071777,
15
+ "eval_precision": 0.7480314960629921,
16
+ "eval_recall": 0.20430107526881722,
17
+ "eval_runtime": 25.7194,
18
+ "eval_samples_per_second": 52.334,
19
+ "eval_steps_per_second": 13.103,
20
+ "step": 1000
21
+ },
22
+ {
23
+ "epoch": 0.66,
24
+ "eval_accuracy": 0.7332838038632986,
25
+ "eval_f1": 0.4665676077265974,
26
+ "eval_loss": 0.514832079410553,
27
+ "eval_precision": 0.7548076923076923,
28
+ "eval_recall": 0.33763440860215055,
29
+ "eval_runtime": 25.719,
30
+ "eval_samples_per_second": 52.335,
31
+ "eval_steps_per_second": 13.103,
32
+ "step": 2000
33
+ },
34
+ {
35
+ "epoch": 0.99,
36
+ "eval_accuracy": 0.7540861812778603,
37
+ "eval_f1": 0.6814244465832531,
38
+ "eval_loss": 0.49710002541542053,
39
+ "eval_precision": 0.6167247386759582,
40
+ "eval_recall": 0.7612903225806451,
41
+ "eval_runtime": 25.7279,
42
+ "eval_samples_per_second": 52.317,
43
+ "eval_steps_per_second": 13.099,
44
+ "step": 3000
45
+ },
46
+ {
47
+ "epoch": 1.0,
48
+ "learning_rate": 1.9e-05,
49
+ "loss": 0.58,
50
+ "step": 3026
51
+ },
52
+ {
53
+ "epoch": 1.32,
54
+ "eval_accuracy": 0.7674591381872214,
55
+ "eval_f1": 0.6916256157635469,
56
+ "eval_loss": 0.628989577293396,
57
+ "eval_precision": 0.6381818181818182,
58
+ "eval_recall": 0.7548387096774194,
59
+ "eval_runtime": 25.7155,
60
+ "eval_samples_per_second": 52.342,
61
+ "eval_steps_per_second": 13.105,
62
+ "step": 4000
63
+ },
64
+ {
65
+ "epoch": 1.65,
66
+ "eval_accuracy": 0.7867756315007429,
67
+ "eval_f1": 0.6849615806805708,
68
+ "eval_loss": 0.6614874005317688,
69
+ "eval_precision": 0.6995515695067265,
70
+ "eval_recall": 0.6709677419354839,
71
+ "eval_runtime": 26.0668,
72
+ "eval_samples_per_second": 51.636,
73
+ "eval_steps_per_second": 12.928,
74
+ "step": 5000
75
+ },
76
+ {
77
+ "epoch": 1.98,
78
+ "eval_accuracy": 0.7838038632986627,
79
+ "eval_f1": 0.7063572149344096,
80
+ "eval_loss": 0.6481547951698303,
81
+ "eval_precision": 0.6653992395437263,
82
+ "eval_recall": 0.7526881720430108,
83
+ "eval_runtime": 25.9916,
84
+ "eval_samples_per_second": 51.786,
85
+ "eval_steps_per_second": 12.966,
86
+ "step": 6000
87
+ }
88
+ ],
89
+ "max_steps": 60520,
90
+ "num_train_epochs": 20,
91
+ "total_flos": 0.0,
92
+ "trial_name": null,
93
+ "trial_params": null
94
+ }
{neuralspace-reverie/indic-transformers-bn-bert/checkpoint-9000 → csebuetnlp/banglabert/checkpoint-6000}/training_args.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e947bfd5f033f36ba9608cfa052f8b713ccb9e11306a7a6c982746cf6595121e
3
- size 2799
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f2e5952205fe295833f75995454e9b8718a4b278409827724e27ec10f8788ff1
3
+ size 2991
csebuetnlp/banglabert/checkpoint-6000/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
csebuetnlp/banglabert/eval_results.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 0.33,
3
+ "eval_samples": 1346,
4
+ "test_accuracy": 0.7838038632986627,
5
+ "test_f1": 0.7063572149344096,
6
+ "test_loss": 0.6481547951698303,
7
+ "test_precision": 0.6653992395437263,
8
+ "test_recall": 0.7526881720430108,
9
+ "test_runtime": 25.7216,
10
+ "test_samples_per_second": 52.33,
11
+ "test_steps_per_second": 13.102
12
+ }
csebuetnlp/banglabert/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a3ddeaed8431e6cfc5e1bcee875048299cfd136dee17d09155643129db7368f
3
+ size 440196977
neuralspace-reverie/indic-transformers-bn-bert/runs/Jan13_22-23-10_e3061b854d02/events.out.tfevents.1642112608.e3061b854d02.83.0 → csebuetnlp/banglabert/runs/Jan14_10-42-02_4ac0fd88951a/1642156936.1436265/events.out.tfevents.1642156936.4ac0fd88951a.2300.11 RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b7c7c2bc3ff38a84eda46ae44d6f9ebf2cf4266865ad786edfcd84a53eeaa6c3
3
- size 9070
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e0a8c741725d06cfc25ed9dc39923be0d12db4000e929af0494538dc43972ac
3
+ size 4830
csebuetnlp/banglabert/runs/Jan14_10-42-02_4ac0fd88951a/events.out.tfevents.1642156936.4ac0fd88951a.2300.10 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:816f0b2be9728594b1d5331f395bfa66a60d2b3081d736d2d0ef004986fba1bf
3
+ size 7723
csebuetnlp/banglabert/special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
csebuetnlp/banglabert/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
csebuetnlp/banglabert/tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"do_lower_case": false, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": false, "strip_accents": null, "special_tokens_map_file": null, "full_tokenizer_file": null, "name_or_path": "csebuetnlp/banglabert", "do_basic_tokenize": true, "never_split": null, "tokenizer_class": "ElectraTokenizer"}
csebuetnlp/banglabert/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f2e5952205fe295833f75995454e9b8718a4b278409827724e27ec10f8788ff1
3
+ size 2991
csebuetnlp/banglabert/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
distilbert-base-multilingual-cased/all_results.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 0.33,
3
+ "eval_samples": 1346,
4
+ "test_accuracy": 0.7578008915304606,
5
+ "test_f1": 0.6632231404958677,
6
+ "test_loss": 0.7381333708763123,
7
+ "test_precision": 0.6381709741550696,
8
+ "test_recall": 0.6903225806451613,
9
+ "test_runtime": 15.8433,
10
+ "test_samples_per_second": 84.957,
11
+ "test_steps_per_second": 21.271
12
+ }
distilbert-base-multilingual-cased/checkpoint-13000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46fad17c165fb4906369b88a570c6ccd372d38d309a685861be52b97b275f620
3
+ size 1077943721
distilbert-base-multilingual-cased/checkpoint-13000/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:025a283d17baff317ddb014ac0edf7d928cf174f5fd7adc0e92a0d36fd8d0bd8
3
+ size 538980917
distilbert-base-multilingual-cased/checkpoint-13000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8603747db40b433ca7b897885b152f245f1673f1691cecdc7eb027c0cc359691
3
+ size 14503
distilbert-base-multilingual-cased/checkpoint-13000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7cebc06887b1ea5796f15d5311dd791a8f4bf296812f347ff6f02ef2b6525a8b
3
+ size 623
distilbert-base-multilingual-cased/checkpoint-13000/special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
distilbert-base-multilingual-cased/checkpoint-13000/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
distilbert-base-multilingual-cased/checkpoint-13000/tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"do_lower_case": false, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "distilbert-base-multilingual-cased", "tokenizer_class": "DistilBertTokenizer"}
distilbert-base-multilingual-cased/checkpoint-13000/trainer_state.json ADDED
@@ -0,0 +1,196 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.6632231404958677,
3
+ "best_model_checkpoint": "./clickbait_identification/distilbert-base-multilingual-cased/checkpoint-13000",
4
+ "epoch": 4.296100462656973,
5
+ "global_step": 13000,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.33,
12
+ "eval_accuracy": 0.6909361069836553,
13
+ "eval_f1": 0.3396825396825397,
14
+ "eval_loss": 0.598863959312439,
15
+ "eval_precision": 0.6484848484848484,
16
+ "eval_recall": 0.23010752688172043,
17
+ "eval_runtime": 15.8265,
18
+ "eval_samples_per_second": 85.047,
19
+ "eval_steps_per_second": 21.293,
20
+ "step": 1000
21
+ },
22
+ {
23
+ "epoch": 0.66,
24
+ "eval_accuracy": 0.7139673105497771,
25
+ "eval_f1": 0.48598130841121495,
26
+ "eval_loss": 0.5456291437149048,
27
+ "eval_precision": 0.6408450704225352,
28
+ "eval_recall": 0.3913978494623656,
29
+ "eval_runtime": 15.7981,
30
+ "eval_samples_per_second": 85.2,
31
+ "eval_steps_per_second": 21.332,
32
+ "step": 2000
33
+ },
34
+ {
35
+ "epoch": 0.99,
36
+ "eval_accuracy": 0.736998514115899,
37
+ "eval_f1": 0.5378590078328981,
38
+ "eval_loss": 0.601733386516571,
39
+ "eval_precision": 0.6843853820598007,
40
+ "eval_recall": 0.443010752688172,
41
+ "eval_runtime": 16.1605,
42
+ "eval_samples_per_second": 83.29,
43
+ "eval_steps_per_second": 20.853,
44
+ "step": 3000
45
+ },
46
+ {
47
+ "epoch": 1.0,
48
+ "learning_rate": 1.9e-05,
49
+ "loss": 0.6042,
50
+ "step": 3026
51
+ },
52
+ {
53
+ "epoch": 1.32,
54
+ "eval_accuracy": 0.7288261515601783,
55
+ "eval_f1": 0.6178010471204188,
56
+ "eval_loss": 0.5779110789299011,
57
+ "eval_precision": 0.6020408163265306,
58
+ "eval_recall": 0.6344086021505376,
59
+ "eval_runtime": 15.9992,
60
+ "eval_samples_per_second": 84.129,
61
+ "eval_steps_per_second": 21.064,
62
+ "step": 4000
63
+ },
64
+ {
65
+ "epoch": 1.65,
66
+ "eval_accuracy": 0.7288261515601783,
67
+ "eval_f1": 0.5866364665911664,
68
+ "eval_loss": 0.5593090057373047,
69
+ "eval_precision": 0.6196172248803827,
70
+ "eval_recall": 0.556989247311828,
71
+ "eval_runtime": 16.0298,
72
+ "eval_samples_per_second": 83.968,
73
+ "eval_steps_per_second": 21.023,
74
+ "step": 5000
75
+ },
76
+ {
77
+ "epoch": 1.98,
78
+ "eval_accuracy": 0.7347696879643388,
79
+ "eval_f1": 0.6375634517766496,
80
+ "eval_loss": 0.6610430479049683,
81
+ "eval_precision": 0.6038461538461538,
82
+ "eval_recall": 0.6752688172043011,
83
+ "eval_runtime": 15.807,
84
+ "eval_samples_per_second": 85.152,
85
+ "eval_steps_per_second": 21.32,
86
+ "step": 6000
87
+ },
88
+ {
89
+ "epoch": 2.0,
90
+ "learning_rate": 1.8e-05,
91
+ "loss": 0.5693,
92
+ "step": 6052
93
+ },
94
+ {
95
+ "epoch": 2.31,
96
+ "eval_accuracy": 0.7421991084695394,
97
+ "eval_f1": 0.5510996119016817,
98
+ "eval_loss": 0.7018317580223083,
99
+ "eval_precision": 0.6915584415584416,
100
+ "eval_recall": 0.45806451612903226,
101
+ "eval_runtime": 15.8126,
102
+ "eval_samples_per_second": 85.122,
103
+ "eval_steps_per_second": 21.312,
104
+ "step": 7000
105
+ },
106
+ {
107
+ "epoch": 2.64,
108
+ "eval_accuracy": 0.7488855869242199,
109
+ "eval_f1": 0.6434599156118143,
110
+ "eval_loss": 0.6016787886619568,
111
+ "eval_precision": 0.6314699792960663,
112
+ "eval_recall": 0.6559139784946236,
113
+ "eval_runtime": 16.1537,
114
+ "eval_samples_per_second": 83.324,
115
+ "eval_steps_per_second": 20.862,
116
+ "step": 8000
117
+ },
118
+ {
119
+ "epoch": 2.97,
120
+ "eval_accuracy": 0.7473997028231798,
121
+ "eval_f1": 0.6473029045643153,
122
+ "eval_loss": 0.5445184707641602,
123
+ "eval_precision": 0.625250501002004,
124
+ "eval_recall": 0.6709677419354839,
125
+ "eval_runtime": 15.9994,
126
+ "eval_samples_per_second": 84.128,
127
+ "eval_steps_per_second": 21.063,
128
+ "step": 9000
129
+ },
130
+ {
131
+ "epoch": 3.0,
132
+ "learning_rate": 1.7e-05,
133
+ "loss": 0.5564,
134
+ "step": 9078
135
+ },
136
+ {
137
+ "epoch": 3.3,
138
+ "eval_accuracy": 0.7243684992570579,
139
+ "eval_f1": 0.6522961574507966,
140
+ "eval_loss": 0.6023839116096497,
141
+ "eval_precision": 0.5780730897009967,
142
+ "eval_recall": 0.7483870967741936,
143
+ "eval_runtime": 15.9931,
144
+ "eval_samples_per_second": 84.161,
145
+ "eval_steps_per_second": 21.072,
146
+ "step": 10000
147
+ },
148
+ {
149
+ "epoch": 3.64,
150
+ "eval_accuracy": 0.7147102526002972,
151
+ "eval_f1": 0.6489945155393052,
152
+ "eval_loss": 0.6206891536712646,
153
+ "eval_precision": 0.5643879173290938,
154
+ "eval_recall": 0.7634408602150538,
155
+ "eval_runtime": 15.7964,
156
+ "eval_samples_per_second": 85.21,
157
+ "eval_steps_per_second": 21.334,
158
+ "step": 11000
159
+ },
160
+ {
161
+ "epoch": 3.97,
162
+ "eval_accuracy": 0.7473997028231798,
163
+ "eval_f1": 0.6620278330019881,
164
+ "eval_loss": 0.6622892618179321,
165
+ "eval_precision": 0.6155268022181146,
166
+ "eval_recall": 0.7161290322580646,
167
+ "eval_runtime": 15.8234,
168
+ "eval_samples_per_second": 85.064,
169
+ "eval_steps_per_second": 21.298,
170
+ "step": 12000
171
+ },
172
+ {
173
+ "epoch": 4.0,
174
+ "learning_rate": 1.6000000000000003e-05,
175
+ "loss": 0.5472,
176
+ "step": 12104
177
+ },
178
+ {
179
+ "epoch": 4.3,
180
+ "eval_accuracy": 0.7578008915304606,
181
+ "eval_f1": 0.6632231404958677,
182
+ "eval_loss": 0.7381333708763123,
183
+ "eval_precision": 0.6381709741550696,
184
+ "eval_recall": 0.6903225806451613,
185
+ "eval_runtime": 16.1759,
186
+ "eval_samples_per_second": 83.21,
187
+ "eval_steps_per_second": 20.833,
188
+ "step": 13000
189
+ }
190
+ ],
191
+ "max_steps": 60520,
192
+ "num_train_epochs": 20,
193
+ "total_flos": 0.0,
194
+ "trial_name": null,
195
+ "trial_params": null
196
+ }
distilbert-base-multilingual-cased/checkpoint-13000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:47d830c2782ebe8b10368751d6867c1f878d2df28a7ef85e6f2b2eb6962848f9
3
+ size 3055
distilbert-base-multilingual-cased/checkpoint-13000/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
distilbert-base-multilingual-cased/eval_results.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 0.33,
3
+ "eval_samples": 1346,
4
+ "test_accuracy": 0.7578008915304606,
5
+ "test_f1": 0.6632231404958677,
6
+ "test_loss": 0.7381333708763123,
7
+ "test_precision": 0.6381709741550696,
8
+ "test_recall": 0.6903225806451613,
9
+ "test_runtime": 15.8433,
10
+ "test_samples_per_second": 84.957,
11
+ "test_steps_per_second": 21.271
12
+ }
distilbert-base-multilingual-cased/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:025a283d17baff317ddb014ac0edf7d928cf174f5fd7adc0e92a0d36fd8d0bd8
3
+ size 538980917
distilbert-base-multilingual-cased/runs/Jan14_09-24-05_4ac0fd88951a/1642152252.538074/events.out.tfevents.1642152252.4ac0fd88951a.2300.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a3b21e093ded94909557bae23bfadddaeca3e3ccd10735ddce98eecbfb949996
3
+ size 4869
distilbert-base-multilingual-cased/runs/Jan14_09-24-05_4ac0fd88951a/events.out.tfevents.1642152252.4ac0fd88951a.2300.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:baaf0f9243e82c55cfc2d40c06965f419cea15217789e6b265073b39a38c759f
3
+ size 11537
distilbert-base-multilingual-cased/special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
distilbert-base-multilingual-cased/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
distilbert-base-multilingual-cased/tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"do_lower_case": false, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "distilbert-base-multilingual-cased", "tokenizer_class": "DistilBertTokenizer"}
distilbert-base-multilingual-cased/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:47d830c2782ebe8b10368751d6867c1f878d2df28a7ef85e6f2b2eb6962848f9
3
+ size 3055
distilbert-base-multilingual-cased/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
final_result.csv ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ ,train_epoch,test_accuracy,test_f1,test_precision,test_recall,test_loss
2
+ neuralspace-reverie/indic-transformers-bn-bert,0.33,0.7689450222882616,0.661588683351469,0.6696035242290749,0.6537634408602151,0.6427938938140869
3
+ neuralspace-reverie/indic-transformers-bn-roberta,0.33,0.7466567607726597,0.6827906976744187,0.601639344262295,0.789247311827957,0.5567601323127747
4
+ distilbert-base-multilingual-cased,0.33,0.7578008915304606,0.6632231404958677,0.6381709741550696,0.6903225806451613,0.7381333708763123
5
+ neuralspace-reverie/indic-transformers-bn-distilbert,0.33,0.7979197622585439,0.7213114754098361,0.6888454011741683,0.7569892473118279,0.7468417882919312
6
+ monsoon-nlp/bangla-electra,0.33,0.7414561664190193,0.6692015209125476,0.5996592844974447,0.7569892473118279,0.547999382019043
7
+ csebuetnlp/banglabert,0.33,0.7838038632986627,0.7063572149344096,0.6653992395437263,0.7526881720430108,0.6481547951698303
8
+ xlm-roberta-large,0.17,0.6545319465081724,0.0,0.0,0.0,0.6548592448234558
9
+ neuralspace-reverie/indic-transformers-bn-xlmroberta,0.17,0.7771173848439822,0.6781115879828327,0.6766595289079229,0.6795698924731183,1.1499747037887573
monsoon-nlp/bangla-electra/all_results.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 0.33,
3
+ "eval_samples": 1346,
4
+ "test_accuracy": 0.7414561664190193,
5
+ "test_f1": 0.6692015209125476,
6
+ "test_loss": 0.547999382019043,
7
+ "test_precision": 0.5996592844974447,
8
+ "test_recall": 0.7569892473118279,
9
+ "test_runtime": 6.8808,
10
+ "test_samples_per_second": 195.617,
11
+ "test_steps_per_second": 48.977
12
+ }
monsoon-nlp/bangla-electra/checkpoint-9000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f160ff966f41851aed2bc2d11cbfb60bb5705631344c6f748c890325c48d59b7
3
+ size 107345693
monsoon-nlp/bangla-electra/checkpoint-9000/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a08f51ae88a5ae3d20797e32400fe98bbb38ac1d17348fd0966bd36055f7a862
3
+ size 53699053
monsoon-nlp/bangla-electra/checkpoint-9000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fbf949d1440a036887b89511c3f2943839dda58cd7391e3d3032f221b696bc5f
3
+ size 14503
{neuralspace-reverie/indic-transformers-bn-bert → monsoon-nlp/bangla-electra}/checkpoint-9000/scheduler.pt RENAMED
File without changes
monsoon-nlp/bangla-electra/checkpoint-9000/special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
monsoon-nlp/bangla-electra/checkpoint-9000/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
monsoon-nlp/bangla-electra/checkpoint-9000/tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"do_lower_case": false, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": false, "special_tokens_map_file": "/root/.cache/huggingface/transformers/03b59f5325d31d4b9c4899993115ee1540291ed877e0d2ff3fd4dff814ada400.dd8bd9bfd3664b530ea4e645105f557769387b3da9f79bdb55ed556bdd80611d", "name_or_path": "monsoon-nlp/bangla-electra", "do_basic_tokenize": true, "never_split": null, "tokenizer_class": "ElectraTokenizer"}
monsoon-nlp/bangla-electra/checkpoint-9000/trainer_state.json ADDED
@@ -0,0 +1,136 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.6692015209125476,
3
+ "best_model_checkpoint": "./clickbait_identification/monsoon-nlp/bangla-electra/checkpoint-9000",
4
+ "epoch": 2.9742233972240584,
5
+ "global_step": 9000,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.33,
12
+ "eval_accuracy": 0.7013372956909361,
13
+ "eval_f1": 0.3116438356164384,
14
+ "eval_loss": 0.5911309123039246,
15
+ "eval_precision": 0.7647058823529411,
16
+ "eval_recall": 0.1956989247311828,
17
+ "eval_runtime": 6.9517,
18
+ "eval_samples_per_second": 193.622,
19
+ "eval_steps_per_second": 48.477,
20
+ "step": 1000
21
+ },
22
+ {
23
+ "epoch": 0.66,
24
+ "eval_accuracy": 0.711738484398217,
25
+ "eval_f1": 0.4947916666666667,
26
+ "eval_loss": 0.5567843317985535,
27
+ "eval_precision": 0.6270627062706271,
28
+ "eval_recall": 0.40860215053763443,
29
+ "eval_runtime": 6.8804,
30
+ "eval_samples_per_second": 195.628,
31
+ "eval_steps_per_second": 48.98,
32
+ "step": 2000
33
+ },
34
+ {
35
+ "epoch": 0.99,
36
+ "eval_accuracy": 0.6731054977711739,
37
+ "eval_f1": 0.608540925266904,
38
+ "eval_loss": 0.5933745503425598,
39
+ "eval_precision": 0.5189681335356601,
40
+ "eval_recall": 0.7354838709677419,
41
+ "eval_runtime": 6.8721,
42
+ "eval_samples_per_second": 195.863,
43
+ "eval_steps_per_second": 49.039,
44
+ "step": 3000
45
+ },
46
+ {
47
+ "epoch": 1.0,
48
+ "learning_rate": 1.9e-05,
49
+ "loss": 0.5987,
50
+ "step": 3026
51
+ },
52
+ {
53
+ "epoch": 1.32,
54
+ "eval_accuracy": 0.7325408618127786,
55
+ "eval_f1": 0.6086956521739132,
56
+ "eval_loss": 0.5164241790771484,
57
+ "eval_precision": 0.6153846153846154,
58
+ "eval_recall": 0.6021505376344086,
59
+ "eval_runtime": 6.8817,
60
+ "eval_samples_per_second": 195.59,
61
+ "eval_steps_per_second": 48.97,
62
+ "step": 4000
63
+ },
64
+ {
65
+ "epoch": 1.65,
66
+ "eval_accuracy": 0.7169390787518574,
67
+ "eval_f1": 0.6514181152790485,
68
+ "eval_loss": 0.5508525371551514,
69
+ "eval_precision": 0.5668789808917197,
70
+ "eval_recall": 0.7655913978494624,
71
+ "eval_runtime": 6.8787,
72
+ "eval_samples_per_second": 195.677,
73
+ "eval_steps_per_second": 48.992,
74
+ "step": 5000
75
+ },
76
+ {
77
+ "epoch": 1.98,
78
+ "eval_accuracy": 0.736998514115899,
79
+ "eval_f1": 0.6522593320235757,
80
+ "eval_loss": 0.5713074803352356,
81
+ "eval_precision": 0.6003616636528029,
82
+ "eval_recall": 0.7139784946236559,
83
+ "eval_runtime": 6.8865,
84
+ "eval_samples_per_second": 195.456,
85
+ "eval_steps_per_second": 48.937,
86
+ "step": 6000
87
+ },
88
+ {
89
+ "epoch": 2.0,
90
+ "learning_rate": 1.8e-05,
91
+ "loss": 0.5535,
92
+ "step": 6052
93
+ },
94
+ {
95
+ "epoch": 2.31,
96
+ "eval_accuracy": 0.7518573551263001,
97
+ "eval_f1": 0.6563786008230453,
98
+ "eval_loss": 0.5634607076644897,
99
+ "eval_precision": 0.6291913214990138,
100
+ "eval_recall": 0.6860215053763441,
101
+ "eval_runtime": 6.8807,
102
+ "eval_samples_per_second": 195.619,
103
+ "eval_steps_per_second": 48.977,
104
+ "step": 7000
105
+ },
106
+ {
107
+ "epoch": 2.64,
108
+ "eval_accuracy": 0.7273402674591382,
109
+ "eval_f1": 0.6654512306289881,
110
+ "eval_loss": 0.6494086980819702,
111
+ "eval_precision": 0.5775316455696202,
112
+ "eval_recall": 0.7849462365591398,
113
+ "eval_runtime": 6.878,
114
+ "eval_samples_per_second": 195.696,
115
+ "eval_steps_per_second": 48.997,
116
+ "step": 8000
117
+ },
118
+ {
119
+ "epoch": 2.97,
120
+ "eval_accuracy": 0.7414561664190193,
121
+ "eval_f1": 0.6692015209125476,
122
+ "eval_loss": 0.547999382019043,
123
+ "eval_precision": 0.5996592844974447,
124
+ "eval_recall": 0.7569892473118279,
125
+ "eval_runtime": 6.8738,
126
+ "eval_samples_per_second": 195.815,
127
+ "eval_steps_per_second": 49.027,
128
+ "step": 9000
129
+ }
130
+ ],
131
+ "max_steps": 60520,
132
+ "num_train_epochs": 20,
133
+ "total_flos": 0.0,
134
+ "trial_name": null,
135
+ "trial_params": null
136
+ }