ThilinaGunathilaka commited on
Commit
b609707
·
verified ·
1 Parent(s): 555fab4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +12 -24
README.md CHANGED
@@ -1,35 +1,25 @@
1
- <!-- ---
2
  license: mit
3
-
4
  base_model:
5
  - google-bert/bert-base-cased
6
  tags:
7
  - generated_from_trainer
8
  - finetune
9
- --- -->
10
-
11
- ---
12
- tags:
13
- - generated_from_trainer
14
- - finetune
15
- model-index:
16
- - name: fine-tune-sinhala-bert-v2
17
- results: []
18
  datasets:
19
  - vrclc/dakshina-lexicons-ml
20
- base_model:
 
 
 
21
  - Ransaka/sinhala-bert-medium-v2
22
- widget:
23
  - text: "අපි තමයි [MASK] කරේ."
24
  - text: "මට හෙට එන්න වෙන්නේ [MASK]."
25
  - text: "අපි ගෙදර [MASK]."
26
  - text: 'සිංහල සහ [MASK] අලුත් අවුරුද්ද.'
27
- license: mit
28
- language:
29
- - si
30
  ---
31
 
32
- # sinhala-bert-medium
33
 
34
  This model is pretrained on Sinhala data resources.
35
 
@@ -55,13 +45,14 @@ More information needed
55
  The following hyperparameters were used during training:
56
  - learning_rate: 5e-05
57
  - train_batch_size: 64
58
- - eval_batch_size: 8
59
  - seed: 42
60
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
61
  - lr_scheduler_type: linear
62
  - num_epochs: 6
63
 
64
  ### Training results
 
65
  Epoch Training Loss Validation Loss
66
  1 3.946600 3.898129
67
  2 3.782100 3.800080
@@ -70,12 +61,9 @@ Epoch Training Loss Validation Loss
70
  5 3.480900 3.601913
71
  6 3.420000 3.615573
72
 
73
-
74
-
75
  ### Framework versions
76
 
77
- - Transformers 4.33.3
78
  - Pytorch 2.0.0
79
- - Datasets 2.14.5
80
- - Tokenizers 0.13.3
81
-
 
1
+ ---
2
  license: mit
 
3
  base_model:
4
  - google-bert/bert-base-cased
5
  tags:
6
  - generated_from_trainer
7
  - finetune
 
 
 
 
 
 
 
 
 
8
  datasets:
9
  - vrclc/dakshina-lexicons-ml
10
+ language:
11
+ - si
12
+ new_version: google-bert/bert-base-cased
13
+ base-model:
14
  - Ransaka/sinhala-bert-medium-v2
15
+ wedget:
16
  - text: "අපි තමයි [MASK] කරේ."
17
  - text: "මට හෙට එන්න වෙන්නේ [MASK]."
18
  - text: "අපි ගෙදර [MASK]."
19
  - text: 'සිංහල සහ [MASK] අලුත් අවුරුද්ද.'
 
 
 
20
  ---
21
 
22
+ # fine-tune-sinhala-bert
23
 
24
  This model is pretrained on Sinhala data resources.
25
 
 
45
  The following hyperparameters were used during training:
46
  - learning_rate: 5e-05
47
  - train_batch_size: 64
48
+ - eval_batch_size: 16
49
  - seed: 42
50
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
51
  - lr_scheduler_type: linear
52
  - num_epochs: 6
53
 
54
  ### Training results
55
+
56
  Epoch Training Loss Validation Loss
57
  1 3.946600 3.898129
58
  2 3.782100 3.800080
 
61
  5 3.480900 3.601913
62
  6 3.420000 3.615573
63
 
 
 
64
  ### Framework versions
65
 
66
+ - Transformers 4.47.0
67
  - Pytorch 2.0.0
68
+ - Datasets 3.2.0
69
+ - Tokenizers 0.21.0