ThilinaGunathilaka commited on
Commit
555fab4
·
verified ·
1 Parent(s): 99d8dcc

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +75 -6
README.md CHANGED
@@ -1,12 +1,81 @@
1
- ---
2
  license: mit
3
- datasets:
4
- - vrclc/dakshina-lexicons-ml
5
- language:
6
- - si
7
  base_model:
8
  - google-bert/bert-base-cased
9
  tags:
10
  - generated_from_trainer
11
  - finetune
12
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ <!-- ---
2
  license: mit
3
+
 
 
 
4
  base_model:
5
  - google-bert/bert-base-cased
6
  tags:
7
  - generated_from_trainer
8
  - finetune
9
+ --- -->
10
+
11
+ ---
12
+ tags:
13
+ - generated_from_trainer
14
+ - finetune
15
+ model-index:
16
+ - name: fine-tune-sinhala-bert-v2
17
+ results: []
18
+ datasets:
19
+ - vrclc/dakshina-lexicons-ml
20
+ base_model:
21
+ - Ransaka/sinhala-bert-medium-v2
22
+ widget:
23
+ - text: "අපි තමයි [MASK] කරේ."
24
+ - text: "මට හෙට එන්න වෙන්නේ [MASK]."
25
+ - text: "අපි ගෙදර [MASK]."
26
+ - text: 'සිංහල සහ [MASK] අලුත් අවුරුද්ද.'
27
+ license: mit
28
+ language:
29
+ - si
30
+ ---
31
+
32
+ # sinhala-bert-medium
33
+
34
+ This model is pretrained on Sinhala data resources.
35
+
36
+ ## Model description
37
+
38
+ hidden_size = 786
39
+ num_hidden_layers = 6
40
+ num_attention_heads = 6
41
+ intermediate_size = 1024
42
+
43
+ ## Intended uses & limitations
44
+
45
+ More information needed
46
+
47
+ ## Training and evaluation data
48
+
49
+ More information needed
50
+
51
+ ## Training procedure
52
+
53
+ ### Training hyperparameters
54
+
55
+ The following hyperparameters were used during training:
56
+ - learning_rate: 5e-05
57
+ - train_batch_size: 64
58
+ - eval_batch_size: 8
59
+ - seed: 42
60
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
61
+ - lr_scheduler_type: linear
62
+ - num_epochs: 6
63
+
64
+ ### Training results
65
+ Epoch Training Loss Validation Loss
66
+ 1 3.946600 3.898129
67
+ 2 3.782100 3.800080
68
+ 3 3.678300 3.706316
69
+ 4 3.485600 3.646217
70
+ 5 3.480900 3.601913
71
+ 6 3.420000 3.615573
72
+
73
+
74
+
75
+ ### Framework versions
76
+
77
+ - Transformers 4.33.3
78
+ - Pytorch 2.0.0
79
+ - Datasets 2.14.5
80
+ - Tokenizers 0.13.3
81
+