hans00 commited on
Commit
318c65b
·
verified ·
1 Parent(s): fea927b

Upload 8 files

Browse files
bert_zh/special_tokens_map.json CHANGED
@@ -1,37 +1,7 @@
1
- {
2
- "cls_token": {
3
- "content": "[CLS]",
4
- "lstrip": false,
5
- "normalized": false,
6
- "rstrip": false,
7
- "single_word": false
8
- },
9
- "mask_token": {
10
- "content": "[MASK]",
11
- "lstrip": false,
12
- "normalized": false,
13
- "rstrip": false,
14
- "single_word": false
15
- },
16
- "pad_token": {
17
- "content": "[PAD]",
18
- "lstrip": false,
19
- "normalized": false,
20
- "rstrip": false,
21
- "single_word": false
22
- },
23
- "sep_token": {
24
- "content": "[SEP]",
25
- "lstrip": false,
26
- "normalized": false,
27
- "rstrip": false,
28
- "single_word": false
29
- },
30
- "unk_token": {
31
- "content": "[UNK]",
32
- "lstrip": false,
33
- "normalized": false,
34
- "rstrip": false,
35
- "single_word": false
36
- }
37
- }
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
bert_zh/tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
bert_zh/tokenizer_config.json CHANGED
@@ -1,62 +1,58 @@
1
- {
2
- "added_tokens_decoder": {
3
- "0": {
4
- "content": "[PAD]",
5
- "lstrip": false,
6
- "normalized": false,
7
- "rstrip": false,
8
- "single_word": false,
9
- "special": true
10
- },
11
- "10": {
12
- "content": "[UNK]",
13
- "lstrip": false,
14
- "normalized": false,
15
- "rstrip": false,
16
- "single_word": false,
17
- "special": true
18
- },
19
- "11": {
20
- "content": "[CLS]",
21
- "lstrip": false,
22
- "normalized": false,
23
- "rstrip": false,
24
- "single_word": false,
25
- "special": true
26
- },
27
- "12": {
28
- "content": "[SEP]",
29
- "lstrip": false,
30
- "normalized": false,
31
- "rstrip": false,
32
- "single_word": false,
33
- "special": true
34
- },
35
- "13": {
36
- "content": "[MASK]",
37
- "lstrip": false,
38
- "normalized": false,
39
- "rstrip": false,
40
- "single_word": false,
41
- "special": true
42
- }
43
- },
44
- "clean_up_tokenization_spaces": true,
45
- "cls_token": "[CLS]",
46
- "do_basic_tokenize": true,
47
- "do_lower_case": false,
48
- "extra_special_tokens": {},
49
- "mask_token": "[MASK]",
50
- "max_length": null,
51
- "model_max_length": 512,
52
- "never_split": null,
53
- "pad_to_multiple_of": null,
54
- "pad_token": "[PAD]",
55
- "pad_token_type_id": 0,
56
- "padding_side": "right",
57
- "sep_token": "[SEP]",
58
- "strip_accents": null,
59
- "tokenize_chinese_chars": true,
60
- "tokenizer_class": "DistilBertTokenizer",
61
- "unk_token": "[UNK]"
62
- }
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_basic_tokenize": true,
47
+ "do_lower_case": false,
48
+ "extra_special_tokens": {},
49
+ "mask_token": "[MASK]",
50
+ "model_max_length": 512,
51
+ "never_split": null,
52
+ "pad_token": "[PAD]",
53
+ "sep_token": "[SEP]",
54
+ "strip_accents": null,
55
+ "tokenize_chinese_chars": true,
56
+ "tokenizer_class": "BertTokenizer",
57
+ "unk_token": "[UNK]"
58
+ }
 
 
 
 
bert_zh/vocab.txt CHANGED
The diff for this file is too large to render. See raw diff
 
config.json CHANGED
@@ -1,6 +1,5 @@
1
  {
2
  "_attn_implementation_autoset": true,
3
- "_name_or_path": "../../ellie-BertVITS2-mini",
4
  "activation_dropout": 0.1,
5
  "architectures": [
6
  "BertVits2Model"
@@ -13,41 +12,39 @@
13
  "bert_configs": [
14
  {
15
  "_attn_implementation_autoset": true,
16
- "_name_or_path": "distilbert-base-multilingual-cased",
17
- "activation": "gelu",
18
  "add_cross_attention": false,
19
  "architectures": [
20
- "DistilBertForMaskedLM"
21
  ],
22
- "attention_dropout": 0.1,
23
- "attention_probs_dropout_prob": 0.1,
24
  "bad_words_ids": null,
25
  "begin_suppress_tokens": null,
26
- "bos_token_id": null,
27
  "chunk_size_feed_forward": 0,
28
  "classifier_dropout": null,
 
29
  "cross_attention_hidden_size": null,
30
  "decoder_start_token_id": null,
31
- "dim": 768,
32
  "diversity_penalty": 0.0,
33
  "do_sample": false,
34
- "dropout": 0.1,
35
  "early_stopping": false,
 
36
  "encoder_no_repeat_ngram_size": 0,
37
- "eos_token_id": null,
38
  "exponential_decay_length_penalty": null,
39
  "finetuning_task": null,
40
  "forced_bos_token_id": null,
41
  "forced_eos_token_id": null,
42
- "hidden_act": "gelu",
43
- "hidden_dim": 3072,
44
- "hidden_dropout_prob": 0.1,
45
  "hidden_size": 768,
46
  "id2label": {
47
  "0": "LABEL_0",
48
  "1": "LABEL_1"
49
  },
50
  "initializer_range": 0.02,
 
51
  "intermediate_size": 3072,
52
  "is_decoder": false,
53
  "is_encoder_decoder": false,
@@ -61,49 +58,43 @@
61
  "max_position_embeddings": 512,
62
  "min_length": 0,
63
  "model_type": "bert",
64
- "n_heads": 12,
65
- "n_layers": 6,
66
  "no_repeat_ngram_size": 0,
67
  "num_attention_heads": 12,
68
  "num_beam_groups": 1,
69
  "num_beams": 1,
 
70
  "num_hidden_layers": 12,
71
  "num_return_sequences": 1,
72
  "output_attentions": false,
73
  "output_hidden_states": false,
74
- "output_past": true,
75
  "output_scores": false,
76
  "pad_token_id": 0,
77
  "position_embedding_type": "absolute",
78
  "prefix": null,
79
  "problem_type": null,
80
  "pruned_heads": {},
81
- "qa_dropout": 0.1,
82
  "remove_invalid_values": false,
83
  "repetition_penalty": 1.0,
84
  "return_dict": true,
85
  "return_dict_in_generate": false,
86
  "sep_token_id": null,
87
- "seq_classif_dropout": 0.2,
88
- "sinusoidal_pos_embds": false,
89
  "suppress_tokens": null,
90
  "task_specific_params": null,
91
  "temperature": 1.0,
92
  "tf_legacy_loss": false,
93
  "tie_encoder_decoder": false,
94
- "tie_weights_": true,
95
  "tie_word_embeddings": true,
96
- "tokenizer_class": null,
97
  "top_k": 50,
98
  "top_p": 1.0,
99
  "torch_dtype": null,
100
  "torchscript": false,
101
- "transformers_version": "4.49.0",
102
  "type_vocab_size": 2,
103
  "typical_p": 1.0,
104
  "use_bfloat16": false,
105
  "use_cache": true,
106
- "vocab_size": 41986
107
  }
108
  ],
109
  "cond_layer_index": 2,
@@ -166,7 +157,7 @@
166
  "spectrogram_bins": 1025,
167
  "stochastic_duration_prediction_ratio": 0.0,
168
  "torch_dtype": "float32",
169
- "transformers_version": "4.49.0",
170
  "upsample_initial_channel": 512,
171
  "upsample_kernel_sizes": [
172
  16,
 
1
  {
2
  "_attn_implementation_autoset": true,
 
3
  "activation_dropout": 0.1,
4
  "architectures": [
5
  "BertVits2Model"
 
12
  "bert_configs": [
13
  {
14
  "_attn_implementation_autoset": true,
15
+ "_name_or_path": "albert",
 
16
  "add_cross_attention": false,
17
  "architectures": [
18
+ "AlbertForMaskedLM"
19
  ],
20
+ "attention_probs_dropout_prob": 0,
 
21
  "bad_words_ids": null,
22
  "begin_suppress_tokens": null,
23
+ "bos_token_id": 2,
24
  "chunk_size_feed_forward": 0,
25
  "classifier_dropout": null,
26
+ "classifier_dropout_prob": 0.1,
27
  "cross_attention_hidden_size": null,
28
  "decoder_start_token_id": null,
 
29
  "diversity_penalty": 0.0,
30
  "do_sample": false,
 
31
  "early_stopping": false,
32
+ "embedding_size": 128,
33
  "encoder_no_repeat_ngram_size": 0,
34
+ "eos_token_id": 3,
35
  "exponential_decay_length_penalty": null,
36
  "finetuning_task": null,
37
  "forced_bos_token_id": null,
38
  "forced_eos_token_id": null,
39
+ "hidden_act": "relu",
40
+ "hidden_dropout_prob": 0,
 
41
  "hidden_size": 768,
42
  "id2label": {
43
  "0": "LABEL_0",
44
  "1": "LABEL_1"
45
  },
46
  "initializer_range": 0.02,
47
+ "inner_group_num": 1,
48
  "intermediate_size": 3072,
49
  "is_decoder": false,
50
  "is_encoder_decoder": false,
 
58
  "max_position_embeddings": 512,
59
  "min_length": 0,
60
  "model_type": "bert",
 
 
61
  "no_repeat_ngram_size": 0,
62
  "num_attention_heads": 12,
63
  "num_beam_groups": 1,
64
  "num_beams": 1,
65
+ "num_hidden_groups": 1,
66
  "num_hidden_layers": 12,
67
  "num_return_sequences": 1,
68
  "output_attentions": false,
69
  "output_hidden_states": false,
 
70
  "output_scores": false,
71
  "pad_token_id": 0,
72
  "position_embedding_type": "absolute",
73
  "prefix": null,
74
  "problem_type": null,
75
  "pruned_heads": {},
 
76
  "remove_invalid_values": false,
77
  "repetition_penalty": 1.0,
78
  "return_dict": true,
79
  "return_dict_in_generate": false,
80
  "sep_token_id": null,
 
 
81
  "suppress_tokens": null,
82
  "task_specific_params": null,
83
  "temperature": 1.0,
84
  "tf_legacy_loss": false,
85
  "tie_encoder_decoder": false,
 
86
  "tie_word_embeddings": true,
87
+ "tokenizer_class": "BertTokenizer",
88
  "top_k": 50,
89
  "top_p": 1.0,
90
  "torch_dtype": null,
91
  "torchscript": false,
92
+ "transformers_version": "4.50.2",
93
  "type_vocab_size": 2,
94
  "typical_p": 1.0,
95
  "use_bfloat16": false,
96
  "use_cache": true,
97
+ "vocab_size": 21128
98
  }
99
  ],
100
  "cond_layer_index": 2,
 
157
  "spectrogram_bins": 1025,
158
  "stochastic_duration_prediction_ratio": 0.0,
159
  "torch_dtype": "float32",
160
+ "transformers_version": "4.50.2",
161
  "upsample_initial_channel": 512,
162
  "upsample_kernel_sizes": [
163
  16,
onnx/model.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d74aa783e97a646c19c60c32e125fb994f3578b78715a71ccc258b566b3dd19a
3
- size 662923164
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dad34ada6add382125b12514bc61f0c911e81296b6cfae01ebcb3d0fd8f236b7
3
+ size 599032271
onnx/model_q4.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:37066564d2dfe450319a1cc6e6d608321cf44bf3fb634a31d5c5fdf23ba36abc
3
- size 402022953
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba8576f0f77ccddf3120efb2297f7f5da60d2a1eceac5bd2596180859849611a
3
+ size 338132060
onnx/model_quantized.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:43712524472ecd97ce78fb7f455a5561c70fef0f9f767cd0a68fcb7dc1da6e08
3
- size 277053380
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aaef55d7a487da0d842d0d697bb20503b9f28247e57fb6b8f8d5ddadbbd4ae7c
3
+ size 374317850