lfoppiano commited on
Commit
47f9e3b
·
1 Parent(s): 134a99c

migrate models to HF

Browse files
Files changed (37) hide show
  1. .gitattributes +1 -0
  2. context_bert/config.json +20 -0
  3. context_bert/model_weights.hdf5 +3 -0
  4. context_bert/transformer-config.json +25 -0
  5. context_bert/transformer-tokenizer/special_tokens_map.json +1 -0
  6. context_bert/transformer-tokenizer/tokenizer.json +0 -0
  7. context_bert/transformer-tokenizer/tokenizer_config.json +1 -0
  8. context_bert/transformer-tokenizer/vocab.txt +0 -0
  9. context_creation_bert/config.json +19 -0
  10. context_creation_bert/model_weights.hdf5 +3 -0
  11. context_creation_bert/transformer-config.json +25 -0
  12. context_creation_bert/transformer-tokenizer/special_tokens_map.json +1 -0
  13. context_creation_bert/transformer-tokenizer/tokenizer.json +0 -0
  14. context_creation_bert/transformer-tokenizer/tokenizer_config.json +1 -0
  15. context_creation_bert/transformer-tokenizer/vocab.txt +0 -0
  16. context_shared_bert/config.json +19 -0
  17. context_shared_bert/model_weights.hdf5 +3 -0
  18. context_shared_bert/transformer-config.json +25 -0
  19. context_shared_bert/transformer-tokenizer/special_tokens_map.json +1 -0
  20. context_shared_bert/transformer-tokenizer/tokenizer.json +0 -0
  21. context_shared_bert/transformer-tokenizer/tokenizer_config.json +1 -0
  22. context_shared_bert/transformer-tokenizer/vocab.txt +0 -0
  23. context_used_bert/config.json +19 -0
  24. context_used_bert/model_weights.hdf5 +3 -0
  25. context_used_bert/transformer-config.json +25 -0
  26. context_used_bert/transformer-tokenizer/special_tokens_map.json +1 -0
  27. context_used_bert/transformer-tokenizer/tokenizer.json +0 -0
  28. context_used_bert/transformer-tokenizer/tokenizer_config.json +1 -0
  29. context_used_bert/transformer-tokenizer/vocab.txt +0 -0
  30. software-BERT/config.json +38 -0
  31. software-BERT/model_weights.hdf5 +3 -0
  32. software-BERT/preprocessor.json +646 -0
  33. software-BERT/transformer-config.json +21 -0
  34. software-BERT/transformer-tokenizer/special_tokens_map.json +7 -0
  35. software-BERT/transformer-tokenizer/tokenizer.json +0 -0
  36. software-BERT/transformer-tokenizer/tokenizer_config.json +19 -0
  37. software-BERT/transformer-tokenizer/vocab.txt +0 -0
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ *.hdf5 filter=lfs diff=lfs merge=lfs -text
context_bert/config.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model_name": "context_bert",
3
+ "architecture": "bert",
4
+ "embeddings_name": null,
5
+ "char_embedding_size": 25,
6
+ "word_embedding_size": 0,
7
+ "dropout": 0.5,
8
+ "recurrent_dropout": 0.25,
9
+ "maxlen": 100,
10
+ "dense_size": 32,
11
+ "use_char_feature": false,
12
+ "list_classes": [
13
+ "used",
14
+ "creation",
15
+ "shared"
16
+ ],
17
+ "fold_number": 1,
18
+ "batch_size": 32,
19
+ "transformer_name": "michiyasunaga/LinkBERT-base"
20
+ }
context_bert/model_weights.hdf5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0ea2189fa380300c38d697ee7edf59881177ec8fa844275c7718152b66be449
3
+ size 433524128
context_bert/transformer-config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "michiyasunaga/LinkBERT-base",
3
+ "architectures": [
4
+ "BertModel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 0,
20
+ "position_embedding_type": "absolute",
21
+ "transformers_version": "4.15.0",
22
+ "type_vocab_size": 2,
23
+ "use_cache": true,
24
+ "vocab_size": 28996
25
+ }
context_bert/transformer-tokenizer/special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
context_bert/transformer-tokenizer/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
context_bert/transformer-tokenizer/tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"do_lower_case": false, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "michiyasunaga/LinkBERT-base", "add_special_tokens": true, "max_length": 100, "add_prefix_space": true, "tokenizer_class": "BertTokenizer"}
context_bert/transformer-tokenizer/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
context_creation_bert/config.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model_name": "context_creation_bert",
3
+ "architecture": "bert",
4
+ "embeddings_name": null,
5
+ "char_embedding_size": 25,
6
+ "word_embedding_size": 0,
7
+ "dropout": 0.5,
8
+ "recurrent_dropout": 0.25,
9
+ "maxlen": 100,
10
+ "dense_size": 32,
11
+ "use_char_feature": false,
12
+ "list_classes": [
13
+ "creation",
14
+ "not_creation"
15
+ ],
16
+ "fold_number": 1,
17
+ "batch_size": 32,
18
+ "transformer_name": "michiyasunaga/LinkBERT-base"
19
+ }
context_creation_bert/model_weights.hdf5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:82c65aa8020380c2ac32b144e3a803549b9343370b3cb458f4a1c494f26dc2a4
3
+ size 433521056
context_creation_bert/transformer-config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "michiyasunaga/LinkBERT-base",
3
+ "architectures": [
4
+ "BertModel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 0,
20
+ "position_embedding_type": "absolute",
21
+ "transformers_version": "4.15.0",
22
+ "type_vocab_size": 2,
23
+ "use_cache": true,
24
+ "vocab_size": 28996
25
+ }
context_creation_bert/transformer-tokenizer/special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
context_creation_bert/transformer-tokenizer/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
context_creation_bert/transformer-tokenizer/tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"do_lower_case": false, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "michiyasunaga/LinkBERT-base", "add_special_tokens": true, "max_length": 100, "add_prefix_space": true, "tokenizer_class": "BertTokenizer"}
context_creation_bert/transformer-tokenizer/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
context_shared_bert/config.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model_name": "context_shared_bert",
3
+ "architecture": "bert",
4
+ "embeddings_name": null,
5
+ "char_embedding_size": 25,
6
+ "word_embedding_size": 0,
7
+ "dropout": 0.5,
8
+ "recurrent_dropout": 0.25,
9
+ "maxlen": 100,
10
+ "dense_size": 32,
11
+ "use_char_feature": false,
12
+ "list_classes": [
13
+ "shared",
14
+ "not_shared"
15
+ ],
16
+ "fold_number": 1,
17
+ "batch_size": 32,
18
+ "transformer_name": "michiyasunaga/LinkBERT-base"
19
+ }
context_shared_bert/model_weights.hdf5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7844fdf560ac0ad413503940457d2a5b9ea6069eca3e1a04ced69fca58f03c47
3
+ size 433521056
context_shared_bert/transformer-config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "michiyasunaga/LinkBERT-base",
3
+ "architectures": [
4
+ "BertModel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 0,
20
+ "position_embedding_type": "absolute",
21
+ "transformers_version": "4.15.0",
22
+ "type_vocab_size": 2,
23
+ "use_cache": true,
24
+ "vocab_size": 28996
25
+ }
context_shared_bert/transformer-tokenizer/special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
context_shared_bert/transformer-tokenizer/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
context_shared_bert/transformer-tokenizer/tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"do_lower_case": false, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "michiyasunaga/LinkBERT-base", "add_special_tokens": true, "max_length": 100, "add_prefix_space": true, "tokenizer_class": "BertTokenizer"}
context_shared_bert/transformer-tokenizer/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
context_used_bert/config.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model_name": "context_used_bert",
3
+ "architecture": "bert",
4
+ "embeddings_name": null,
5
+ "char_embedding_size": 25,
6
+ "word_embedding_size": 0,
7
+ "dropout": 0.5,
8
+ "recurrent_dropout": 0.25,
9
+ "maxlen": 100,
10
+ "dense_size": 32,
11
+ "use_char_feature": false,
12
+ "list_classes": [
13
+ "used",
14
+ "not_used"
15
+ ],
16
+ "fold_number": 1,
17
+ "batch_size": 32,
18
+ "transformer_name": "michiyasunaga/LinkBERT-base"
19
+ }
context_used_bert/model_weights.hdf5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:95ea068e2ab8ed28b22c984e32ac8ad0f8b6105c100d00ef9cad8b4dffb1be75
3
+ size 433521056
context_used_bert/transformer-config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "michiyasunaga/LinkBERT-base",
3
+ "architectures": [
4
+ "BertModel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 0,
20
+ "position_embedding_type": "absolute",
21
+ "transformers_version": "4.15.0",
22
+ "type_vocab_size": 2,
23
+ "use_cache": true,
24
+ "vocab_size": 28996
25
+ }
context_used_bert/transformer-tokenizer/special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
context_used_bert/transformer-tokenizer/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
context_used_bert/transformer-tokenizer/tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"do_lower_case": false, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "michiyasunaga/LinkBERT-base", "add_special_tokens": true, "max_length": 100, "add_prefix_space": true, "tokenizer_class": "BertTokenizer"}
context_used_bert/transformer-tokenizer/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
software-BERT/config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model_name": "software-BERT",
3
+ "architecture": "BERT",
4
+ "embeddings_name": null,
5
+ "char_vocab_size": 599,
6
+ "case_vocab_size": 8,
7
+ "char_embedding_size": 25,
8
+ "num_char_lstm_units": 25,
9
+ "max_char_length": 30,
10
+ "features_vocabulary_size": 12,
11
+ "features_indices": null,
12
+ "features_embedding_size": 4,
13
+ "features_lstm_units": 4,
14
+ "max_sequence_length": 512,
15
+ "word_embedding_size": 0,
16
+ "num_word_lstm_units": 100,
17
+ "case_embedding_size": 5,
18
+ "dropout": 0.5,
19
+ "recurrent_dropout": 0.5,
20
+ "use_crf": false,
21
+ "use_chain_crf": false,
22
+ "fold_number": 1,
23
+ "batch_size": 8,
24
+ "transformer_name": "allenai/scibert_scivocab_cased",
25
+ "use_ELMo": false,
26
+ "labels": {
27
+ "<PAD>": 0,
28
+ "B-<creator>": 1,
29
+ "B-<software>": 2,
30
+ "B-<url>": 3,
31
+ "B-<version>": 4,
32
+ "I-<creator>": 5,
33
+ "I-<software>": 6,
34
+ "I-<url>": 7,
35
+ "I-<version>": 8,
36
+ "O": 9
37
+ }
38
+ }
software-BERT/model_weights.hdf5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ea5e50e4d778c464a75f6bab2901107c5f50dbb16d62249efb4994390183a4d
3
+ size 440060072
software-BERT/preprocessor.json ADDED
@@ -0,0 +1,646 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "padding": true,
3
+ "return_lengths": false,
4
+ "return_word_embeddings": false,
5
+ "return_casing": false,
6
+ "return_features": false,
7
+ "return_chars": false,
8
+ "return_bert_embeddings": true,
9
+ "vocab_char": {
10
+ "<PAD>": 0,
11
+ "<UNK>": 1,
12
+ "!": 2,
13
+ "\"": 3,
14
+ "#": 4,
15
+ "$": 5,
16
+ "%": 6,
17
+ "&": 7,
18
+ "'": 8,
19
+ "(": 9,
20
+ ")": 10,
21
+ "*": 11,
22
+ "+": 12,
23
+ ",": 13,
24
+ "-": 14,
25
+ ".": 15,
26
+ "/": 16,
27
+ "0": 17,
28
+ "1": 18,
29
+ "2": 19,
30
+ "3": 20,
31
+ "4": 21,
32
+ "5": 22,
33
+ "6": 23,
34
+ "7": 24,
35
+ "8": 25,
36
+ "9": 26,
37
+ ":": 27,
38
+ ";": 28,
39
+ "<": 29,
40
+ "=": 30,
41
+ ">": 31,
42
+ "?": 32,
43
+ "@": 33,
44
+ "A": 34,
45
+ "B": 35,
46
+ "C": 36,
47
+ "D": 37,
48
+ "E": 38,
49
+ "F": 39,
50
+ "G": 40,
51
+ "H": 41,
52
+ "I": 42,
53
+ "J": 43,
54
+ "K": 44,
55
+ "L": 45,
56
+ "M": 46,
57
+ "N": 47,
58
+ "O": 48,
59
+ "P": 49,
60
+ "Q": 50,
61
+ "R": 51,
62
+ "S": 52,
63
+ "T": 53,
64
+ "U": 54,
65
+ "V": 55,
66
+ "W": 56,
67
+ "X": 57,
68
+ "Y": 58,
69
+ "Z": 59,
70
+ "[": 60,
71
+ "\\": 61,
72
+ "]": 62,
73
+ "^": 63,
74
+ "_": 64,
75
+ "`": 65,
76
+ "a": 66,
77
+ "b": 67,
78
+ "c": 68,
79
+ "d": 69,
80
+ "e": 70,
81
+ "f": 71,
82
+ "g": 72,
83
+ "h": 73,
84
+ "i": 74,
85
+ "j": 75,
86
+ "k": 76,
87
+ "l": 77,
88
+ "m": 78,
89
+ "n": 79,
90
+ "o": 80,
91
+ "p": 81,
92
+ "q": 82,
93
+ "r": 83,
94
+ "s": 84,
95
+ "t": 85,
96
+ "u": 86,
97
+ "v": 87,
98
+ "w": 88,
99
+ "x": 89,
100
+ "y": 90,
101
+ "z": 91,
102
+ "{": 92,
103
+ "|": 93,
104
+ "}": 94,
105
+ "~": 95,
106
+ "\u00a1": 96,
107
+ "\u00a2": 97,
108
+ "\u00a3": 98,
109
+ "\u00a5": 99,
110
+ "\u00a7": 100,
111
+ "\u00a8": 101,
112
+ "\u00a9": 102,
113
+ "\u00aa": 103,
114
+ "\u00ab": 104,
115
+ "\u00ac": 105,
116
+ "\u00ae": 106,
117
+ "\u00af": 107,
118
+ "\u00b0": 108,
119
+ "\u00b1": 109,
120
+ "\u00b2": 110,
121
+ "\u00b3": 111,
122
+ "\u00b4": 112,
123
+ "\u00b5": 113,
124
+ "\u00b6": 114,
125
+ "\u00b7": 115,
126
+ "\u00b8": 116,
127
+ "\u00b9": 117,
128
+ "\u00ba": 118,
129
+ "\u00bb": 119,
130
+ "\u00bc": 120,
131
+ "\u00bd": 121,
132
+ "\u00be": 122,
133
+ "\u00bf": 123,
134
+ "\u00c0": 124,
135
+ "\u00c1": 125,
136
+ "\u00c2": 126,
137
+ "\u00c3": 127,
138
+ "\u00c4": 128,
139
+ "\u00c5": 129,
140
+ "\u00c7": 130,
141
+ "\u00c8": 131,
142
+ "\u00c9": 132,
143
+ "\u00ca": 133,
144
+ "\u00cd": 134,
145
+ "\u00d0": 135,
146
+ "\u00d1": 136,
147
+ "\u00d2": 137,
148
+ "\u00d3": 138,
149
+ "\u00d4": 139,
150
+ "\u00d5": 140,
151
+ "\u00d6": 141,
152
+ "\u00d7": 142,
153
+ "\u00d8": 143,
154
+ "\u00d9": 144,
155
+ "\u00da": 145,
156
+ "\u00dc": 146,
157
+ "\u00de": 147,
158
+ "\u00df": 148,
159
+ "\u00e0": 149,
160
+ "\u00e1": 150,
161
+ "\u00e2": 151,
162
+ "\u00e3": 152,
163
+ "\u00e4": 153,
164
+ "\u00e5": 154,
165
+ "\u00e7": 155,
166
+ "\u00e8": 156,
167
+ "\u00e9": 157,
168
+ "\u00ea": 158,
169
+ "\u00eb": 159,
170
+ "\u00ec": 160,
171
+ "\u00ed": 161,
172
+ "\u00ee": 162,
173
+ "\u00ef": 163,
174
+ "\u00f0": 164,
175
+ "\u00f1": 165,
176
+ "\u00f2": 166,
177
+ "\u00f3": 167,
178
+ "\u00f4": 168,
179
+ "\u00f5": 169,
180
+ "\u00f6": 170,
181
+ "\u00f7": 171,
182
+ "\u00f8": 172,
183
+ "\u00f9": 173,
184
+ "\u00fa": 174,
185
+ "\u00fb": 175,
186
+ "\u00fc": 176,
187
+ "\u00fd": 177,
188
+ "\u00fe": 178,
189
+ "\u0100": 179,
190
+ "\u0101": 180,
191
+ "\u0103": 181,
192
+ "\u0105": 182,
193
+ "\u0106": 183,
194
+ "\u0107": 184,
195
+ "\u0108": 185,
196
+ "\u0109": 186,
197
+ "\u010c": 187,
198
+ "\u010d": 188,
199
+ "\u0113": 189,
200
+ "\u0117": 190,
201
+ "\u0118": 191,
202
+ "\u011c": 192,
203
+ "\u012a": 193,
204
+ "\u012b": 194,
205
+ "\u0130": 195,
206
+ "\u0131": 196,
207
+ "\u0138": 197,
208
+ "\u013e": 198,
209
+ "\u0141": 199,
210
+ "\u0142": 200,
211
+ "\u0144": 201,
212
+ "\u0148": 202,
213
+ "\u014d": 203,
214
+ "\u0151": 204,
215
+ "\u0159": 205,
216
+ "\u015a": 206,
217
+ "\u015c": 207,
218
+ "\u015d": 208,
219
+ "\u015e": 209,
220
+ "\u015f": 210,
221
+ "\u0160": 211,
222
+ "\u0161": 212,
223
+ "\u0168": 213,
224
+ "\u0169": 214,
225
+ "\u016b": 215,
226
+ "\u0173": 216,
227
+ "\u0175": 217,
228
+ "\u0176": 218,
229
+ "\u0177": 219,
230
+ "\u017a": 220,
231
+ "\u017d": 221,
232
+ "\u017e": 222,
233
+ "\u0192": 223,
234
+ "\u021b": 224,
235
+ "\u025b": 225,
236
+ "\u0263": 226,
237
+ "\u0288": 227,
238
+ "\u02a6": 228,
239
+ "\u02b9": 229,
240
+ "\u02bc": 230,
241
+ "\u02c2": 231,
242
+ "\u02c3": 232,
243
+ "\u02c6": 233,
244
+ "\u02c7": 234,
245
+ "\u02c8": 235,
246
+ "\u02c9": 236,
247
+ "\u02d8": 237,
248
+ "\u02d9": 238,
249
+ "\u02da": 239,
250
+ "\u02db": 240,
251
+ "\u02dc": 241,
252
+ "\u0302": 242,
253
+ "\u0303": 243,
254
+ "\u0305": 244,
255
+ "\u030a": 245,
256
+ "\u0313": 246,
257
+ "\u0314": 247,
258
+ "\u034c": 248,
259
+ "\u0350": 249,
260
+ "\u0351": 250,
261
+ "\u0352": 251,
262
+ "\u0354": 252,
263
+ "\u0357": 253,
264
+ "\u0358": 254,
265
+ "\u0371": 255,
266
+ "\u0374": 256,
267
+ "\u0392": 257,
268
+ "\u0393": 258,
269
+ "\u0394": 259,
270
+ "\u0397": 260,
271
+ "\u0398": 261,
272
+ "\u039b": 262,
273
+ "\u039e": 263,
274
+ "\u03a0": 264,
275
+ "\u03a3": 265,
276
+ "\u03a4": 266,
277
+ "\u03a5": 267,
278
+ "\u03a6": 268,
279
+ "\u03a7": 269,
280
+ "\u03a8": 270,
281
+ "\u03a9": 271,
282
+ "\u03b1": 272,
283
+ "\u03b2": 273,
284
+ "\u03b3": 274,
285
+ "\u03b4": 275,
286
+ "\u03b5": 276,
287
+ "\u03b6": 277,
288
+ "\u03b7": 278,
289
+ "\u03b8": 279,
290
+ "\u03b9": 280,
291
+ "\u03ba": 281,
292
+ "\u03bb": 282,
293
+ "\u03bc": 283,
294
+ "\u03bd": 284,
295
+ "\u03be": 285,
296
+ "\u03c0": 286,
297
+ "\u03c1": 287,
298
+ "\u03c2": 288,
299
+ "\u03c3": 289,
300
+ "\u03c4": 290,
301
+ "\u03c5": 291,
302
+ "\u03c6": 292,
303
+ "\u03c7": 293,
304
+ "\u03c8": 294,
305
+ "\u03c9": 295,
306
+ "\u03d1": 296,
307
+ "\u03d2": 297,
308
+ "\u03d5": 298,
309
+ "\u03d6": 299,
310
+ "\u03e9": 300,
311
+ "\u03ea": 301,
312
+ "\u03eb": 302,
313
+ "\u03ed": 303,
314
+ "\u03ee": 304,
315
+ "\u03f1": 305,
316
+ "\u03f3": 306,
317
+ "\u03f5": 307,
318
+ "\u03f7": 308,
319
+ "\u03fd": 309,
320
+ "\u03fe": 310,
321
+ "\u0408": 311,
322
+ "\u0412": 312,
323
+ "\u0413": 313,
324
+ "\u041a": 314,
325
+ "\u041d": 315,
326
+ "\u0424": 316,
327
+ "\u0430": 317,
328
+ "\u0545": 318,
329
+ "\u0546": 319,
330
+ "\u0609": 320,
331
+ "\u060a": 321,
332
+ "\u060c": 322,
333
+ "\u060d": 323,
334
+ "\u060e": 324,
335
+ "\u0621": 325,
336
+ "\u0623": 326,
337
+ "\u0626": 327,
338
+ "\u0627": 328,
339
+ "\u0628": 329,
340
+ "\u0629": 330,
341
+ "\u062a": 331,
342
+ "\u062b": 332,
343
+ "\u062d": 333,
344
+ "\u062e": 334,
345
+ "\u062f": 335,
346
+ "\u0631": 336,
347
+ "\u0633": 337,
348
+ "\u0634": 338,
349
+ "\u0637": 339,
350
+ "\u0639": 340,
351
+ "\u0641": 341,
352
+ "\u0642": 342,
353
+ "\u0643": 343,
354
+ "\u0644": 344,
355
+ "\u0645": 345,
356
+ "\u0646": 346,
357
+ "\u0648": 347,
358
+ "\u0649": 348,
359
+ "\u064a": 349,
360
+ "\u064b": 350,
361
+ "\u0792": 351,
362
+ "\u1b80": 352,
363
+ "\u1e54": 353,
364
+ "\u1e63": 354,
365
+ "\u1e7c": 355,
366
+ "\u1e83": 356,
367
+ "\u1e90": 357,
368
+ "\u1e91": 358,
369
+ "\u1ebc": 359,
370
+ "\u1ebd": 360,
371
+ "\u1ef9": 361,
372
+ "\u1fb1": 362,
373
+ "\u1fbd": 363,
374
+ "\u2016": 364,
375
+ "\u201a": 365,
376
+ "\u2020": 366,
377
+ "\u2021": 367,
378
+ "\u2022": 368,
379
+ "\u2025": 369,
380
+ "\u2026": 370,
381
+ "\u202b": 371,
382
+ "\u202c": 372,
383
+ "\u2030": 373,
384
+ "\u2032": 374,
385
+ "\u2033": 375,
386
+ "\u2034": 376,
387
+ "\u203a": 377,
388
+ "\u203e": 378,
389
+ "\u2044": 379,
390
+ "\u204e": 380,
391
+ "\u20a9": 381,
392
+ "\u20ac": 382,
393
+ "\u2103": 383,
394
+ "\u2119": 384,
395
+ "\u211c": 385,
396
+ "\u211d": 386,
397
+ "\u2122": 387,
398
+ "\u2126": 388,
399
+ "\u2150": 389,
400
+ "\u2161": 390,
401
+ "\u2163": 391,
402
+ "\u2190": 392,
403
+ "\u2191": 393,
404
+ "\u2192": 394,
405
+ "\u2193": 395,
406
+ "\u2194": 396,
407
+ "\u21a6": 397,
408
+ "\u21c4": 398,
409
+ "\u21d2": 399,
410
+ "\u21d4": 400,
411
+ "\u2200": 401,
412
+ "\u2202": 402,
413
+ "\u2203": 403,
414
+ "\u2205": 404,
415
+ "\u2206": 405,
416
+ "\u2208": 406,
417
+ "\u220e": 407,
418
+ "\u2211": 408,
419
+ "\u2212": 409,
420
+ "\u2213": 410,
421
+ "\u2215": 411,
422
+ "\u221a": 412,
423
+ "\u221d": 413,
424
+ "\u221e": 414,
425
+ "\u2227": 415,
426
+ "\u2228": 416,
427
+ "\u2229": 417,
428
+ "\u222a": 418,
429
+ "\u222b": 419,
430
+ "\u2236": 420,
431
+ "\u223c": 421,
432
+ "\u2243": 422,
433
+ "\u2248": 423,
434
+ "\u2260": 424,
435
+ "\u2261": 425,
436
+ "\u2264": 426,
437
+ "\u2265": 427,
438
+ "\u226b": 428,
439
+ "\u227a": 429,
440
+ "\u227b": 430,
441
+ "\u2282": 431,
442
+ "\u2283": 432,
443
+ "\u2286": 433,
444
+ "\u2287": 434,
445
+ "\u2295": 435,
446
+ "\u2297": 436,
447
+ "\u22a5": 437,
448
+ "\u22c5": 438,
449
+ "\u22ef": 439,
450
+ "\u2329": 440,
451
+ "\u232a": 441,
452
+ "\u232c": 442,
453
+ "\u2338": 443,
454
+ "\u233a": 444,
455
+ "\u233d": 445,
456
+ "\u239b": 446,
457
+ "\u239d": 447,
458
+ "\u239e": 448,
459
+ "\u23a0": 449,
460
+ "\u2423": 450,
461
+ "\u2424": 451,
462
+ "\u2425": 452,
463
+ "\u2426": 453,
464
+ "\u2440": 454,
465
+ "\u24c7": 455,
466
+ "\u2500": 456,
467
+ "\u2502": 457,
468
+ "\u25a0": 458,
469
+ "\u25a1": 459,
470
+ "\u25aa": 460,
471
+ "\u25ab": 461,
472
+ "\u25b2": 462,
473
+ "\u25b3": 463,
474
+ "\u25b5": 464,
475
+ "\u25b6": 465,
476
+ "\u25b8": 466,
477
+ "\u25c6": 467,
478
+ "\u25c7": 468,
479
+ "\u25cb": 469,
480
+ "\u2605": 470,
481
+ "\u2610": 471,
482
+ "\u2713": 472,
483
+ "\u274f": 473,
484
+ "\u27e8": 474,
485
+ "\u27e9": 475,
486
+ "\u29cb": 476,
487
+ "\u2a7d": 477,
488
+ "\u2a7e": 478,
489
+ "\u3002": 479,
490
+ "\u3008": 480,
491
+ "\u3009": 481,
492
+ "\ud545": 482,
493
+ "\ue023": 483,
494
+ "\ue024": 484,
495
+ "\ue02c": 485,
496
+ "\ue02e": 486,
497
+ "\ue031": 487,
498
+ "\ue032": 488,
499
+ "\ue035": 489,
500
+ "\ue039": 490,
501
+ "\ue044": 491,
502
+ "\ue061": 492,
503
+ "\ue062": 493,
504
+ "\ue063": 494,
505
+ "\ue067": 495,
506
+ "\ue06b": 496,
507
+ "\ue06c": 497,
508
+ "\ue06d": 498,
509
+ "\ue06e": 499,
510
+ "\ue073": 500,
511
+ "\ue074": 501,
512
+ "\ue07a": 502,
513
+ "\ue103": 503,
514
+ "\ue104": 504,
515
+ "\uf025": 505,
516
+ "\uf03c": 506,
517
+ "\uf043": 507,
518
+ "\uf061": 508,
519
+ "\uf062": 509,
520
+ "\uf063": 510,
521
+ "\uf065": 511,
522
+ "\uf067": 512,
523
+ "\uf06b": 513,
524
+ "\uf06c": 514,
525
+ "\uf06d": 515,
526
+ "\uf070": 516,
527
+ "\uf073": 517,
528
+ "\uf074": 518,
529
+ "\uf077": 519,
530
+ "\uf0a2": 520,
531
+ "\uf0a3": 521,
532
+ "\uf0b0": 522,
533
+ "\uf0b1": 523,
534
+ "\uf0b4": 524,
535
+ "\uf0b6": 525,
536
+ "\uf0b9": 526,
537
+ "\uf0bb": 527,
538
+ "\uf0e0": 528,
539
+ "\uf0e2": 529,
540
+ "\uf0e4": 530,
541
+ "\uf0fc": 531,
542
+ "\uf643": 532,
543
+ "\uf644": 533,
544
+ "\uf645": 534,
545
+ "\uf646": 535,
546
+ "\uf647": 536,
547
+ "\uf648": 537,
548
+ "\uf649": 538,
549
+ "\uf64a": 539,
550
+ "\uf64b": 540,
551
+ "\uf64c": 541,
552
+ "\uf6f6": 542,
553
+ "\uf769": 543,
554
+ "\uf76a": 544,
555
+ "\uf76d": 545,
556
+ "\uf775": 546,
557
+ "\uf777": 547,
558
+ "\uf8e8": 548,
559
+ "\uff0b": 549,
560
+ "\uff0c": 550,
561
+ "\uff1a": 551,
562
+ "\uff1c": 552,
563
+ "\uff1d": 553,
564
+ "\uff1e": 554,
565
+ "\ufffd": 555,
566
+ "\ud835\udc34": 556,
567
+ "\ud835\udc36": 557,
568
+ "\ud835\udc37": 558,
569
+ "\ud835\udc39": 559,
570
+ "\ud835\udc3a": 560,
571
+ "\ud835\udc3b": 561,
572
+ "\ud835\udc3e": 562,
573
+ "\ud835\udc3f": 563,
574
+ "\ud835\udc43": 564,
575
+ "\ud835\udc47": 565,
576
+ "\ud835\udc4a": 566,
577
+ "\ud835\udc4c": 567,
578
+ "\ud835\udc50": 568,
579
+ "\ud835\udc51": 569,
580
+ "\ud835\udc52": 570,
581
+ "\ud835\udc53": 571,
582
+ "\ud835\udc54": 572,
583
+ "\ud835\udc56": 573,
584
+ "\ud835\udc57": 574,
585
+ "\ud835\udc58": 575,
586
+ "\ud835\udc5a": 576,
587
+ "\ud835\udc5e": 577,
588
+ "\ud835\udc5f": 578,
589
+ "\ud835\udc60": 579,
590
+ "\ud835\udc61": 580,
591
+ "\ud835\udc63": 581,
592
+ "\ud835\udc65": 582,
593
+ "\ud835\udd3c": 583,
594
+ "\ud835\udd40": 584,
595
+ "\ud835\udefc": 585,
596
+ "\ud835\udefd": 586,
597
+ "\ud835\udefe": 587,
598
+ "\ud835\udeff": 588,
599
+ "\ud835\udf00": 589,
600
+ "\ud835\udf03": 590,
601
+ "\ud835\udf06": 591,
602
+ "\ud835\udf07": 592,
603
+ "\ud835\udf0b": 593,
604
+ "\ud835\udf0d": 594,
605
+ "\ud835\udf0f": 595,
606
+ "\ud835\udf14": 596,
607
+ "\ud835\udf15": 597,
608
+ "\ud835\udf16": 598
609
+ },
610
+ "vocab_tag": {
611
+ "<PAD>": 0,
612
+ "B-<creator>": 1,
613
+ "B-<software>": 2,
614
+ "B-<url>": 3,
615
+ "B-<version>": 4,
616
+ "I-<creator>": 5,
617
+ "I-<software>": 6,
618
+ "I-<url>": 7,
619
+ "I-<version>": 8,
620
+ "O": 9
621
+ },
622
+ "vocab_case": [
623
+ "<PAD>",
624
+ "numeric",
625
+ "allLower",
626
+ "allUpper",
627
+ "initialUpper",
628
+ "other",
629
+ "mainly_numeric",
630
+ "contains_digit"
631
+ ],
632
+ "max_char_length": 30,
633
+ "feature_preprocessor": null,
634
+ "indice_tag": {
635
+ "0": "<PAD>",
636
+ "1": "B-<creator>",
637
+ "2": "B-<software>",
638
+ "3": "B-<url>",
639
+ "4": "B-<version>",
640
+ "5": "I-<creator>",
641
+ "6": "I-<software>",
642
+ "7": "I-<url>",
643
+ "8": "I-<version>",
644
+ "9": "O"
645
+ }
646
+ }
software-BERT/transformer-config.json ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "allenai/scibert_scivocab_cased",
3
+ "attention_probs_dropout_prob": 0.1,
4
+ "classifier_dropout": null,
5
+ "hidden_act": "gelu",
6
+ "hidden_dropout_prob": 0.1,
7
+ "hidden_size": 768,
8
+ "initializer_range": 0.02,
9
+ "intermediate_size": 3072,
10
+ "layer_norm_eps": 1e-12,
11
+ "max_position_embeddings": 512,
12
+ "model_type": "bert",
13
+ "num_attention_heads": 12,
14
+ "num_hidden_layers": 12,
15
+ "pad_token_id": 0,
16
+ "position_embedding_type": "absolute",
17
+ "transformers_version": "4.25.1",
18
+ "type_vocab_size": 2,
19
+ "use_cache": true,
20
+ "vocab_size": 31116
21
+ }
software-BERT/transformer-tokenizer/special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
software-BERT/transformer-tokenizer/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
software-BERT/transformer-tokenizer/tokenizer_config.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": true,
3
+ "add_special_tokens": true,
4
+ "cls_token": "[CLS]",
5
+ "do_basic_tokenize": true,
6
+ "do_lower_case": false,
7
+ "mask_token": "[MASK]",
8
+ "max_length": 512,
9
+ "model_max_length": 1000000000000000019884624838656,
10
+ "name_or_path": "allenai/scibert_scivocab_cased",
11
+ "never_split": null,
12
+ "pad_token": "[PAD]",
13
+ "sep_token": "[SEP]",
14
+ "special_tokens_map_file": null,
15
+ "strip_accents": null,
16
+ "tokenize_chinese_chars": true,
17
+ "tokenizer_class": "BertTokenizer",
18
+ "unk_token": "[UNK]"
19
+ }
software-BERT/transformer-tokenizer/vocab.txt ADDED
The diff for this file is too large to render. See raw diff