fdalvi commited on
Commit
d9ad8e7
1 Parent(s): 9da215e

NAACL 2022 trained version

Browse files
config.json ADDED
@@ -0,0 +1,122 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "BertForTokenClassification"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "directionality": "bidi",
7
+ "gradient_checkpointing": false,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.1,
10
+ "hidden_size": 768,
11
+ "id2label": {
12
+ "0": "O",
13
+ "1": "``",
14
+ "2": ",",
15
+ "3": ":",
16
+ "4": ".",
17
+ "5": "''",
18
+ "6": "$",
19
+ "7": "#",
20
+ "8": "CC",
21
+ "9": "CD",
22
+ "10": "DT",
23
+ "11": "EX",
24
+ "12": "FW",
25
+ "13": "IN",
26
+ "14": "JJ",
27
+ "15": "JJR",
28
+ "16": "JJS",
29
+ "17": "-LRB-",
30
+ "18": "LS",
31
+ "19": "MD",
32
+ "20": "NN",
33
+ "21": "NNP",
34
+ "22": "NNPS",
35
+ "23": "NNS",
36
+ "24": "PDT",
37
+ "25": "POS",
38
+ "26": "PRP",
39
+ "27": "PRP$",
40
+ "28": "RB",
41
+ "29": "RBR",
42
+ "30": "RBS",
43
+ "31": "RP",
44
+ "32": "-RRB-",
45
+ "33": "SYM",
46
+ "34": "TO",
47
+ "35": "UH",
48
+ "36": "VB",
49
+ "37": "VBD",
50
+ "38": "VBG",
51
+ "39": "VBN",
52
+ "40": "VBP",
53
+ "41": "VBZ",
54
+ "42": "WDT",
55
+ "43": "WP",
56
+ "44": "WP$",
57
+ "45": "WRB"
58
+ },
59
+ "initializer_range": 0.02,
60
+ "intermediate_size": 3072,
61
+ "label2id": {
62
+ "#": 7,
63
+ "$": 6,
64
+ "''": 5,
65
+ ",": 2,
66
+ "-LRB-": 17,
67
+ "-RRB-": 32,
68
+ ".": 4,
69
+ ":": 3,
70
+ "CC": 8,
71
+ "CD": 9,
72
+ "DT": 10,
73
+ "EX": 11,
74
+ "FW": 12,
75
+ "IN": 13,
76
+ "JJ": 14,
77
+ "JJR": 15,
78
+ "JJS": 16,
79
+ "LS": 18,
80
+ "MD": 19,
81
+ "NN": 20,
82
+ "NNP": 21,
83
+ "NNPS": 22,
84
+ "NNS": 23,
85
+ "O": 0,
86
+ "PDT": 24,
87
+ "POS": 25,
88
+ "PRP": 26,
89
+ "PRP$": 27,
90
+ "RB": 28,
91
+ "RBR": 29,
92
+ "RBS": 30,
93
+ "RP": 31,
94
+ "SYM": 33,
95
+ "TO": 34,
96
+ "UH": 35,
97
+ "VB": 36,
98
+ "VBD": 37,
99
+ "VBG": 38,
100
+ "VBN": 39,
101
+ "VBP": 40,
102
+ "VBZ": 41,
103
+ "WDT": 42,
104
+ "WP": 43,
105
+ "WP$": 44,
106
+ "WRB": 45,
107
+ "``": 1
108
+ },
109
+ "layer_norm_eps": 1e-12,
110
+ "max_position_embeddings": 512,
111
+ "model_type": "bert",
112
+ "num_attention_heads": 12,
113
+ "num_hidden_layers": 12,
114
+ "pad_token_id": 0,
115
+ "pooler_fc_size": 768,
116
+ "pooler_num_attention_heads": 12,
117
+ "pooler_num_fc_layers": 3,
118
+ "pooler_size_per_head": 128,
119
+ "pooler_type": "first_token_transform",
120
+ "type_vocab_size": 2,
121
+ "vocab_size": 119547
122
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b316dc8bf69a94910eb648873e22f1bdf8088d930324535ff3569ba6b09a388a
3
+ size 711609823
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"do_lower_case": false, "model_max_length": 512}
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a2f5914996661a654b872a5e32914a5739509cd8ff3e63fd6b9733591f03d26
3
+ size 1300
vocab.txt ADDED
The diff for this file is too large to render. See raw diff