sileod commited on
Commit
1ed07c1
1 Parent(s): a37a5d5

Upload DebertaV2ForSequenceClassification

Browse files
Files changed (2) hide show
  1. config.json +65 -0
  2. pytorch_model.bin +3 -0
config.json ADDED
@@ -0,0 +1,65 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "sileod/deberta-v3-base-tasksource-nli",
3
+ "architectures": [
4
+ "DebertaV2ForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifiers_size": [
8
+ 2,
9
+ 2,
10
+ 2,
11
+ 3,
12
+ 1,
13
+ 1,
14
+ 1,
15
+ 2,
16
+ 1
17
+ ],
18
+ "hidden_act": "gelu",
19
+ "hidden_dropout_prob": 0.1,
20
+ "hidden_size": 768,
21
+ "id2label": {
22
+ "0": 0,
23
+ "1": 1
24
+ },
25
+ "initializer_range": 0.02,
26
+ "intermediate_size": 3072,
27
+ "label2id": {
28
+ "0": 0,
29
+ "1": 1
30
+ },
31
+ "layer_norm_eps": 1e-07,
32
+ "max_position_embeddings": 512,
33
+ "max_relative_positions": -1,
34
+ "model_type": "deberta-v2",
35
+ "norm_rel_ebd": "layer_norm",
36
+ "num_attention_heads": 12,
37
+ "num_hidden_layers": 12,
38
+ "pad_token_id": 0,
39
+ "pooler_dropout": 0,
40
+ "pooler_hidden_act": "gelu",
41
+ "pooler_hidden_size": 768,
42
+ "pos_att_type": [
43
+ "p2c",
44
+ "c2p"
45
+ ],
46
+ "position_biased_input": false,
47
+ "position_buckets": 256,
48
+ "relative_attention": true,
49
+ "share_att_key": true,
50
+ "tasks": [
51
+ "toxic_conversations",
52
+ "tweet_eval/hate",
53
+ "dynahate",
54
+ "implicit-hate-stg1",
55
+ "oasst2_dense_flat/toxicity",
56
+ "civil_comments/toxicity",
57
+ "civil_comments/severe_toxicity",
58
+ "HatemojiBuild",
59
+ "toxigen-data"
60
+ ],
61
+ "torch_dtype": "float32",
62
+ "transformers_version": "4.34.1",
63
+ "type_vocab_size": 0,
64
+ "vocab_size": 128100
65
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:666dff2d62bda764e8f5224bdff6a5c5372d065b42e6555cd4d1f1400a55d44c
3
+ size 737764786