zhezi12138 commited on
Commit
94e28a4
·
verified ·
1 Parent(s): c25f6d3

copy Ray2333/gpt2-large-harmless-reward_model

Browse files
Files changed (1) hide show
  1. config.json +5 -7
config.json CHANGED
@@ -1,9 +1,8 @@
1
  {
2
- "_name_or_path": "./weak_rm/best_checkpoint",
3
- "pipeline_tag": "text-classification",
4
  "activation_function": "gelu_new",
5
  "architectures": [
6
- "GPT2Model"
7
  ],
8
  "attn_pdrop": 0.1,
9
  "bos_token_id": 50256,
@@ -24,7 +23,6 @@
24
  "n_inner": null,
25
  "n_layer": 36,
26
  "n_positions": 1024,
27
- "pad_token_id": 50256,
28
  "reorder_and_upcast_attn": false,
29
  "resid_pdrop": 0.1,
30
  "scale_attn_by_inverse_layer_idx": false,
@@ -40,8 +38,8 @@
40
  "max_length": 50
41
  }
42
  },
43
- "torch_dtype": "float32",
44
- "transformers_version": "4.34.1",
45
- "use_cache": false,
46
  "vocab_size": 50257
47
  }
 
1
  {
2
+ "_name_or_path": "Ray2333/gpt2-large-harmless-reward_model",
 
3
  "activation_function": "gelu_new",
4
  "architectures": [
5
+ "GPT2ForSequenceClassification"
6
  ],
7
  "attn_pdrop": 0.1,
8
  "bos_token_id": 50256,
 
23
  "n_inner": null,
24
  "n_layer": 36,
25
  "n_positions": 1024,
 
26
  "reorder_and_upcast_attn": false,
27
  "resid_pdrop": 0.1,
28
  "scale_attn_by_inverse_layer_idx": false,
 
38
  "max_length": 50
39
  }
40
  },
41
+ "torch_dtype": "bfloat16",
42
+ "transformers_version": "4.36.2",
43
+ "use_cache": true,
44
  "vocab_size": 50257
45
  }