p1atdev commited on
Commit
64b7031
1 Parent(s): 03194e9

Upload OPTForCausalLM

Browse files
Files changed (2) hide show
  1. config.json +2 -1
  2. model.safetensors +2 -2
config.json CHANGED
@@ -1,4 +1,5 @@
1
  {
 
2
  "_remove_final_layer_norm": false,
3
  "activation_function": "relu",
4
  "architectures": [
@@ -20,7 +21,7 @@
20
  "num_attention_heads": 12,
21
  "num_hidden_layers": 12,
22
  "pad_token_id": 2,
23
- "torch_dtype": "float32",
24
  "transformers_version": "4.38.0",
25
  "use_cache": true,
26
  "use_position_embeddings": false,
 
1
  {
2
+ "_name_or_path": "p1atdev/dart-v1-sft",
3
  "_remove_final_layer_norm": false,
4
  "activation_function": "relu",
5
  "architectures": [
 
21
  "num_attention_heads": 12,
22
  "num_hidden_layers": 12,
23
  "pad_token_id": 2,
24
+ "torch_dtype": "bfloat16",
25
  "transformers_version": "4.38.0",
26
  "use_cache": true,
27
  "use_position_embeddings": false,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:46d575643bec000f7ec8d04fe2980f4bd5dc42e6b8bd7374fb3bb2ec2885f418
3
- size 555427728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ac51e337885c1f2c3d21164d68468a5e21a0d906916fa416004201af58e8432
3
+ size 277725272