burtenshaw HF Staff commited on
Commit
8a7bd77
·
verified ·
1 Parent(s): 9c3978e

Upload config.json with huggingface_hub

Browse files
Files changed (1) hide show
  1. config.json +3 -5
config.json CHANGED
@@ -3,13 +3,12 @@
3
  "NanoChatForCausalLM"
4
  ],
5
  "attention_dropout": 0.0,
6
- "bos_token": "<|bos|>",
7
- "bos_token_id": 0,
8
- "chat_template": "{% if messages[0]['role'] == 'system' %}<|bos|><|user_start|>{{ messages[0]['content'] }}\n\n{{ messages[1]['content'] }}<|user_end|>{% set messages = messages[2:] %}{% else %}<|bos|>{% endif %}{% for message in messages %}{% if loop.index0 % 2 == 0 %}<|user_start|>{{ message['content'] }}<|user_end|>{% else %}<|assistant_start|>{{ message['content'] }}<|assistant_end|>{% endif %}{% endfor %}",
9
  "dtype": "bfloat16",
10
  "eos_token": "<|assistant_end|>",
11
  "eos_token_id": 65531,
12
  "hidden_act": "relu2",
 
13
  "initializer_range": 0.02,
14
  "intermediate_size": 8192,
15
  "logits_soft_cap": 15.0,
@@ -18,7 +17,6 @@
18
  "n_embd": 2048,
19
  "n_head": 16,
20
  "n_layer": 32,
21
- "hidden_size": 2048,
22
  "num_attention_heads": 16,
23
  "num_hidden_layers": 32,
24
  "num_key_value_heads": 16,
@@ -33,4 +31,4 @@
33
  "transformers_version": "4.57.0.dev0",
34
  "use_cache": true,
35
  "vocab_size": 65536
36
- }
 
3
  "NanoChatForCausalLM"
4
  ],
5
  "attention_dropout": 0.0,
6
+ "bos_token_id": 65527,
 
 
7
  "dtype": "bfloat16",
8
  "eos_token": "<|assistant_end|>",
9
  "eos_token_id": 65531,
10
  "hidden_act": "relu2",
11
+ "hidden_size": 2048,
12
  "initializer_range": 0.02,
13
  "intermediate_size": 8192,
14
  "logits_soft_cap": 15.0,
 
17
  "n_embd": 2048,
18
  "n_head": 16,
19
  "n_layer": 32,
 
20
  "num_attention_heads": 16,
21
  "num_hidden_layers": 32,
22
  "num_key_value_heads": 16,
 
31
  "transformers_version": "4.57.0.dev0",
32
  "use_cache": true,
33
  "vocab_size": 65536
34
+ }