Bedovyy commited on
Commit
44acc7e
·
verified ·
1 Parent(s): 6899c3d

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
config.json ADDED
@@ -0,0 +1,59 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Cohere2ForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 5,
8
+ "cache_implementation": "hybrid",
9
+ "eos_token_id": 255001,
10
+ "head_dim": 128,
11
+ "hidden_act": "silu",
12
+ "hidden_size": 12288,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 36864,
15
+ "layer_norm_eps": 1e-05,
16
+ "logit_scale": 0.25,
17
+ "max_position_embeddings": 131072,
18
+ "model_type": "cohere2",
19
+ "num_attention_heads": 96,
20
+ "num_hidden_layers": 64,
21
+ "num_key_value_heads": 8,
22
+ "order_of_interleaved_layers": "local_attn_first",
23
+ "pad_token_id": 0,
24
+ "position_embedding_type": "rope_gptj",
25
+ "quantization_config": {
26
+ "bits": 4,
27
+ "checkpoint_format": "gptq",
28
+ "desc_act": true,
29
+ "group_size": 128,
30
+ "lm_head": false,
31
+ "meta": {
32
+ "damp_auto_increment": 0.0025,
33
+ "damp_percent": 0.01,
34
+ "mse": 0.0,
35
+ "quantizer": [
36
+ "gptqmodel:2.3.0-dev"
37
+ ],
38
+ "static_groups": false,
39
+ "true_sequential": true,
40
+ "uri": "https://github.com/modelcloud/gptqmodel"
41
+ },
42
+ "pack_dtype": "int32",
43
+ "quant_method": "gptq",
44
+ "sym": true
45
+ },
46
+ "rope_scaling": null,
47
+ "rope_theta": 50000,
48
+ "rotary_pct": 1.0,
49
+ "sliding_window": 4096,
50
+ "sliding_window_pattern": 4,
51
+ "torch_dtype": "bfloat16",
52
+ "transformers_version": "4.50.1",
53
+ "use_cache": true,
54
+ "use_embedding_sharing": true,
55
+ "use_gated_activation": true,
56
+ "use_parallel_block": true,
57
+ "use_parallel_embedding": true,
58
+ "vocab_size": 256000
59
+ }
generation_config.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 5,
4
+ "cache_implementation": "hybrid",
5
+ "eos_token_id": 255001,
6
+ "pad_token_id": 0,
7
+ "transformers_version": "4.50.1"
8
+ }
model-00001-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a8b510a976376c61d1387063aec33104a962b451416d69d901089758bb00d9e8
3
+ size 6291456144
model-00002-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8cebc75efbc90462398e34dee046a66026df7fad72497bf3b36f79bb427ec824
3
+ size 3911384040
model-00003-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d31cf3607a14819de92fa9c9f1509dff5f41c80c03f622069c822b10c14b1e0
3
+ size 3976730368
model-00004-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da41a4c912c6949765d927764456d703ff9c6581e44e8216805743ee57b9d1d1
3
+ size 3911384144
model-00005-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e6d90e9204245b8aed459c0940f1bb31edd8523851ba6a513aab61f6177691bb
3
+ size 3976730496
model-00006-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:21d19d451d1d3573467b09fab0509c2484690db1c6ab71f7de7489bb923bd10c
3
+ size 3911384176
model-00007-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e7c97bd0b2cc159c402a9724d5894d1cc6f4067e420af90e02a1384cc57ed26b
3
+ size 3976730496
model-00008-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a893a2921b48ca85e672c72115da630e941dbfc8ef1a0d8fd9c24105425b2dc3
3
+ size 3911384176
model-00009-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:79ea307f4c6fea03ab0720353d5e1bd35e34574236f74fe9156478e88a44ff4a
3
+ size 3976730496
model-00010-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92c7f4ae60714f3ead424a74cc7a03ffd44e9223294ef133c6a94eeb5a67dd5c
3
+ size 3911384176
model-00011-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1840f5ff46151beab6443a87af59d3ca5c8e8be26e36fcdc2eef64e85d931757
3
+ size 3976730496
model-00012-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5bfe3444a959bd0bbbe245200606459d3aab738ec10d1cf1745e9af98d112669
3
+ size 3911384176
model-00013-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b7e02b199ec26897648b458afbe41fc527ec089dd14197df70fcf3e6f091a93
3
+ size 3976730496
model-00014-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ecd5374959efd4645744d172c0e27f99df85d9f93c0794a7aa795a8395911aa8
3
+ size 3911384176
model-00015-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3fc0de0f482754ba39b184b3e2a3c372366989c286754e88b80c4aa578ce7304
3
+ size 3976730496
model-00016-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e6fecc782563327a676675dc9cf2226d49808d9e91401eb179bfc133f2cbd65
3
+ size 876481824
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
quant_log.csv ADDED
@@ -0,0 +1,449 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ layer,module,loss,samples,damp,time
2
+ 0,self_attn.k_proj,0.00237118,0.01000,3.127
3
+ 0,self_attn.v_proj,0.00014123,0.01000,2.779
4
+ 0,self_attn.q_proj,0.01183380,0.01000,3.106
5
+ 0,self_attn.o_proj,0.00028543,0.01000,3.116
6
+ 0,mlp.up_proj,0.01609593,0.01000,4.300
7
+ 0,mlp.gate_proj,0.01776301,0.01000,4.316
8
+ 0,mlp.down_proj,0.00567739,0.01000,20.775
9
+ 1,self_attn.k_proj,0.01959774,0.01000,2.800
10
+ 1,self_attn.v_proj,0.00128063,0.01000,2.817
11
+ 1,self_attn.q_proj,0.06702538,0.01000,3.121
12
+ 1,self_attn.o_proj,0.00178431,0.01000,3.120
13
+ 1,mlp.up_proj,0.02137567,0.01000,4.330
14
+ 1,mlp.gate_proj,0.02512531,0.01000,4.328
15
+ 1,mlp.down_proj,0.35868970,0.01000,20.743
16
+ 2,self_attn.k_proj,0.00618781,0.01000,2.828
17
+ 2,self_attn.v_proj,0.00153573,0.01000,2.877
18
+ 2,self_attn.q_proj,0.02293800,0.01000,3.165
19
+ 2,self_attn.o_proj,0.00572403,0.01000,3.112
20
+ 2,mlp.up_proj,0.04553442,0.01000,4.327
21
+ 2,mlp.gate_proj,0.04639249,0.01000,4.331
22
+ 2,mlp.down_proj,0.00389441,0.01000,20.653
23
+ 3,self_attn.k_proj,0.02781182,0.01000,2.813
24
+ 3,self_attn.v_proj,0.00432311,0.01000,2.817
25
+ 3,self_attn.q_proj,0.06809862,0.01000,3.117
26
+ 3,self_attn.o_proj,0.00073100,0.01000,3.149
27
+ 3,mlp.up_proj,0.18160847,0.01000,4.325
28
+ 3,mlp.gate_proj,0.18848120,0.01000,4.327
29
+ 3,mlp.down_proj,0.01171539,0.01000,20.729
30
+ 4,self_attn.k_proj,0.02351411,0.01000,2.788
31
+ 4,self_attn.v_proj,0.00622305,0.01000,2.792
32
+ 4,self_attn.q_proj,0.08002016,0.01000,3.115
33
+ 4,self_attn.o_proj,0.00523799,0.01000,3.098
34
+ 4,mlp.up_proj,0.29500917,0.01000,4.324
35
+ 4,mlp.gate_proj,0.30944484,0.01000,4.320
36
+ 4,mlp.down_proj,0.01355328,0.01000,20.755
37
+ 5,self_attn.k_proj,0.03124500,0.01000,2.808
38
+ 5,self_attn.v_proj,0.00812443,0.01000,2.801
39
+ 5,self_attn.q_proj,0.11061060,0.01000,3.131
40
+ 5,self_attn.o_proj,0.00859872,0.01000,3.148
41
+ 5,mlp.up_proj,0.38112611,0.01000,4.316
42
+ 5,mlp.gate_proj,0.40232828,0.01000,4.329
43
+ 5,mlp.down_proj,0.02169481,0.01000,20.708
44
+ 6,self_attn.k_proj,0.05406944,0.01000,2.798
45
+ 6,self_attn.v_proj,0.01080229,0.01000,2.793
46
+ 6,self_attn.q_proj,0.20431036,0.01000,3.121
47
+ 6,self_attn.o_proj,0.02157875,0.01000,3.105
48
+ 6,mlp.up_proj,0.49781340,0.01000,4.311
49
+ 6,mlp.gate_proj,0.51407611,0.01000,4.306
50
+ 6,mlp.down_proj,0.03056708,0.01000,20.723
51
+ 7,self_attn.k_proj,0.13908443,0.01000,2.801
52
+ 7,self_attn.v_proj,0.02066772,0.01000,2.796
53
+ 7,self_attn.q_proj,0.34970999,0.01000,3.123
54
+ 7,self_attn.o_proj,0.00358162,0.01000,3.106
55
+ 7,mlp.up_proj,0.85166085,0.01000,4.323
56
+ 7,mlp.gate_proj,0.91031492,0.01000,4.338
57
+ 7,mlp.down_proj,0.06406650,0.01000,20.736
58
+ 8,self_attn.k_proj,0.07373472,0.01000,2.791
59
+ 8,self_attn.v_proj,0.01990465,0.01000,2.800
60
+ 8,self_attn.q_proj,0.30415455,0.01000,3.118
61
+ 8,self_attn.o_proj,0.01392892,0.01000,3.113
62
+ 8,mlp.up_proj,1.01542103,0.01000,4.316
63
+ 8,mlp.gate_proj,1.08411729,0.01000,4.321
64
+ 8,mlp.down_proj,0.07083103,0.01000,20.709
65
+ 9,self_attn.k_proj,0.10313431,0.01000,2.798
66
+ 9,self_attn.v_proj,0.02676009,0.01000,2.796
67
+ 9,self_attn.q_proj,0.43401644,0.01000,3.176
68
+ 9,self_attn.o_proj,0.02035430,0.01000,3.116
69
+ 9,mlp.up_proj,1.31375146,0.01000,4.320
70
+ 9,mlp.gate_proj,1.38119090,0.01000,4.343
71
+ 9,mlp.down_proj,0.08337718,0.01000,20.657
72
+ 10,self_attn.k_proj,0.14717038,0.01000,2.786
73
+ 10,self_attn.v_proj,0.03329685,0.01000,2.788
74
+ 10,self_attn.q_proj,0.59851396,0.01000,3.103
75
+ 10,self_attn.o_proj,0.02543477,0.01000,3.102
76
+ 10,mlp.up_proj,1.69003797,0.01000,4.308
77
+ 10,mlp.gate_proj,1.82192278,0.01000,4.314
78
+ 10,mlp.down_proj,0.13317853,0.01000,20.652
79
+ 11,self_attn.k_proj,0.27080441,0.01000,2.798
80
+ 11,self_attn.v_proj,0.05442170,0.01000,2.798
81
+ 11,self_attn.q_proj,0.79030275,0.01000,3.108
82
+ 11,self_attn.o_proj,0.00988693,0.01000,3.111
83
+ 11,mlp.up_proj,1.97805333,0.01000,4.326
84
+ 11,mlp.gate_proj,2.11406827,0.01000,4.327
85
+ 11,mlp.down_proj,0.15446520,0.01000,20.679
86
+ 12,self_attn.k_proj,0.17791295,0.01000,2.803
87
+ 12,self_attn.v_proj,0.04965502,0.01000,2.797
88
+ 12,self_attn.q_proj,0.81974483,0.01000,3.127
89
+ 12,self_attn.o_proj,0.03834637,0.01000,3.122
90
+ 12,mlp.up_proj,2.40238333,0.01000,4.315
91
+ 12,mlp.gate_proj,2.60567045,0.01000,4.324
92
+ 12,mlp.down_proj,0.21052358,0.01000,20.673
93
+ 13,self_attn.k_proj,0.22546425,0.01000,2.802
94
+ 13,self_attn.v_proj,0.05633362,0.01000,2.811
95
+ 13,self_attn.q_proj,0.98123246,0.01000,3.103
96
+ 13,self_attn.o_proj,0.05594819,0.01000,3.092
97
+ 13,mlp.up_proj,2.58392668,0.01000,4.321
98
+ 13,mlp.gate_proj,2.77088070,0.01000,4.323
99
+ 13,mlp.down_proj,0.23992851,0.01000,20.574
100
+ 14,self_attn.k_proj,0.22727673,0.01000,2.791
101
+ 14,self_attn.v_proj,0.06221664,0.01000,2.801
102
+ 14,self_attn.q_proj,1.04620826,0.01000,3.117
103
+ 14,self_attn.o_proj,0.07045884,0.01000,3.086
104
+ 14,mlp.up_proj,2.71724558,0.01000,4.318
105
+ 14,mlp.gate_proj,2.93374062,0.01000,4.320
106
+ 14,mlp.down_proj,0.28235990,0.01000,20.659
107
+ 15,self_attn.k_proj,0.52308160,0.01000,2.799
108
+ 15,self_attn.v_proj,0.08687409,0.01000,2.818
109
+ 15,self_attn.q_proj,1.81790495,0.01000,3.104
110
+ 15,self_attn.o_proj,0.03997960,0.01000,3.090
111
+ 15,mlp.up_proj,3.49855995,0.01000,4.325
112
+ 15,mlp.gate_proj,3.81484175,0.01000,4.320
113
+ 15,mlp.down_proj,0.37759125,0.01000,20.719
114
+ 16,self_attn.k_proj,0.26788694,0.01000,2.806
115
+ 16,self_attn.v_proj,0.06629637,0.01000,2.816
116
+ 16,self_attn.q_proj,1.27799642,0.01000,3.115
117
+ 16,self_attn.o_proj,0.09724845,0.01000,3.103
118
+ 16,mlp.up_proj,3.24072337,0.01000,4.326
119
+ 16,mlp.gate_proj,3.47319508,0.01000,4.323
120
+ 16,mlp.down_proj,0.40419251,0.01000,20.757
121
+ 17,self_attn.k_proj,0.30704799,0.01000,2.795
122
+ 17,self_attn.v_proj,0.09384538,0.01000,2.817
123
+ 17,self_attn.q_proj,1.51047468,0.01000,3.114
124
+ 17,self_attn.o_proj,0.14133909,0.01000,3.112
125
+ 17,mlp.up_proj,3.72232294,0.01000,4.320
126
+ 17,mlp.gate_proj,4.00111389,0.01000,4.320
127
+ 17,mlp.down_proj,0.48936534,0.01000,20.733
128
+ 18,self_attn.k_proj,0.39536184,0.01000,2.796
129
+ 18,self_attn.v_proj,0.10785440,0.01000,2.789
130
+ 18,self_attn.q_proj,1.79001546,0.01000,3.119
131
+ 18,self_attn.o_proj,0.25551438,0.01000,3.101
132
+ 18,mlp.up_proj,4.27974701,0.01000,4.319
133
+ 18,mlp.gate_proj,4.59731483,0.01000,4.331
134
+ 18,mlp.down_proj,0.68496990,0.01000,20.730
135
+ 19,self_attn.k_proj,0.68125188,0.01000,2.824
136
+ 19,self_attn.v_proj,0.19433120,0.01000,2.820
137
+ 19,self_attn.q_proj,2.58441305,0.01000,3.122
138
+ 19,self_attn.o_proj,0.06257914,0.01000,3.093
139
+ 19,mlp.up_proj,6.02338457,0.01000,4.321
140
+ 19,mlp.gate_proj,6.82512856,0.01000,4.322
141
+ 19,mlp.down_proj,0.96757936,0.01000,20.767
142
+ 20,self_attn.k_proj,0.47361368,0.01000,2.796
143
+ 20,self_attn.v_proj,0.14743000,0.01000,2.793
144
+ 20,self_attn.q_proj,2.37846446,0.01000,3.116
145
+ 20,self_attn.o_proj,0.23110107,0.01000,3.078
146
+ 20,mlp.up_proj,6.47508574,0.01000,4.321
147
+ 20,mlp.gate_proj,7.07527924,0.01000,4.317
148
+ 20,mlp.down_proj,1.00784147,0.01000,20.825
149
+ 21,self_attn.k_proj,0.62638307,0.01000,2.791
150
+ 21,self_attn.v_proj,0.16114473,0.01000,2.814
151
+ 21,self_attn.q_proj,2.91944075,0.01000,3.100
152
+ 21,self_attn.o_proj,0.27754447,0.01000,3.095
153
+ 21,mlp.up_proj,7.26681948,0.01000,4.322
154
+ 21,mlp.gate_proj,7.88508415,0.01000,4.323
155
+ 21,mlp.down_proj,1.13626218,0.01000,20.800
156
+ 22,self_attn.k_proj,0.49365854,0.01000,2.790
157
+ 22,self_attn.v_proj,0.18713154,0.01000,2.818
158
+ 22,self_attn.q_proj,2.64725590,0.01000,3.104
159
+ 22,self_attn.o_proj,0.38021657,0.01000,3.112
160
+ 22,mlp.up_proj,7.15935612,0.01000,4.317
161
+ 22,mlp.gate_proj,7.51539850,0.01000,4.333
162
+ 22,mlp.down_proj,1.21487689,0.01000,20.779
163
+ 23,self_attn.k_proj,0.97644448,0.01000,2.803
164
+ 23,self_attn.v_proj,0.31881452,0.01000,2.806
165
+ 23,self_attn.q_proj,3.98502779,0.01000,3.114
166
+ 23,self_attn.o_proj,0.19667208,0.01000,3.109
167
+ 23,mlp.up_proj,8.73191643,0.01000,4.325
168
+ 23,mlp.gate_proj,9.20885563,0.01000,4.331
169
+ 23,mlp.down_proj,1.61171770,0.01000,20.771
170
+ 24,self_attn.k_proj,0.74062228,0.01000,2.808
171
+ 24,self_attn.v_proj,0.26684782,0.01000,2.796
172
+ 24,self_attn.q_proj,3.40645289,0.01000,3.119
173
+ 24,self_attn.o_proj,0.70490086,0.01000,3.121
174
+ 24,mlp.up_proj,8.75358868,0.01000,4.326
175
+ 24,mlp.gate_proj,9.01323509,0.01000,4.326
176
+ 24,mlp.down_proj,2.02982712,0.01000,20.856
177
+ 25,self_attn.k_proj,0.68248963,0.01000,2.806
178
+ 25,self_attn.v_proj,0.28857774,0.01000,2.878
179
+ 25,self_attn.q_proj,3.46604204,0.01000,3.167
180
+ 25,self_attn.o_proj,0.49425381,0.01000,3.115
181
+ 25,mlp.up_proj,9.47107220,0.01000,4.321
182
+ 25,mlp.gate_proj,10.28519821,0.01000,4.324
183
+ 25,mlp.down_proj,2.07002401,0.01000,20.835
184
+ 26,self_attn.k_proj,0.82497019,0.01000,2.820
185
+ 26,self_attn.v_proj,0.31065938,0.01000,2.802
186
+ 26,self_attn.q_proj,4.06170321,0.01000,3.150
187
+ 26,self_attn.o_proj,0.43173498,0.01000,3.119
188
+ 26,mlp.up_proj,11.16811943,0.01000,4.321
189
+ 26,mlp.gate_proj,12.52889824,0.01000,4.305
190
+ 26,mlp.down_proj,2.20023608,0.01000,20.845
191
+ 27,self_attn.k_proj,1.22410727,0.01000,2.825
192
+ 27,self_attn.v_proj,0.43076056,0.01000,2.798
193
+ 27,self_attn.q_proj,5.18575001,0.01000,3.123
194
+ 27,self_attn.o_proj,0.27877492,0.01000,3.096
195
+ 27,mlp.up_proj,12.58386040,0.01000,4.327
196
+ 27,mlp.gate_proj,13.95620251,0.01000,4.333
197
+ 27,mlp.down_proj,2.38565969,0.01000,20.839
198
+ 28,self_attn.k_proj,0.85403657,0.01000,2.806
199
+ 28,self_attn.v_proj,0.29076397,0.01000,2.787
200
+ 28,self_attn.q_proj,4.03672314,0.01000,3.101
201
+ 28,self_attn.o_proj,0.59794313,0.01000,3.104
202
+ 28,mlp.up_proj,11.74896240,0.01000,4.320
203
+ 28,mlp.gate_proj,12.48750114,0.01000,4.319
204
+ 28,mlp.down_proj,3.22983122,0.01000,20.771
205
+ 29,self_attn.k_proj,0.79713917,0.01000,2.825
206
+ 29,self_attn.v_proj,0.28039247,0.01000,2.871
207
+ 29,self_attn.q_proj,3.93984890,0.01000,3.125
208
+ 29,self_attn.o_proj,0.78089821,0.01000,3.117
209
+ 29,mlp.up_proj,11.33225822,0.01000,4.317
210
+ 29,mlp.gate_proj,11.53958988,0.01000,4.334
211
+ 29,mlp.down_proj,2.68685436,0.01000,20.762
212
+ 30,self_attn.k_proj,0.79357457,0.01000,2.792
213
+ 30,self_attn.v_proj,0.34253484,0.01000,2.788
214
+ 30,self_attn.q_proj,4.11761475,0.01000,3.112
215
+ 30,self_attn.o_proj,0.91359365,0.01000,3.125
216
+ 30,mlp.up_proj,11.63817596,0.01000,4.321
217
+ 30,mlp.gate_proj,11.82233524,0.01000,4.313
218
+ 30,mlp.down_proj,3.03167748,0.01000,20.861
219
+ 31,self_attn.k_proj,1.21440160,0.01000,2.801
220
+ 31,self_attn.v_proj,0.56404942,0.01000,2.832
221
+ 31,self_attn.q_proj,5.67956305,0.01000,3.118
222
+ 31,self_attn.o_proj,0.48865020,0.01000,3.115
223
+ 31,mlp.up_proj,14.14960670,0.01000,4.329
224
+ 31,mlp.gate_proj,14.72683907,0.01000,4.333
225
+ 31,mlp.down_proj,3.68465829,0.01000,20.774
226
+ 32,self_attn.k_proj,0.90606254,0.01000,2.826
227
+ 32,self_attn.v_proj,0.38378727,0.01000,2.815
228
+ 32,self_attn.q_proj,4.70458794,0.01000,3.093
229
+ 32,self_attn.o_proj,1.31476831,0.01000,3.092
230
+ 32,mlp.up_proj,13.16320038,0.01000,4.322
231
+ 32,mlp.gate_proj,13.45796967,0.01000,4.336
232
+ 32,mlp.down_proj,4.18441963,0.01000,20.792
233
+ 33,self_attn.k_proj,0.76760435,0.01000,2.822
234
+ 33,self_attn.v_proj,0.42847544,0.01000,2.819
235
+ 33,self_attn.q_proj,4.56134892,0.01000,3.119
236
+ 33,self_attn.o_proj,1.08251286,0.01000,3.113
237
+ 33,mlp.up_proj,14.34337330,0.01000,4.323
238
+ 33,mlp.gate_proj,14.79744720,0.01000,4.333
239
+ 33,mlp.down_proj,4.69971466,0.01000,20.854
240
+ 34,self_attn.k_proj,0.89555413,0.01000,2.810
241
+ 34,self_attn.v_proj,0.52807939,0.01000,2.849
242
+ 34,self_attn.q_proj,5.28146935,0.01000,3.121
243
+ 34,self_attn.o_proj,1.48747659,0.01000,3.119
244
+ 34,mlp.up_proj,15.42789745,0.01000,4.258
245
+ 34,mlp.gate_proj,16.50379753,0.01000,4.254
246
+ 34,mlp.down_proj,5.58877373,0.01000,19.393
247
+ 35,self_attn.k_proj,1.37647629,0.01000,2.770
248
+ 35,self_attn.v_proj,0.78910184,0.01000,2.944
249
+ 35,self_attn.q_proj,7.08261204,0.01000,3.076
250
+ 35,self_attn.o_proj,1.01057386,0.01000,3.073
251
+ 35,mlp.up_proj,18.30294037,0.01000,4.269
252
+ 35,mlp.gate_proj,20.64340401,0.01000,4.277
253
+ 35,mlp.down_proj,6.54550982,0.01000,20.593
254
+ 36,self_attn.k_proj,0.97656679,0.01000,2.811
255
+ 36,self_attn.v_proj,0.57391113,0.01000,2.792
256
+ 36,self_attn.q_proj,5.68178225,0.01000,3.112
257
+ 36,self_attn.o_proj,1.29969442,0.01000,3.110
258
+ 36,mlp.up_proj,17.94355011,0.01000,4.299
259
+ 36,mlp.gate_proj,20.13683701,0.01000,4.304
260
+ 36,mlp.down_proj,7.21205807,0.01000,20.909
261
+ 37,self_attn.k_proj,0.94301772,0.01000,2.847
262
+ 37,self_attn.v_proj,0.57824910,0.01000,2.932
263
+ 37,self_attn.q_proj,5.95239687,0.01000,3.128
264
+ 37,self_attn.o_proj,1.37940025,0.01000,3.131
265
+ 37,mlp.up_proj,18.52091217,0.01000,4.337
266
+ 37,mlp.gate_proj,21.02642441,0.01000,4.325
267
+ 37,mlp.down_proj,7.64422464,0.01000,27.377
268
+ 38,self_attn.k_proj,1.03581893,0.01000,2.804
269
+ 38,self_attn.v_proj,0.70786691,0.01000,2.806
270
+ 38,self_attn.q_proj,6.59887505,0.01000,3.119
271
+ 38,self_attn.o_proj,1.17235184,0.01000,3.099
272
+ 38,mlp.up_proj,20.69558716,0.01000,4.387
273
+ 38,mlp.gate_proj,24.59513092,0.01000,4.339
274
+ 38,mlp.down_proj,7.93910074,0.01000,20.944
275
+ 39,self_attn.k_proj,1.52943969,0.01000,2.791
276
+ 39,self_attn.v_proj,1.04627979,0.01000,2.928
277
+ 39,self_attn.q_proj,8.58604050,0.01000,3.132
278
+ 39,self_attn.o_proj,0.76626897,0.01000,3.121
279
+ 39,mlp.up_proj,23.22537613,0.01000,4.328
280
+ 39,mlp.gate_proj,28.91666985,0.01000,4.343
281
+ 39,mlp.down_proj,8.40031242,0.01000,20.985
282
+ 40,self_attn.k_proj,1.09509122,0.01000,2.802
283
+ 40,self_attn.v_proj,0.78926635,0.01000,2.807
284
+ 40,self_attn.q_proj,7.18499231,0.01000,3.124
285
+ 40,self_attn.o_proj,1.27122092,0.01000,3.120
286
+ 40,mlp.up_proj,22.79548645,0.01000,4.303
287
+ 40,mlp.gate_proj,27.59724045,0.01000,4.323
288
+ 40,mlp.down_proj,8.57806015,0.01000,20.957
289
+ 41,self_attn.k_proj,1.19796991,0.01000,2.807
290
+ 41,self_attn.v_proj,0.87375212,0.01000,2.809
291
+ 41,self_attn.q_proj,7.69725323,0.01000,3.129
292
+ 41,self_attn.o_proj,1.05193281,0.01000,3.112
293
+ 41,mlp.up_proj,23.96932602,0.01000,4.322
294
+ 41,mlp.gate_proj,29.43371964,0.01000,4.324
295
+ 41,mlp.down_proj,8.79974461,0.01000,20.929
296
+ 42,self_attn.k_proj,1.21347082,0.01000,2.815
297
+ 42,self_attn.v_proj,0.80230772,0.01000,2.818
298
+ 42,self_attn.q_proj,7.82734251,0.01000,3.113
299
+ 42,self_attn.o_proj,1.46261120,0.01000,3.123
300
+ 42,mlp.up_proj,24.96371841,0.01000,4.324
301
+ 42,mlp.gate_proj,31.55431366,0.01000,4.319
302
+ 42,mlp.down_proj,8.53590584,0.01000,20.950
303
+ 43,self_attn.k_proj,1.56291699,0.01000,2.821
304
+ 43,self_attn.v_proj,1.20205545,0.01000,2.870
305
+ 43,self_attn.q_proj,9.32339096,0.01000,3.125
306
+ 43,self_attn.o_proj,0.53463578,0.01000,3.115
307
+ 43,mlp.up_proj,26.88477707,0.01000,4.308
308
+ 43,mlp.gate_proj,35.31211853,0.01000,4.306
309
+ 43,mlp.down_proj,8.85589981,0.01000,21.078
310
+ 44,self_attn.k_proj,1.21192908,0.01000,2.816
311
+ 44,self_attn.v_proj,0.87010479,0.01000,8.459
312
+ 44,self_attn.q_proj,8.13472939,0.01000,3.108
313
+ 44,self_attn.o_proj,1.07565022,0.01000,3.149
314
+ 44,mlp.up_proj,26.33778381,0.01000,4.323
315
+ 44,mlp.gate_proj,33.19284058,0.01000,4.315
316
+ 44,mlp.down_proj,8.94399166,0.01000,20.928
317
+ 45,self_attn.k_proj,1.26758444,0.01000,2.829
318
+ 45,self_attn.v_proj,0.93762308,0.01000,2.807
319
+ 45,self_attn.q_proj,8.30747700,0.01000,3.119
320
+ 45,self_attn.o_proj,1.35970306,0.01000,3.119
321
+ 45,mlp.up_proj,27.30161667,0.01000,4.332
322
+ 45,mlp.gate_proj,34.61208344,0.01000,4.326
323
+ 45,mlp.down_proj,8.97346687,0.01000,20.960
324
+ 46,self_attn.k_proj,1.28506112,0.01000,2.826
325
+ 46,self_attn.v_proj,0.94637656,0.01000,2.919
326
+ 46,self_attn.q_proj,8.71655273,0.01000,3.110
327
+ 46,self_attn.o_proj,1.31473839,0.01000,3.114
328
+ 46,mlp.up_proj,28.32736397,0.01000,4.311
329
+ 46,mlp.gate_proj,35.69065475,0.01000,4.330
330
+ 46,mlp.down_proj,9.26610470,0.01000,20.941
331
+ 47,self_attn.k_proj,1.58398616,0.01000,2.807
332
+ 47,self_attn.v_proj,1.45605755,0.01000,2.839
333
+ 47,self_attn.q_proj,10.10481834,0.01000,3.130
334
+ 47,self_attn.o_proj,0.88219702,0.01000,3.114
335
+ 47,mlp.up_proj,31.00548363,0.01000,4.325
336
+ 47,mlp.gate_proj,40.00745392,0.01000,4.324
337
+ 47,mlp.down_proj,9.52927017,0.01000,20.986
338
+ 48,self_attn.k_proj,1.29730964,0.01000,2.865
339
+ 48,self_attn.v_proj,1.25468504,0.01000,2.812
340
+ 48,self_attn.q_proj,9.12980461,0.01000,3.120
341
+ 48,self_attn.o_proj,1.22588074,0.01000,3.108
342
+ 48,mlp.up_proj,31.16883087,0.01000,4.318
343
+ 48,mlp.gate_proj,38.73802948,0.01000,4.329
344
+ 48,mlp.down_proj,10.10922623,0.01000,20.966
345
+ 49,self_attn.k_proj,1.31353641,0.01000,2.841
346
+ 49,self_attn.v_proj,1.13106871,0.01000,2.797
347
+ 49,self_attn.q_proj,9.42452240,0.01000,3.110
348
+ 49,self_attn.o_proj,1.36230147,0.01000,3.127
349
+ 49,mlp.up_proj,32.10203171,0.01000,4.328
350
+ 49,mlp.gate_proj,39.23027039,0.01000,4.300
351
+ 49,mlp.down_proj,10.73592186,0.01000,21.070
352
+ 50,self_attn.k_proj,1.23155284,0.01000,2.797
353
+ 50,self_attn.v_proj,1.12254333,0.01000,2.824
354
+ 50,self_attn.q_proj,9.45220280,0.01000,3.122
355
+ 50,self_attn.o_proj,1.79769087,0.01000,3.131
356
+ 50,mlp.up_proj,33.46105576,0.01000,4.326
357
+ 50,mlp.gate_proj,40.31344223,0.01000,4.332
358
+ 50,mlp.down_proj,11.42701721,0.01000,21.021
359
+ 51,self_attn.k_proj,1.51977789,0.01000,2.816
360
+ 51,self_attn.v_proj,1.63614023,0.01000,2.820
361
+ 51,self_attn.q_proj,11.08936691,0.01000,3.126
362
+ 51,self_attn.o_proj,0.86253881,0.01000,3.123
363
+ 51,mlp.up_proj,36.72800064,0.01000,4.329
364
+ 51,mlp.gate_proj,44.42227173,0.01000,4.332
365
+ 51,mlp.down_proj,12.24649525,0.01000,21.035
366
+ 52,self_attn.k_proj,1.26305985,0.01000,2.806
367
+ 52,self_attn.v_proj,1.29762578,0.01000,2.801
368
+ 52,self_attn.q_proj,10.05467892,0.01000,3.121
369
+ 52,self_attn.o_proj,1.56924605,0.01000,3.097
370
+ 52,mlp.up_proj,36.72959900,0.01000,4.331
371
+ 52,mlp.gate_proj,42.71206284,0.01000,4.334
372
+ 52,mlp.down_proj,13.34146309,0.01000,21.000
373
+ 53,self_attn.k_proj,1.24679852,0.01000,2.797
374
+ 53,self_attn.v_proj,1.46717882,0.01000,2.826
375
+ 53,self_attn.q_proj,9.73361015,0.01000,3.124
376
+ 53,self_attn.o_proj,2.62760925,0.01000,3.113
377
+ 53,mlp.up_proj,39.03444290,0.01000,4.331
378
+ 53,mlp.gate_proj,44.58834839,0.01000,4.329
379
+ 53,mlp.down_proj,15.31289387,0.01000,21.021
380
+ 54,self_attn.k_proj,1.19278765,0.01000,2.811
381
+ 54,self_attn.v_proj,1.44246030,0.01000,2.884
382
+ 54,self_attn.q_proj,10.12059593,0.01000,3.082
383
+ 54,self_attn.o_proj,3.47702193,0.01000,3.116
384
+ 54,mlp.up_proj,40.57743835,0.01000,4.330
385
+ 54,mlp.gate_proj,45.48489380,0.01000,4.330
386
+ 54,mlp.down_proj,17.86631775,0.01000,20.980
387
+ 55,self_attn.k_proj,1.45634103,0.01000,2.800
388
+ 55,self_attn.v_proj,2.27894878,0.01000,2.809
389
+ 55,self_attn.q_proj,12.08972454,0.01000,3.111
390
+ 55,self_attn.o_proj,2.46146488,0.01000,3.113
391
+ 55,mlp.up_proj,45.36315536,0.01000,4.315
392
+ 55,mlp.gate_proj,50.01605988,0.01000,4.321
393
+ 55,mlp.down_proj,21.87582016,0.01000,21.069
394
+ 56,self_attn.k_proj,1.10241914,0.01000,2.789
395
+ 56,self_attn.v_proj,1.82905793,0.01000,2.869
396
+ 56,self_attn.q_proj,10.14868736,0.01000,3.114
397
+ 56,self_attn.o_proj,4.03703833,0.01000,3.105
398
+ 56,mlp.up_proj,45.74471283,0.01000,4.324
399
+ 56,mlp.gate_proj,49.25518036,0.01000,4.334
400
+ 56,mlp.down_proj,26.32149887,0.01000,20.717
401
+ 57,self_attn.k_proj,0.94596267,0.01000,2.809
402
+ 57,self_attn.v_proj,1.91650033,0.01000,2.809
403
+ 57,self_attn.q_proj,9.83541203,0.01000,3.102
404
+ 57,self_attn.o_proj,4.67447901,0.01000,3.080
405
+ 57,mlp.up_proj,45.34143829,0.01000,4.319
406
+ 57,mlp.gate_proj,47.47423553,0.01000,4.314
407
+ 57,mlp.down_proj,31.68601799,0.01000,20.718
408
+ 58,self_attn.k_proj,0.90078247,0.01000,2.780
409
+ 58,self_attn.v_proj,2.12932301,0.01000,2.797
410
+ 58,self_attn.q_proj,10.10538292,0.01000,3.088
411
+ 58,self_attn.o_proj,6.05742645,0.01000,3.095
412
+ 58,mlp.up_proj,49.03110886,0.01000,4.318
413
+ 58,mlp.gate_proj,50.97365189,0.01000,4.315
414
+ 58,mlp.down_proj,36.67863464,0.01000,20.820
415
+ 59,self_attn.k_proj,1.21062970,0.01000,2.789
416
+ 59,self_attn.v_proj,3.84760022,0.01000,2.876
417
+ 59,self_attn.q_proj,12.37336540,0.01000,3.112
418
+ 59,self_attn.o_proj,5.57980919,0.01000,3.113
419
+ 59,mlp.up_proj,54.85280609,0.01000,4.317
420
+ 59,mlp.gate_proj,56.50590897,0.01000,4.311
421
+ 59,mlp.down_proj,44.20305634,0.01000,20.808
422
+ 60,self_attn.k_proj,0.91982675,0.01000,2.804
423
+ 60,self_attn.v_proj,2.71451902,0.01000,2.832
424
+ 60,self_attn.q_proj,10.73928070,0.01000,3.116
425
+ 60,self_attn.o_proj,7.95396280,0.01000,3.115
426
+ 60,mlp.up_proj,55.34839249,0.01000,4.333
427
+ 60,mlp.gate_proj,54.90727615,0.01000,4.306
428
+ 60,mlp.down_proj,56.50479507,0.01000,20.848
429
+ 61,self_attn.k_proj,0.85721207,0.01000,2.817
430
+ 61,self_attn.v_proj,2.78861809,0.01000,2.835
431
+ 61,self_attn.q_proj,10.20442200,0.01000,3.115
432
+ 61,self_attn.o_proj,12.74089622,0.01000,3.184
433
+ 61,mlp.up_proj,57.27671051,0.01000,4.320
434
+ 61,mlp.gate_proj,54.85419083,0.01000,4.326
435
+ 61,mlp.down_proj,67.24620819,0.01000,20.779
436
+ 62,self_attn.k_proj,2.68184018,0.01000,2.817
437
+ 62,self_attn.v_proj,4.22722721,0.01000,2.808
438
+ 62,self_attn.q_proj,6.90515041,0.01000,3.110
439
+ 62,self_attn.o_proj,42.22541809,0.01000,3.081
440
+ 62,mlp.up_proj,193.36753845,0.01000,4.308
441
+ 62,mlp.gate_proj,227.66082764,0.01000,4.319
442
+ 62,mlp.down_proj,26389.76171875,0.01000,20.821
443
+ 63,self_attn.k_proj,0.09801108,0.01000,2.790
444
+ 63,self_attn.v_proj,0.08450449,0.01000,2.803
445
+ 63,self_attn.q_proj,0.52656490,0.01000,3.114
446
+ 63,self_attn.o_proj,0.09586952,0.01000,3.108
447
+ 63,mlp.up_proj,37.69497299,0.01000,4.316
448
+ 63,mlp.gate_proj,42.26115417,0.01000,4.315
449
+ 63,mlp.down_proj,48684.17187500,0.01250,21.967
quantize_config.json ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bits": 4,
3
+ "group_size": 128,
4
+ "desc_act": true,
5
+ "sym": true,
6
+ "lm_head": false,
7
+ "quant_method": "gptq",
8
+ "checkpoint_format": "gptq",
9
+ "pack_dtype": "int32",
10
+ "meta": {
11
+ "quantizer": [
12
+ "gptqmodel:2.3.0-dev"
13
+ ],
14
+ "uri": "https://github.com/modelcloud/gptqmodel",
15
+ "damp_percent": 0.01,
16
+ "damp_auto_increment": 0.0025,
17
+ "static_groups": false,
18
+ "true_sequential": true,
19
+ "mse": 0.0
20
+ }
21
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<BOS_TOKEN>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|END_OF_TURN_TOKEN|>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<PAD>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "<UNK>",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "additional_special_tokens": [
31
+ "<|START_RESPONSE|>",
32
+ "<|END_RESPONSE|>"
33
+ ]
34
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5bad1341109519f60380bdefb784dbe92e1e044390109143baddf1fa53539e47
3
+ size 19597349
tokenizer_config.json ADDED
@@ -0,0 +1,364 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": false,
5
+ "clean_up_tokenization_spaces": false,
6
+ "legacy": true,
7
+ "spaces_between_special_tokens": false,
8
+ "use_default_system_prompt": false,
9
+ "bos_token": "<BOS_TOKEN>",
10
+ "eos_token": "<|END_OF_TURN_TOKEN|>",
11
+ "pad_token": "<PAD>",
12
+ "unk_token": "<UNK>",
13
+ "tokenizer_class": "CohereTokenizerFast",
14
+ "model_max_length": 1000000000000000019884624838656,
15
+ "sp_model_kwargs": {},
16
+ "chat_template": [
17
+ {
18
+ "name": "default",
19
+ "template": "{{ bos_token }}{% if documents %}\n{% set tools = [] %}\n{%- macro document_turn(documents) -%}\n{# format documents into chat turn #}\n<|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|><|START_THINKING|>I will look through the document to address the users needs.<|END_THINKING|><|START_ACTION|>[\n {\"tool_call_id\": \"0\", \"tool_name\": \"direct-injected-document\", \"parameters\": {}}\n]<|END_ACTION|><|END_OF_TURN_TOKEN|><|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|><|START_TOOL_RESULT|>[\n {\n \"tool_call_id\": \"0\",\n \"results\": {\n{% for doc in documents %}\n \"{{ loop.index0 }}\": {{doc|tojson}}{% if not loop.last %},\n {% endif %}\n{% endfor %}\n\n },\n \"is_error\": null\n }\n]<|END_TOOL_RESULT|><|END_OF_TURN_TOKEN|>{%- endmacro %}\n{%- macro tool_call_id_to_int(messages, tool_call_id) %}\n{%- set counter = namespace(value=0) %}\n{%- set tool_call_id_seen = namespace(value=false) %}\n{%- for msg in messages %}\n {%- if msg.tool_calls %}\n {%- for tool_call in msg.tool_calls %}\n {%- if tool_call.id == tool_call_id and not tool_call_id_seen.value -%}\n {{ counter.value }}\n {%- set tool_call_id_seen.value = true %}\n {%- endif %}\n {%- set counter.value = counter.value + 1 %}\n {%- endfor %}\n {%- endif %}\n{%- endfor %}\n{%- endmacro %}\n{%- macro format_tool_message(messages, tool_msg) -%}\n{# format tool message #}\n {\n \"tool_call_id\": \"{{ tool_call_id_to_int(messages, tool_msg.tool_call_id) }}\",\n \"results\": {\n \"0\": {{ tool_msg.content|tojson }}\n },\n \"is_error\": null\n }\n{%- endmacro -%}\n{%- if messages and messages[0]['role']|lower == 'system' %}{%- set developer_preamble = messages[0]['content'] %}{% endif %}\n{%- set tool_idx = namespace(value=0) %}\n{%- set tool_ids_seen = namespace(value=[]) %}\n{%- set sent_documents = namespace(value=false) %}\n<|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|># System Preamble\nYou are in contextual safety mode. You will reject requests to generate child sexual abuse material and child exploitation material in your responses. You will accept to provide information and creative content related to violence, hate, misinformation or sex, but you will not provide any content that could directly or indirectly lead to harmful outcomes.\n\nYour information cutoff date is June 2024.\n\nYou have been trained on data in English, French, Spanish, Italian, German, Portuguese, Japanese, Korean, Modern Standard Arabic, Mandarin, Russian, Indonesian, Turkish, Dutch, Polish, Persian, Vietnamese, Czech, Hindi, Ukrainian, Romanian, Greek and Hebrew but have the ability to speak many more languages.\n{% if tools or documents %}\n\nYou have been trained to have advanced reasoning and tool-use capabilities and you should make best use of these skills to serve user's requests.\n\n## Tool Use\nThink about how you can make best use of the provided tools to help with the task and come up with a high level plan that you will execute first.\n\n0. Start by writing <|START_THINKING|> followed by a detailed step by step plan of how you will solve the problem. For each step explain your thinking fully and give details of required tool calls (if needed). Unless specified otherwise, you write your plan in natural language. When you finish, close it out with <|END_THINKING|>.\n You can optionally choose to skip this step when the user request is so straightforward to address that only a trivial plan would be needed.\n NOTE: You MUST skip this step when you are directly responding to the user's request without using any tools.\n\nThen carry out your plan by repeatedly executing the following steps.\n1. Action: write <|START_ACTION|> followed by a list of JSON-formatted tool calls, with each one containing \"tool_name\" and \"parameters\" fields.\n When there are multiple tool calls which are completely independent of each other (i.e. they can be executed in parallel), you should list them out all together in one step. When you finish, close it out with <|END_ACTION|>.\n2. Observation: you will then receive results of those tool calls in JSON format in the very next turn, wrapped around by <|START_TOOL_RESULT|> and <|END_TOOL_RESULT|>. Carefully observe those results and think about what to do next. Note that these results will be provided to you in a separate turn. NEVER hallucinate results.\n Every tool call produces a list of results (when a tool call produces no result or a single result, it'll still get wrapped inside a list). Each result is clearly linked to its originating tool call via its \"tool_call_id\".\n3. Reflection: start the next turn by writing <|START_THINKING|> followed by what you've figured out so far, any changes you need to make to your plan, and what you will do next. When you finish, close it out with <|END_THINKING|>.\n You can optionally choose to skip this step when everything is going according to plan and no special pieces of information or reasoning chains need to be recorded.\n NOTE: You MUST skip this step when you are done with tool-use actions and are ready to respond to the user.\n\nYou can repeat the above 3 steps multiple times (could be 0 times too if no suitable tool calls are available or needed), until you decide it's time to finally respond to the user.\n\n4. Response: then break out of the loop and write <|START_RESPONSE|> followed by a piece of text which serves as a response to the user's last request. Use all previous tool calls and results to help you when formulating your response. When you finish, close it out with <|END_RESPONSE|>.\n{% if enable_citations %}\n\n## Grounding\nImportantly, note that \"Reflection\" and \"Response\" above can be grounded.\nGrounding means you associate pieces of texts (called \"spans\") with those specific tool results that support them (called \"sources\"). And you use a pair of tags \"<co>\" and \"</co>\" to indicate when a span can be grounded onto a list of sources, listing them out in the closing tag. Sources from the same tool call are grouped together and listed as \"{tool_call_id}:[{list of result indices}]\", before they are joined together by \",\". E.g., \"<co>span</co: 0:[1,2],1:[0]>\" means that \"span\" is supported by result 1 and 2 from \"tool_call_id=0\" as well as result 0 from \"tool_call_id=1\".\n{% endif %}\n\n## Available Tools\nHere is the list of tools that you have available to you.\nYou can ONLY use the tools listed here. When a tool is not listed below, it is NOT available and you should NEVER attempt to use it.\nEach tool is represented as a JSON object with fields like \"name\", \"description\", \"parameters\" (per JSON Schema), and optionally, \"responses\" (per JSON Schema).\n\n```json\n[\n{% if documents %}\n {\"name\": \"direct-injected-document\", \"description\": \"This is a special tool to directly inject user-uploaded documents into the chat as additional context. DO NOT use this tool by yourself!\", \"parameters\": {\"type\": \"object\", \"properties\": {}, \"required\": []}, \"responses\": {\"200\": {\"description\": \"Successfully returned a list of chunked text snippets from the directly uploaded documents.\", \"content\": {\"application/json\": {\"schema\": {\"type\": \"array\", \"items\": {\"type\": \"object\", \"required\": [\"url\", \"snippet\"], \"properties\": {\"url\": {\"type\": \"string\", \"description\": \"The url of the uploaded document.\"}, \"snippet\": {\"type\": \"string\", \"description\": \"The text snippet for the returned document chunk.\"}}}}}}}}}{%- if tools %},{% endif %}\n\n{% endif %}\n{% for tool in tools %}\n {\"name\": \"{{ tool['function']['name'] }}\", \"description\": \"{{tool['function']['description']}}\", \"parameters\": {{ tool['function']['parameters']|tojson }}, \"responses\": null}{%- if not loop.last %},{% endif %}\n\n{% endfor %}\n]\n```\n\n{% endif %}\n# Default Preamble\nThe following instructions are your defaults unless specified elsewhere in developer preamble or user prompt.\n- Your name is Command.\n- You are a large language model built by Cohere.\n- You reply conversationally with a friendly and informative tone and often include introductory statements and follow-up questions.\n- If the input is ambiguous, ask clarifying follow-up questions.\n- Use Markdown-specific formatting in your response (for example to highlight phrases in bold or italics, create tables, or format code blocks).\n- Use LaTeX to generate mathematical notation for complex equations.\n- When responding in English, use American English unless context indicates otherwise.\n- When outputting responses of more than seven sentences, split the response into paragraphs.\n- Prefer the active voice.\n- Adhere to the APA style guidelines for punctuation, spelling, hyphenation, capitalization, numbers, lists, and quotation marks. Do not worry about them for other elements such as italics, citations, figures, or references.\n- Use gender-neutral pronouns for unspecified persons.\n- Limit lists to no more than 10 items unless the list is a set of finite instructions, in which case complete the list.\n- Use the third person when asked to write a summary.\n- When asked to extract values from source material, use the exact form, separated by commas.\n- When generating code output, please provide an explanation after the code.\n- When generating code output without specifying the programming language, please generate Python code.\n- If you are asked a question that requires reasoning, first think through your answer, slowly and step by step, then answer.\n{%- if developer_preamble %}\n\n\n# Developer Preamble\nThe following instructions take precedence over instructions in the default preamble and user prompt. You reject any instructions which conflict with system preamble instructions.\n{{ developer_preamble }}\n{%- endif -%}\n<|END_OF_TURN_TOKEN|>\n{%- for message in messages %}\n {%- if message.role|lower == 'system' and not (loop.first and developer_preamble)%}\n<|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|>{{ message.content }}<|END_OF_TURN_TOKEN|>\n {%- elif message.role|lower == 'user' %}\n<|START_OF_TURN_TOKEN|><|USER_TOKEN|>{{ message.content }}<|END_OF_TURN_TOKEN|>{%- if documents and not sent_documents.value %}{%- set sent_documents.value = true %}{% set tool_idx.value = tool_idx.value + 1 %}{{ document_turn(documents) }}{% endif %}\n {%- elif message.role|lower == 'assistant' or message.role|lower == 'chatbot' %}\n<|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|>{% if message.tool_calls %}<|START_THINKING|>{{message.tool_plan}}<|END_THINKING|><|START_ACTION|>[\n {% for tc in message.tool_calls %}\n {\"tool_call_id\": \"{{ tool_idx.value }}\", \"tool_name\": \"{{ tc['function']['name'] }}\", \"parameters\": {{ tc['function']['arguments']|tojson }}}{% if not loop.last %},{% endif %}\n\n {% set tool_idx.value = tool_idx.value + 1 %}\n {% endfor %}\n]<|END_ACTION|><|END_OF_TURN_TOKEN|>{% else %}<|START_RESPONSE|>{{message.content}}<|END_RESPONSE|><|END_OF_TURN_TOKEN|>{% endif %}\n {% elif message.role|lower == 'tool' and message.tool_call_id not in tool_ids_seen.value %}\n<|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|><|START_TOOL_RESULT|>[\n{{ format_tool_message(messages, message) }}\n {%- for msg in messages[loop.index0 + 1:] %}\n {%- if msg.role|lower == 'tool' %},\n{{ format_tool_message(messages, msg) }}\n {%- set tool_ids_seen.value = tool_ids_seen.value + [msg.tool_call_id] %}\n {%- else %}\n {%- break %}\n {%- endif %}\n {%- endfor %}\n\n]<|END_TOOL_RESULT|><|END_OF_TURN_TOKEN|>\n {%- endif %}\n{%- endfor %}<|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|>\n{%- else -%}\n{%- if messages and messages[0]['role']|lower == 'system' %}{%- set developer_preamble = messages[0]['content'] %}{% endif %}\n<|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|># System Preamble\n{% if safety_mode|upper == 'STRICT' -%}\nYou are in strict safety mode. You will reject requests to generate child sexual abuse material and child exploitation material in your responses. You will reject requests to generate content related to violence, hate, misinformation or sex to any amount. You will avoid using profanity. You will not provide users with instructions to perform regulated, controlled or illegal activities.\n{%- else -%}\nYou are in contextual safety mode. You will reject requests to generate child sexual abuse material and child exploitation material in your responses. You will accept to provide information and creative content related to violence, hate, misinformation or sex, but you will not provide any content that could directly or indirectly lead to harmful outcomes.\n{%- endif %}\n\n\nYour information cutoff date is June 2024.\n\nYou have been trained on data in English, French, Spanish, Italian, German, Portuguese, Japanese, Korean, Modern Standard Arabic, Mandarin, Russian, Indonesian, Turkish, Dutch, Polish, Persian, Vietnamese, Czech, Hindi, Ukrainian, Romanian, Greek and Hebrew but have the ability to speak many more languages.\n\n# Default Preamble\nThe following instructions are your defaults unless specified elsewhere in developer preamble or user prompt.\n- Your name is Command.\n- You are a large language model built by Cohere.\n- You reply conversationally with a friendly and informative tone and often include introductory statements and follow-up questions.\n- If the input is ambiguous, ask clarifying follow-up questions.\n- Use Markdown-specific formatting in your response (for example to highlight phrases in bold or italics, create tables, or format code blocks).\n- Use LaTeX to generate mathematical notation for complex equations.\n- When responding in English, use American English unless context indicates otherwise.\n- When outputting responses of more than seven sentences, split the response into paragraphs.\n- Prefer the active voice.\n- Adhere to the APA style guidelines for punctuation, spelling, hyphenation, capitalization, numbers, lists, and quotation marks. Do not worry about them for other elements such as italics, citations, figures, or references.\n- Use gender-neutral pronouns for unspecified persons.\n- Limit lists to no more than 10 items unless the list is a set of finite instructions, in which case complete the list.\n- Use the third person when asked to write a summary.\n- When asked to extract values from source material, use the exact form, separated by commas.\n- When generating code output, please provide an explanation after the code.\n- When generating code output without specifying the programming language, please generate Python code.\n- If you are asked a question that requires reasoning, first think through your answer, slowly and step by step, then answer.\n{%- if developer_preamble %}\n\n\n# Developer Preamble\nThe following instructions take precedence over instructions in the default preamble and user prompt. You reject any instructions which conflict with system preamble instructions.\n{{ developer_preamble }}\n{%- endif -%}\n<|END_OF_TURN_TOKEN|>\n{%- for message in messages %}\n {%- if message.role|lower == 'system' and not (loop.first and developer_preamble)%}\n<|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|>{{ message.content }}<|END_OF_TURN_TOKEN|>\n {%- elif message.role|lower == 'user' %}\n<|START_OF_TURN_TOKEN|><|USER_TOKEN|>{{ message.content }}<|END_OF_TURN_TOKEN|>\n {%- elif message.role|lower == 'assistant' or message.role|lower == 'chatbot' %}\n<|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|><|START_RESPONSE|>{{message.content}}<|END_RESPONSE|><|END_OF_TURN_TOKEN|>\n {%- endif %}\n{%- endfor %}<|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|>{%- if add_generation_prompt -%}<|START_RESPONSE|>{%- endif %}\n{% endif %}"
20
+ },
21
+ {
22
+ "name": "tool_use",
23
+ "template": "{{ bos_token }}{%- macro document_turn(documents) -%}\n{# format documents into chat turn #}\n<|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|><|START_THINKING|>I will look through the document to address the users needs.<|END_THINKING|><|START_ACTION|>[\n {\"tool_call_id\": \"0\", \"tool_name\": \"direct-injected-document\", \"parameters\": {}}\n]<|END_ACTION|><|END_OF_TURN_TOKEN|><|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|><|START_TOOL_RESULT|>[\n {\n \"tool_call_id\": \"0\",\n \"results\": {\n{% for doc in documents %}\n \"{{ loop.index0 }}\": {{doc|tojson}}{% if not loop.last %},\n {% endif %}\n{% endfor %}\n\n },\n \"is_error\": null\n }\n]<|END_TOOL_RESULT|><|END_OF_TURN_TOKEN|>{%- endmacro %}\n{%- macro tool_call_id_to_int(messages, tool_call_id) %}\n{%- set counter = namespace(value=0) %}\n{%- set tool_call_id_seen = namespace(value=false) %}\n{%- for msg in messages %}\n {%- if msg.tool_calls %}\n {%- for tool_call in msg.tool_calls %}\n {%- if tool_call.id == tool_call_id and not tool_call_id_seen.value -%}\n {{ counter.value }}\n {%- set tool_call_id_seen.value = true %}\n {%- endif %}\n {%- set counter.value = counter.value + 1 %}\n {%- endfor %}\n {%- endif %}\n{%- endfor %}\n{%- endmacro %}\n{%- macro format_tool_message(messages, tool_msg) -%}\n{# format tool message #}\n {\n \"tool_call_id\": \"{{ tool_call_id_to_int(messages, tool_msg.tool_call_id) }}\",\n \"results\": {\n \"0\": {{ tool_msg.content|tojson }}\n },\n \"is_error\": null\n }\n{%- endmacro -%}\n{%- if messages and messages[0]['role']|lower == 'system' %}{%- set developer_preamble = messages[0]['content'] %}{% endif %}\n{%- set tool_idx = namespace(value=0) %}\n{%- set tool_ids_seen = namespace(value=[]) %}\n{%- set sent_documents = namespace(value=false) %}\n<|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|># System Preamble\nYou are in contextual safety mode. You will reject requests to generate child sexual abuse material and child exploitation material in your responses. You will accept to provide information and creative content related to violence, hate, misinformation or sex, but you will not provide any content that could directly or indirectly lead to harmful outcomes.\n\nYour information cutoff date is June 2024.\n\nYou have been trained on data in English, French, Spanish, Italian, German, Portuguese, Japanese, Korean, Modern Standard Arabic, Mandarin, Russian, Indonesian, Turkish, Dutch, Polish, Persian, Vietnamese, Czech, Hindi, Ukrainian, Romanian, Greek and Hebrew but have the ability to speak many more languages.\n{% if tools or documents %}\n\nYou have been trained to have advanced reasoning and tool-use capabilities and you should make best use of these skills to serve user's requests.\n\n## Tool Use\nThink about how you can make best use of the provided tools to help with the task and come up with a high level plan that you will execute first.\n\n0. Start by writing <|START_THINKING|> followed by a detailed step by step plan of how you will solve the problem. For each step explain your thinking fully and give details of required tool calls (if needed). Unless specified otherwise, you write your plan in natural language. When you finish, close it out with <|END_THINKING|>.\n You can optionally choose to skip this step when the user request is so straightforward to address that only a trivial plan would be needed.\n NOTE: You MUST skip this step when you are directly responding to the user's request without using any tools.\n\nThen carry out your plan by repeatedly executing the following steps.\n1. Action: write <|START_ACTION|> followed by a list of JSON-formatted tool calls, with each one containing \"tool_name\" and \"parameters\" fields.\n When there are multiple tool calls which are completely independent of each other (i.e. they can be executed in parallel), you should list them out all together in one step. When you finish, close it out with <|END_ACTION|>.\n2. Observation: you will then receive results of those tool calls in JSON format in the very next turn, wrapped around by <|START_TOOL_RESULT|> and <|END_TOOL_RESULT|>. Carefully observe those results and think about what to do next. Note that these results will be provided to you in a separate turn. NEVER hallucinate results.\n Every tool call produces a list of results (when a tool call produces no result or a single result, it'll still get wrapped inside a list). Each result is clearly linked to its originating tool call via its \"tool_call_id\".\n3. Reflection: start the next turn by writing <|START_THINKING|> followed by what you've figured out so far, any changes you need to make to your plan, and what you will do next. When you finish, close it out with <|END_THINKING|>.\n You can optionally choose to skip this step when everything is going according to plan and no special pieces of information or reasoning chains need to be recorded.\n NOTE: You MUST skip this step when you are done with tool-use actions and are ready to respond to the user.\n\nYou can repeat the above 3 steps multiple times (could be 0 times too if no suitable tool calls are available or needed), until you decide it's time to finally respond to the user.\n\n4. Response: then break out of the loop and write <|START_RESPONSE|> followed by a piece of text which serves as a response to the user's last request. Use all previous tool calls and results to help you when formulating your response. When you finish, close it out with <|END_RESPONSE|>.\n{% if enable_citations %}\n\n## Grounding\nImportantly, note that \"Reflection\" and \"Response\" above can be grounded.\nGrounding means you associate pieces of texts (called \"spans\") with those specific tool results that support them (called \"sources\"). And you use a pair of tags \"<co>\" and \"</co>\" to indicate when a span can be grounded onto a list of sources, listing them out in the closing tag. Sources from the same tool call are grouped together and listed as \"{tool_call_id}:[{list of result indices}]\", before they are joined together by \",\". E.g., \"<co>span</co: 0:[1,2],1:[0]>\" means that \"span\" is supported by result 1 and 2 from \"tool_call_id=0\" as well as result 0 from \"tool_call_id=1\".\n{% endif %}\n\n## Available Tools\nHere is the list of tools that you have available to you.\nYou can ONLY use the tools listed here. When a tool is not listed below, it is NOT available and you should NEVER attempt to use it.\nEach tool is represented as a JSON object with fields like \"name\", \"description\", \"parameters\" (per JSON Schema), and optionally, \"responses\" (per JSON Schema).\n\n```json\n[\n{% if documents %}\n {\"name\": \"direct-injected-document\", \"description\": \"This is a special tool to directly inject user-uploaded documents into the chat as additional context. DO NOT use this tool by yourself!\", \"parameters\": {\"type\": \"object\", \"properties\": {}, \"required\": []}, \"responses\": {\"200\": {\"description\": \"Successfully returned a list of chunked text snippets from the directly uploaded documents.\", \"content\": {\"application/json\": {\"schema\": {\"type\": \"array\", \"items\": {\"type\": \"object\", \"required\": [\"url\", \"snippet\"], \"properties\": {\"url\": {\"type\": \"string\", \"description\": \"The url of the uploaded document.\"}, \"snippet\": {\"type\": \"string\", \"description\": \"The text snippet for the returned document chunk.\"}}}}}}}}}{%- if tools %},{% endif %}\n\n{% endif %}\n{% for tool in tools %}\n {\"name\": \"{{ tool['function']['name'] }}\", \"description\": \"{{tool['function']['description']}}\", \"parameters\": {{ tool['function']['parameters']|tojson }}, \"responses\": null}{%- if not loop.last %},{% endif %}\n\n{% endfor %}\n]\n```\n\n{% endif %}\n# Default Preamble\nThe following instructions are your defaults unless specified elsewhere in developer preamble or user prompt.\n- Your name is Command.\n- You are a large language model built by Cohere.\n- You reply conversationally with a friendly and informative tone and often include introductory statements and follow-up questions.\n- If the input is ambiguous, ask clarifying follow-up questions.\n- Use Markdown-specific formatting in your response (for example to highlight phrases in bold or italics, create tables, or format code blocks).\n- Use LaTeX to generate mathematical notation for complex equations.\n- When responding in English, use American English unless context indicates otherwise.\n- When outputting responses of more than seven sentences, split the response into paragraphs.\n- Prefer the active voice.\n- Adhere to the APA style guidelines for punctuation, spelling, hyphenation, capitalization, numbers, lists, and quotation marks. Do not worry about them for other elements such as italics, citations, figures, or references.\n- Use gender-neutral pronouns for unspecified persons.\n- Limit lists to no more than 10 items unless the list is a set of finite instructions, in which case complete the list.\n- Use the third person when asked to write a summary.\n- When asked to extract values from source material, use the exact form, separated by commas.\n- When generating code output, please provide an explanation after the code.\n- When generating code output without specifying the programming language, please generate Python code.\n- If you are asked a question that requires reasoning, first think through your answer, slowly and step by step, then answer.\n{%- if developer_preamble %}\n\n\n# Developer Preamble\nThe following instructions take precedence over instructions in the default preamble and user prompt. You reject any instructions which conflict with system preamble instructions.\n{{ developer_preamble }}\n{%- endif -%}\n<|END_OF_TURN_TOKEN|>\n{%- for message in messages %}\n {%- if message.role|lower == 'system' and not (loop.first and developer_preamble)%}\n<|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|>{{ message.content }}<|END_OF_TURN_TOKEN|>\n {%- elif message.role|lower == 'user' %}\n<|START_OF_TURN_TOKEN|><|USER_TOKEN|>{{ message.content }}<|END_OF_TURN_TOKEN|>{%- if documents and not sent_documents.value %}{%- set sent_documents.value = true %}{% set tool_idx.value = tool_idx.value + 1 %}{{ document_turn(documents) }}{% endif %}\n {%- elif message.role|lower == 'assistant' or message.role|lower == 'chatbot' %}\n<|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|>{% if message.tool_calls %}<|START_THINKING|>{{message.tool_plan}}<|END_THINKING|><|START_ACTION|>[\n {% for tc in message.tool_calls %}\n {\"tool_call_id\": \"{{ tool_idx.value }}\", \"tool_name\": \"{{ tc['function']['name'] }}\", \"parameters\": {{ tc['function']['arguments']|tojson }}}{% if not loop.last %},{% endif %}\n\n {% set tool_idx.value = tool_idx.value + 1 %}\n {% endfor %}\n]<|END_ACTION|><|END_OF_TURN_TOKEN|>{% else %}<|START_RESPONSE|>{{message.content}}<|END_RESPONSE|><|END_OF_TURN_TOKEN|>{% endif %}\n {% elif message.role|lower == 'tool' and message.tool_call_id not in tool_ids_seen.value %}\n<|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|><|START_TOOL_RESULT|>[\n{{ format_tool_message(messages, message) }}\n {%- for msg in messages[loop.index0 + 1:] %}\n {%- if msg.role|lower == 'tool' %},\n{{ format_tool_message(messages, msg) }}\n {%- set tool_ids_seen.value = tool_ids_seen.value + [msg.tool_call_id] %}\n {%- else %}\n {%- break %}\n {%- endif %}\n {%- endfor %}\n\n]<|END_TOOL_RESULT|><|END_OF_TURN_TOKEN|>\n {%- endif %}\n{%- endfor %}<|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|>"
24
+ },
25
+ {
26
+ "name": "rag",
27
+ "template": "{{ bos_token }}{% set tools = [] %}\n{%- macro document_turn(documents) -%}\n{# format documents into chat turn #}\n<|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|><|START_THINKING|>I will look through the document to address the users needs.<|END_THINKING|><|START_ACTION|>[\n {\"tool_call_id\": \"0\", \"tool_name\": \"direct-injected-document\", \"parameters\": {}}\n]<|END_ACTION|><|END_OF_TURN_TOKEN|><|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|><|START_TOOL_RESULT|>[\n {\n \"tool_call_id\": \"0\",\n \"results\": {\n{% for doc in documents %}\n \"{{ loop.index0 }}\": {{doc|tojson}}{% if not loop.last %},\n {% endif %}\n{% endfor %}\n\n },\n \"is_error\": null\n }\n]<|END_TOOL_RESULT|><|END_OF_TURN_TOKEN|>{%- endmacro %}\n{%- macro tool_call_id_to_int(messages, tool_call_id) %}\n{%- set counter = namespace(value=0) %}\n{%- set tool_call_id_seen = namespace(value=false) %}\n{%- for msg in messages %}\n {%- if msg.tool_calls %}\n {%- for tool_call in msg.tool_calls %}\n {%- if tool_call.id == tool_call_id and not tool_call_id_seen.value -%}\n {{ counter.value }}\n {%- set tool_call_id_seen.value = true %}\n {%- endif %}\n {%- set counter.value = counter.value + 1 %}\n {%- endfor %}\n {%- endif %}\n{%- endfor %}\n{%- endmacro %}\n{%- macro format_tool_message(messages, tool_msg) -%}\n{# format tool message #}\n {\n \"tool_call_id\": \"{{ tool_call_id_to_int(messages, tool_msg.tool_call_id) }}\",\n \"results\": {\n \"0\": {{ tool_msg.content|tojson }}\n },\n \"is_error\": null\n }\n{%- endmacro -%}\n{%- if messages and messages[0]['role']|lower == 'system' %}{%- set developer_preamble = messages[0]['content'] %}{% endif %}\n{%- set tool_idx = namespace(value=0) %}\n{%- set tool_ids_seen = namespace(value=[]) %}\n{%- set sent_documents = namespace(value=false) %}\n<|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|># System Preamble\nYou are in contextual safety mode. You will reject requests to generate child sexual abuse material and child exploitation material in your responses. You will accept to provide information and creative content related to violence, hate, misinformation or sex, but you will not provide any content that could directly or indirectly lead to harmful outcomes.\n\nYour information cutoff date is June 2024.\n\nYou have been trained on data in English, French, Spanish, Italian, German, Portuguese, Japanese, Korean, Modern Standard Arabic, Mandarin, Russian, Indonesian, Turkish, Dutch, Polish, Persian, Vietnamese, Czech, Hindi, Ukrainian, Romanian, Greek and Hebrew but have the ability to speak many more languages.\n{% if tools or documents %}\n\nYou have been trained to have advanced reasoning and tool-use capabilities and you should make best use of these skills to serve user's requests.\n\n## Tool Use\nThink about how you can make best use of the provided tools to help with the task and come up with a high level plan that you will execute first.\n\n0. Start by writing <|START_THINKING|> followed by a detailed step by step plan of how you will solve the problem. For each step explain your thinking fully and give details of required tool calls (if needed). Unless specified otherwise, you write your plan in natural language. When you finish, close it out with <|END_THINKING|>.\n You can optionally choose to skip this step when the user request is so straightforward to address that only a trivial plan would be needed.\n NOTE: You MUST skip this step when you are directly responding to the user's request without using any tools.\n\nThen carry out your plan by repeatedly executing the following steps.\n1. Action: write <|START_ACTION|> followed by a list of JSON-formatted tool calls, with each one containing \"tool_name\" and \"parameters\" fields.\n When there are multiple tool calls which are completely independent of each other (i.e. they can be executed in parallel), you should list them out all together in one step. When you finish, close it out with <|END_ACTION|>.\n2. Observation: you will then receive results of those tool calls in JSON format in the very next turn, wrapped around by <|START_TOOL_RESULT|> and <|END_TOOL_RESULT|>. Carefully observe those results and think about what to do next. Note that these results will be provided to you in a separate turn. NEVER hallucinate results.\n Every tool call produces a list of results (when a tool call produces no result or a single result, it'll still get wrapped inside a list). Each result is clearly linked to its originating tool call via its \"tool_call_id\".\n3. Reflection: start the next turn by writing <|START_THINKING|> followed by what you've figured out so far, any changes you need to make to your plan, and what you will do next. When you finish, close it out with <|END_THINKING|>.\n You can optionally choose to skip this step when everything is going according to plan and no special pieces of information or reasoning chains need to be recorded.\n NOTE: You MUST skip this step when you are done with tool-use actions and are ready to respond to the user.\n\nYou can repeat the above 3 steps multiple times (could be 0 times too if no suitable tool calls are available or needed), until you decide it's time to finally respond to the user.\n\n4. Response: then break out of the loop and write <|START_RESPONSE|> followed by a piece of text which serves as a response to the user's last request. Use all previous tool calls and results to help you when formulating your response. When you finish, close it out with <|END_RESPONSE|>.\n{% if enable_citations %}\n\n## Grounding\nImportantly, note that \"Reflection\" and \"Response\" above can be grounded.\nGrounding means you associate pieces of texts (called \"spans\") with those specific tool results that support them (called \"sources\"). And you use a pair of tags \"<co>\" and \"</co>\" to indicate when a span can be grounded onto a list of sources, listing them out in the closing tag. Sources from the same tool call are grouped together and listed as \"{tool_call_id}:[{list of result indices}]\", before they are joined together by \",\". E.g., \"<co>span</co: 0:[1,2],1:[0]>\" means that \"span\" is supported by result 1 and 2 from \"tool_call_id=0\" as well as result 0 from \"tool_call_id=1\".\n{% endif %}\n\n## Available Tools\nHere is the list of tools that you have available to you.\nYou can ONLY use the tools listed here. When a tool is not listed below, it is NOT available and you should NEVER attempt to use it.\nEach tool is represented as a JSON object with fields like \"name\", \"description\", \"parameters\" (per JSON Schema), and optionally, \"responses\" (per JSON Schema).\n\n```json\n[\n{% if documents %}\n {\"name\": \"direct-injected-document\", \"description\": \"This is a special tool to directly inject user-uploaded documents into the chat as additional context. DO NOT use this tool by yourself!\", \"parameters\": {\"type\": \"object\", \"properties\": {}, \"required\": []}, \"responses\": {\"200\": {\"description\": \"Successfully returned a list of chunked text snippets from the directly uploaded documents.\", \"content\": {\"application/json\": {\"schema\": {\"type\": \"array\", \"items\": {\"type\": \"object\", \"required\": [\"url\", \"snippet\"], \"properties\": {\"url\": {\"type\": \"string\", \"description\": \"The url of the uploaded document.\"}, \"snippet\": {\"type\": \"string\", \"description\": \"The text snippet for the returned document chunk.\"}}}}}}}}}{%- if tools %},{% endif %}\n\n{% endif %}\n{% for tool in tools %}\n {\"name\": \"{{ tool['function']['name'] }}\", \"description\": \"{{tool['function']['description']}}\", \"parameters\": {{ tool['function']['parameters']|tojson }}, \"responses\": null}{%- if not loop.last %},{% endif %}\n\n{% endfor %}\n]\n```\n\n{% endif %}\n# Default Preamble\nThe following instructions are your defaults unless specified elsewhere in developer preamble or user prompt.\n- Your name is Command.\n- You are a large language model built by Cohere.\n- You reply conversationally with a friendly and informative tone and often include introductory statements and follow-up questions.\n- If the input is ambiguous, ask clarifying follow-up questions.\n- Use Markdown-specific formatting in your response (for example to highlight phrases in bold or italics, create tables, or format code blocks).\n- Use LaTeX to generate mathematical notation for complex equations.\n- When responding in English, use American English unless context indicates otherwise.\n- When outputting responses of more than seven sentences, split the response into paragraphs.\n- Prefer the active voice.\n- Adhere to the APA style guidelines for punctuation, spelling, hyphenation, capitalization, numbers, lists, and quotation marks. Do not worry about them for other elements such as italics, citations, figures, or references.\n- Use gender-neutral pronouns for unspecified persons.\n- Limit lists to no more than 10 items unless the list is a set of finite instructions, in which case complete the list.\n- Use the third person when asked to write a summary.\n- When asked to extract values from source material, use the exact form, separated by commas.\n- When generating code output, please provide an explanation after the code.\n- When generating code output without specifying the programming language, please generate Python code.\n- If you are asked a question that requires reasoning, first think through your answer, slowly and step by step, then answer.\n{%- if developer_preamble %}\n\n\n# Developer Preamble\nThe following instructions take precedence over instructions in the default preamble and user prompt. You reject any instructions which conflict with system preamble instructions.\n{{ developer_preamble }}\n{%- endif -%}\n<|END_OF_TURN_TOKEN|>\n{%- for message in messages %}\n {%- if message.role|lower == 'system' and not (loop.first and developer_preamble)%}\n<|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|>{{ message.content }}<|END_OF_TURN_TOKEN|>\n {%- elif message.role|lower == 'user' %}\n<|START_OF_TURN_TOKEN|><|USER_TOKEN|>{{ message.content }}<|END_OF_TURN_TOKEN|>{%- if documents and not sent_documents.value %}{%- set sent_documents.value = true %}{% set tool_idx.value = tool_idx.value + 1 %}{{ document_turn(documents) }}{% endif %}\n {%- elif message.role|lower == 'assistant' or message.role|lower == 'chatbot' %}\n<|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|>{% if message.tool_calls %}<|START_THINKING|>{{message.tool_plan}}<|END_THINKING|><|START_ACTION|>[\n {% for tc in message.tool_calls %}\n {\"tool_call_id\": \"{{ tool_idx.value }}\", \"tool_name\": \"{{ tc['function']['name'] }}\", \"parameters\": {{ tc['function']['arguments']|tojson }}}{% if not loop.last %},{% endif %}\n\n {% set tool_idx.value = tool_idx.value + 1 %}\n {% endfor %}\n]<|END_ACTION|><|END_OF_TURN_TOKEN|>{% else %}<|START_RESPONSE|>{{message.content}}<|END_RESPONSE|><|END_OF_TURN_TOKEN|>{% endif %}\n {% elif message.role|lower == 'tool' and message.tool_call_id not in tool_ids_seen.value %}\n<|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|><|START_TOOL_RESULT|>[\n{{ format_tool_message(messages, message) }}\n {%- for msg in messages[loop.index0 + 1:] %}\n {%- if msg.role|lower == 'tool' %},\n{{ format_tool_message(messages, msg) }}\n {%- set tool_ids_seen.value = tool_ids_seen.value + [msg.tool_call_id] %}\n {%- else %}\n {%- break %}\n {%- endif %}\n {%- endfor %}\n\n]<|END_TOOL_RESULT|><|END_OF_TURN_TOKEN|>\n {%- endif %}\n{%- endfor %}<|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|>"
28
+ }
29
+ ],
30
+ "added_tokens_decoder": {
31
+ "0": {
32
+ "content": "<PAD>",
33
+ "lstrip": false,
34
+ "normalized": false,
35
+ "rstrip": false,
36
+ "single_word": false,
37
+ "special": true
38
+ },
39
+ "1": {
40
+ "content": "<UNK>",
41
+ "lstrip": false,
42
+ "normalized": false,
43
+ "rstrip": false,
44
+ "single_word": false,
45
+ "special": true
46
+ },
47
+ "2": {
48
+ "content": "<CLS>",
49
+ "lstrip": false,
50
+ "normalized": false,
51
+ "rstrip": false,
52
+ "single_word": false,
53
+ "special": true
54
+ },
55
+ "3": {
56
+ "content": "<SEP>",
57
+ "lstrip": false,
58
+ "normalized": false,
59
+ "rstrip": false,
60
+ "single_word": false,
61
+ "special": true
62
+ },
63
+ "4": {
64
+ "content": "<MASK_TOKEN>",
65
+ "lstrip": false,
66
+ "normalized": false,
67
+ "rstrip": false,
68
+ "single_word": false,
69
+ "special": true
70
+ },
71
+ "5": {
72
+ "content": "<BOS_TOKEN>",
73
+ "lstrip": false,
74
+ "normalized": false,
75
+ "rstrip": false,
76
+ "single_word": false,
77
+ "special": true
78
+ },
79
+ "6": {
80
+ "content": "<EOS_TOKEN>",
81
+ "lstrip": false,
82
+ "normalized": false,
83
+ "rstrip": false,
84
+ "single_word": false,
85
+ "special": true
86
+ },
87
+ "7": {
88
+ "content": "<EOP_TOKEN>",
89
+ "lstrip": false,
90
+ "normalized": false,
91
+ "rstrip": false,
92
+ "single_word": false,
93
+ "special": true
94
+ },
95
+ "255000": {
96
+ "content": "<|START_OF_TURN_TOKEN|>",
97
+ "lstrip": false,
98
+ "normalized": false,
99
+ "rstrip": false,
100
+ "single_word": false,
101
+ "special": false
102
+ },
103
+ "255001": {
104
+ "content": "<|END_OF_TURN_TOKEN|>",
105
+ "lstrip": false,
106
+ "normalized": false,
107
+ "rstrip": false,
108
+ "single_word": false,
109
+ "special": false
110
+ },
111
+ "255002": {
112
+ "content": "<|YES_TOKEN|>",
113
+ "lstrip": false,
114
+ "normalized": false,
115
+ "rstrip": false,
116
+ "single_word": false,
117
+ "special": false
118
+ },
119
+ "255003": {
120
+ "content": "<|NO_TOKEN|>",
121
+ "lstrip": false,
122
+ "normalized": false,
123
+ "rstrip": false,
124
+ "single_word": false,
125
+ "special": false
126
+ },
127
+ "255004": {
128
+ "content": "<|GOOD_TOKEN|>",
129
+ "lstrip": false,
130
+ "normalized": false,
131
+ "rstrip": false,
132
+ "single_word": false,
133
+ "special": false
134
+ },
135
+ "255005": {
136
+ "content": "<|BAD_TOKEN|>",
137
+ "lstrip": false,
138
+ "normalized": false,
139
+ "rstrip": false,
140
+ "single_word": false,
141
+ "special": false
142
+ },
143
+ "255006": {
144
+ "content": "<|USER_TOKEN|>",
145
+ "lstrip": false,
146
+ "normalized": false,
147
+ "rstrip": false,
148
+ "single_word": false,
149
+ "special": false
150
+ },
151
+ "255007": {
152
+ "content": "<|CHATBOT_TOKEN|>",
153
+ "lstrip": false,
154
+ "normalized": false,
155
+ "rstrip": false,
156
+ "single_word": false,
157
+ "special": false
158
+ },
159
+ "255008": {
160
+ "content": "<|SYSTEM_TOKEN|>",
161
+ "lstrip": false,
162
+ "normalized": false,
163
+ "rstrip": false,
164
+ "single_word": false,
165
+ "special": false
166
+ },
167
+ "255009": {
168
+ "content": "<|USER_0_TOKEN|>",
169
+ "lstrip": false,
170
+ "normalized": false,
171
+ "rstrip": false,
172
+ "single_word": false,
173
+ "special": false
174
+ },
175
+ "255010": {
176
+ "content": "<|USER_1_TOKEN|>",
177
+ "lstrip": false,
178
+ "normalized": false,
179
+ "rstrip": false,
180
+ "single_word": false,
181
+ "special": false
182
+ },
183
+ "255011": {
184
+ "content": "<|USER_2_TOKEN|>",
185
+ "lstrip": false,
186
+ "normalized": false,
187
+ "rstrip": false,
188
+ "single_word": false,
189
+ "special": false
190
+ },
191
+ "255012": {
192
+ "content": "<|USER_3_TOKEN|>",
193
+ "lstrip": false,
194
+ "normalized": false,
195
+ "rstrip": false,
196
+ "single_word": false,
197
+ "special": false
198
+ },
199
+ "255013": {
200
+ "content": "<|USER_4_TOKEN|>",
201
+ "lstrip": false,
202
+ "normalized": false,
203
+ "rstrip": false,
204
+ "single_word": false,
205
+ "special": false
206
+ },
207
+ "255014": {
208
+ "content": "<|USER_5_TOKEN|>",
209
+ "lstrip": false,
210
+ "normalized": false,
211
+ "rstrip": false,
212
+ "single_word": false,
213
+ "special": false
214
+ },
215
+ "255015": {
216
+ "content": "<|USER_6_TOKEN|>",
217
+ "lstrip": false,
218
+ "normalized": false,
219
+ "rstrip": false,
220
+ "single_word": false,
221
+ "special": false
222
+ },
223
+ "255016": {
224
+ "content": "<|USER_7_TOKEN|>",
225
+ "lstrip": false,
226
+ "normalized": false,
227
+ "rstrip": false,
228
+ "single_word": false,
229
+ "special": false
230
+ },
231
+ "255017": {
232
+ "content": "<|USER_8_TOKEN|>",
233
+ "lstrip": false,
234
+ "normalized": false,
235
+ "rstrip": false,
236
+ "single_word": false,
237
+ "special": false
238
+ },
239
+ "255018": {
240
+ "content": "<|USER_9_TOKEN|>",
241
+ "lstrip": false,
242
+ "normalized": false,
243
+ "rstrip": false,
244
+ "single_word": false,
245
+ "special": false
246
+ },
247
+ "255019": {
248
+ "content": "<|START_THINKING|>",
249
+ "lstrip": false,
250
+ "normalized": false,
251
+ "rstrip": false,
252
+ "single_word": false,
253
+ "special": false
254
+ },
255
+ "255020": {
256
+ "content": "<|END_THINKING|>",
257
+ "lstrip": false,
258
+ "normalized": false,
259
+ "rstrip": false,
260
+ "single_word": false,
261
+ "special": false
262
+ },
263
+ "255021": {
264
+ "content": "<|START_RESPONSE|>",
265
+ "lstrip": false,
266
+ "normalized": false,
267
+ "rstrip": false,
268
+ "single_word": false,
269
+ "special": true
270
+ },
271
+ "255022": {
272
+ "content": "<|END_RESPONSE|>",
273
+ "lstrip": false,
274
+ "normalized": false,
275
+ "rstrip": false,
276
+ "single_word": false,
277
+ "special": true
278
+ },
279
+ "255023": {
280
+ "content": "<|START_ACTION|>",
281
+ "lstrip": false,
282
+ "normalized": false,
283
+ "rstrip": false,
284
+ "single_word": false,
285
+ "special": false
286
+ },
287
+ "255024": {
288
+ "content": "<|END_ACTION|>",
289
+ "lstrip": false,
290
+ "normalized": false,
291
+ "rstrip": false,
292
+ "single_word": false,
293
+ "special": false
294
+ },
295
+ "255025": {
296
+ "content": "<|START_TOOL_RESULT|>",
297
+ "lstrip": false,
298
+ "normalized": false,
299
+ "rstrip": false,
300
+ "single_word": false,
301
+ "special": false
302
+ },
303
+ "255026": {
304
+ "content": "<|END_TOOL_RESULT|>",
305
+ "lstrip": false,
306
+ "normalized": false,
307
+ "rstrip": false,
308
+ "single_word": false,
309
+ "special": false
310
+ },
311
+ "255027": {
312
+ "content": "<|EXTRA_8_TOKEN|>",
313
+ "lstrip": false,
314
+ "normalized": false,
315
+ "rstrip": false,
316
+ "single_word": false,
317
+ "special": false
318
+ },
319
+ "255028": {
320
+ "content": "<|NEW_FILE|>",
321
+ "lstrip": false,
322
+ "normalized": false,
323
+ "rstrip": false,
324
+ "single_word": false,
325
+ "special": true
326
+ },
327
+ "255029": {
328
+ "content": "<|BEGINNING_OF_PREFIX_FIM_TOKEN|>",
329
+ "lstrip": false,
330
+ "normalized": false,
331
+ "rstrip": false,
332
+ "single_word": false,
333
+ "special": false
334
+ },
335
+ "255030": {
336
+ "content": "<|BEGINNING_OF_MIDDLE_FIM_TOKEN|>",
337
+ "lstrip": false,
338
+ "normalized": false,
339
+ "rstrip": false,
340
+ "single_word": false,
341
+ "special": false
342
+ },
343
+ "255031": {
344
+ "content": "<|BEGINNING_OF_SUFFIX_FIM_TOKEN|>",
345
+ "lstrip": false,
346
+ "normalized": false,
347
+ "rstrip": false,
348
+ "single_word": false,
349
+ "special": false
350
+ },
351
+ "255032": {
352
+ "content": "<|END_OF_MIDDLE_FIM_TOKEN|>",
353
+ "lstrip": false,
354
+ "normalized": false,
355
+ "rstrip": false,
356
+ "single_word": false,
357
+ "special": false
358
+ }
359
+ },
360
+ "additional_special_tokens": [
361
+ "<|START_RESPONSE|>",
362
+ "<|END_RESPONSE|>"
363
+ ]
364
+ }