Add files using upload-large-folder tool
Browse files- chat_template.jinja +7 -7
- config.json +3 -5
- processor_config.json +1 -0
- tokenizer.json +2 -2
chat_template.jinja
CHANGED
|
@@ -1,5 +1,5 @@
|
|
| 1 |
{#- Default system message if no system prompt is passed. #}
|
| 2 |
-
{%- set default_system_message = '
|
| 3 |
|
| 4 |
{#- Begin of sequence token. #}
|
| 5 |
{{- bos_token }}
|
|
@@ -79,8 +79,8 @@
|
|
| 79 |
|
| 80 |
{#- Assistant messages supports text content or text and image chunks. #}
|
| 81 |
{%- elif message['role'] == 'assistant' %}
|
| 82 |
-
{%- if message['content'] is
|
| 83 |
-
{{- raise_exception('Assistant message
|
| 84 |
{%- endif %}
|
| 85 |
|
| 86 |
{%- if message['content'] is string %}
|
|
@@ -93,7 +93,9 @@
|
|
| 93 |
{{- raise_exception('Only text chunks are supported in assistant message contents.') }}
|
| 94 |
{%- endif %}
|
| 95 |
{%- endfor %}
|
| 96 |
-
{%-
|
|
|
|
|
|
|
| 97 |
{%- for tool in message['tool_calls'] %}
|
| 98 |
{%- set arguments = tool['function']['arguments'] %}
|
| 99 |
{%- if arguments is not string %}
|
|
@@ -103,8 +105,6 @@
|
|
| 103 |
{%- endif %}
|
| 104 |
{{- '[TOOL_CALLS]' + tool['function']['name'] + '[ARGS]' + arguments }}
|
| 105 |
{%- endfor %}
|
| 106 |
-
{%- else %}
|
| 107 |
-
{{- raise_exception('Assistant message must have a string or a list of chunks in content or a list of tool calls.') }}
|
| 108 |
{%- endif %}
|
| 109 |
|
| 110 |
{#- End of sequence token for each assistant messages. #}
|
|
@@ -116,6 +116,6 @@
|
|
| 116 |
|
| 117 |
{#- Raise exception for unsupported roles. #}
|
| 118 |
{%- else %}
|
| 119 |
-
{{- raise_exception('Only user, assistant and tool roles are supported, got ' + message) }}
|
| 120 |
{%- endif %}
|
| 121 |
{%- endfor %}
|
|
|
|
| 1 |
{#- Default system message if no system prompt is passed. #}
|
| 2 |
+
{%- set default_system_message = '' %}
|
| 3 |
|
| 4 |
{#- Begin of sequence token. #}
|
| 5 |
{{- bos_token }}
|
|
|
|
| 79 |
|
| 80 |
{#- Assistant messages supports text content or text and image chunks. #}
|
| 81 |
{%- elif message['role'] == 'assistant' %}
|
| 82 |
+
{%- if (message['content'] is none or message['content'] == '' or message['content']|length == 0) and (message['tool_calls'] is not defined or message['tool_calls'] is none or message['tool_calls']|length == 0) %}
|
| 83 |
+
{{- raise_exception('Assistant message must have a string or a list of chunks in content or a list of tool calls.') }}
|
| 84 |
{%- endif %}
|
| 85 |
|
| 86 |
{%- if message['content'] is string %}
|
|
|
|
| 93 |
{{- raise_exception('Only text chunks are supported in assistant message contents.') }}
|
| 94 |
{%- endif %}
|
| 95 |
{%- endfor %}
|
| 96 |
+
{%- endif %}
|
| 97 |
+
|
| 98 |
+
{%- if message['tool_calls'] is defined and message['tool_calls'] is not none and message['tool_calls']|length > 0 %}
|
| 99 |
{%- for tool in message['tool_calls'] %}
|
| 100 |
{%- set arguments = tool['function']['arguments'] %}
|
| 101 |
{%- if arguments is not string %}
|
|
|
|
| 105 |
{%- endif %}
|
| 106 |
{{- '[TOOL_CALLS]' + tool['function']['name'] + '[ARGS]' + arguments }}
|
| 107 |
{%- endfor %}
|
|
|
|
|
|
|
| 108 |
{%- endif %}
|
| 109 |
|
| 110 |
{#- End of sequence token for each assistant messages. #}
|
|
|
|
| 116 |
|
| 117 |
{#- Raise exception for unsupported roles. #}
|
| 118 |
{%- else %}
|
| 119 |
+
{{- raise_exception('Only user, assistant and tool roles are supported, got ' + message['role']) }}
|
| 120 |
{%- endif %}
|
| 121 |
{%- endfor %}
|
config.json
CHANGED
|
@@ -16,8 +16,7 @@
|
|
| 16 |
"initializer_range": 0.02,
|
| 17 |
"intermediate_size": 9216,
|
| 18 |
"max_position_embeddings": 262144,
|
| 19 |
-
"model_type": "
|
| 20 |
-
"mscale": 1,
|
| 21 |
"num_attention_heads": 32,
|
| 22 |
"num_hidden_layers": 26,
|
| 23 |
"num_key_value_heads": 8,
|
|
@@ -27,7 +26,8 @@
|
|
| 27 |
"beta_slow": 1.0,
|
| 28 |
"factor": 16.0,
|
| 29 |
"llama_4_scaling_beta": 0.1,
|
| 30 |
-
"
|
|
|
|
| 31 |
"original_max_position_embeddings": 16384,
|
| 32 |
"rope_theta": 1000000.0,
|
| 33 |
"rope_type": "yarn",
|
|
@@ -44,8 +44,6 @@
|
|
| 44 |
"head_dim": 64,
|
| 45 |
"hidden_act": "silu",
|
| 46 |
"hidden_size": 1024,
|
| 47 |
-
"image_break_id": 12,
|
| 48 |
-
"image_end_id": 13,
|
| 49 |
"image_size": 1540,
|
| 50 |
"initializer_range": 0.02,
|
| 51 |
"intermediate_size": 4096,
|
|
|
|
| 16 |
"initializer_range": 0.02,
|
| 17 |
"intermediate_size": 9216,
|
| 18 |
"max_position_embeddings": 262144,
|
| 19 |
+
"model_type": "ministral3",
|
|
|
|
| 20 |
"num_attention_heads": 32,
|
| 21 |
"num_hidden_layers": 26,
|
| 22 |
"num_key_value_heads": 8,
|
|
|
|
| 26 |
"beta_slow": 1.0,
|
| 27 |
"factor": 16.0,
|
| 28 |
"llama_4_scaling_beta": 0.1,
|
| 29 |
+
"mscale": 1.0,
|
| 30 |
+
"mscale_all_dim": 1.0,
|
| 31 |
"original_max_position_embeddings": 16384,
|
| 32 |
"rope_theta": 1000000.0,
|
| 33 |
"rope_type": "yarn",
|
|
|
|
| 44 |
"head_dim": 64,
|
| 45 |
"hidden_act": "silu",
|
| 46 |
"hidden_size": 1024,
|
|
|
|
|
|
|
| 47 |
"image_size": 1540,
|
| 48 |
"initializer_range": 0.02,
|
| 49 |
"intermediate_size": 4096,
|
processor_config.json
CHANGED
|
@@ -18,6 +18,7 @@
|
|
| 18 |
0.40821073
|
| 19 |
],
|
| 20 |
"image_processor_type": "PixtralImageProcessorFast",
|
|
|
|
| 21 |
"image_std": [
|
| 22 |
0.26862954,
|
| 23 |
0.26130258,
|
|
|
|
| 18 |
0.40821073
|
| 19 |
],
|
| 20 |
"image_processor_type": "PixtralImageProcessorFast",
|
| 21 |
+
"image_seq_length": null,
|
| 22 |
"image_std": [
|
| 23 |
0.26862954,
|
| 24 |
0.26130258,
|
tokenizer.json
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:577575622324b2e099e2648be26bdeb5e5815ffe66d7004e9e3ddbf421db6bf1
|
| 3 |
+
size 17078110
|