|
{ |
|
"architectures": [ |
|
"Qwen3ForCausalLM" |
|
], |
|
"auto_map": { |
|
"AutoModelForCausalLM": "Qwen3ForCausalLM" |
|
}, |
|
"model_type": "qwen3", |
|
"quantization_config": { |
|
"quant_method": "gguf", |
|
"bits": 8, |
|
"group_size": 32, |
|
"zero_point": true, |
|
"desc_act": false |
|
}, |
|
"gguf_metadata": { |
|
"general": { |
|
"name": "Qwen3 4B Function Calling Pro", |
|
"architecture": "qwen3", |
|
"file_type": 7, |
|
"quantization_version": 2, |
|
"size_label": "4B" |
|
}, |
|
"qwen3": { |
|
"context_length": 262144, |
|
"embedding_length": 2560, |
|
"block_count": 36, |
|
"feed_forward_length": 9728, |
|
"attention": { |
|
"head_count": 32, |
|
"head_count_kv": 8, |
|
"key_length": 128, |
|
"value_length": 128, |
|
"layer_norm_rms_epsilon": 1e-06 |
|
}, |
|
"rope": { |
|
"freq_base": 5000000.0 |
|
} |
|
} |
|
}, |
|
"vocab_size": 151936, |
|
"hidden_size": 2560, |
|
"intermediate_size": 9728, |
|
"num_hidden_layers": 36, |
|
"num_attention_heads": 32, |
|
"num_key_value_heads": 8, |
|
"hidden_act": "silu", |
|
"max_position_embeddings": 262144, |
|
"initializer_range": 0.02, |
|
"rms_norm_eps": 1e-06, |
|
"use_cache": true, |
|
"rope_theta": 5000000.0, |
|
"attention_bias": false, |
|
"attention_dropout": 0.0, |
|
"torch_dtype": "float16", |
|
"transformers_version": "4.45.0", |
|
"special_tokens": { |
|
"bos_token": "<|endoftext|>", |
|
"eos_token": "<|im_end|>", |
|
"pad_token": "<|im_end|>", |
|
"additional_special_tokens": [ |
|
"<|im_start|>", |
|
"<|im_end|>", |
|
"<|object_ref_start|>", |
|
"<|object_ref_end|>", |
|
"<|box_start|>", |
|
"<|box_end|>", |
|
"<|quad_start|>", |
|
"<|quad_end|>", |
|
"<|vision_start|>", |
|
"<|vision_end|>", |
|
"<|vision_pad|>", |
|
"<|image_pad|>", |
|
"<|video_pad|>", |
|
"<tool_call>", |
|
"</tool_call>", |
|
"<tool_response>", |
|
"</tool_response>" |
|
] |
|
}, |
|
"chat_template": "{% for message in messages %}{% if message['role'] == 'user' %}{{ '<start_of_turn>user\n' + message['content'] + '<end_of_turn>\n' }}{% elif message['role'] == 'system' %}{{ '<start_of_turn>system\n' + message['content'] + '<end_of_turn>\n' }}{% elif message['role'] == 'assistant' %}{{ '<start_of_turn>model\n' + message['content'] + '<end_of_turn>\n' }}{% endif %}{% if loop.last and add_generation_prompt %}{{ '<start_of_turn>model\n' }}{% endif %}{% endfor %}", |
|
"generation_config": { |
|
"do_sample": true, |
|
"temperature": 0.7, |
|
"top_p": 0.8, |
|
"top_k": 40, |
|
"repetition_penalty": 1.1, |
|
"max_new_tokens": 512, |
|
"stop": [ |
|
"<|im_end|>", |
|
"<|im_start|>", |
|
"<tool_call>", |
|
"</tool_call>", |
|
"<tool_response>", |
|
"</tool_response>" |
|
] |
|
} |
|
} |