--- base_model: - unsloth/Llama-3.2-3B-Instruct - marcuscedricridia/talonp4 - unsloth/Llama-3.2-3B-Instruct - marcuscedricridia/talonp10-safe - unsloth/Llama-3.2-3B-Instruct - marcuscedricridia/talonp8-code.1nemotron - unsloth/Llama-3.2-3B-Instruct - marcuscedricridia/talonp1lora - unsloth/Llama-3.2-3B-Instruct - marcuscedricridia/talonp6-mathv1nemotron - unsloth/Llama-3.2-3B-Instruct - marcuscedricridia/talonp12-helpsteer - unsloth/Llama-3.2-3B-Instruct - marcuscedricridia/talonp5 - unsloth/Llama-3.2-3B-Instruct - marcuscedricridia/talonp13-debug - unsloth/Llama-3.2-3B-Instruct - marcuscedricridia/talonp2lora - unsloth/Llama-3.2-3B-Instruct - marcuscedricridia/talonp7-mathv1.1nemotron - unsloth/Llama-3.2-3B-Instruct - marcuscedricridia/talon-merged-run1 - unsloth/Llama-3.2-3B-Instruct - marcuscedricridia/talonp11-fleschprose - unsloth/Llama-3.2-3B-Instruct - marcuscedricridia/talonp9-scinemotron - unsloth/Llama-3.2-3B-Instruct library_name: transformers tags: - mergekit - merge --- # merge This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the [TIES](https://arxiv.org/abs/2306.01708) merge method using [unsloth/Llama-3.2-3B-Instruct](https://huggingface.co/unsloth/Llama-3.2-3B-Instruct) as a base. ### Models Merged The following models were included in the merge: * [unsloth/Llama-3.2-3B-Instruct](https://huggingface.co/unsloth/Llama-3.2-3B-Instruct) + [marcuscedricridia/talonp4](https://huggingface.co/marcuscedricridia/talonp4) * [unsloth/Llama-3.2-3B-Instruct](https://huggingface.co/unsloth/Llama-3.2-3B-Instruct) + [marcuscedricridia/talonp10-safe](https://huggingface.co/marcuscedricridia/talonp10-safe) * [unsloth/Llama-3.2-3B-Instruct](https://huggingface.co/unsloth/Llama-3.2-3B-Instruct) + [marcuscedricridia/talonp8-code.1nemotron](https://huggingface.co/marcuscedricridia/talonp8-code.1nemotron) * [unsloth/Llama-3.2-3B-Instruct](https://huggingface.co/unsloth/Llama-3.2-3B-Instruct) + [marcuscedricridia/talonp1lora](https://huggingface.co/marcuscedricridia/talonp1lora) * [unsloth/Llama-3.2-3B-Instruct](https://huggingface.co/unsloth/Llama-3.2-3B-Instruct) + [marcuscedricridia/talonp6-mathv1nemotron](https://huggingface.co/marcuscedricridia/talonp6-mathv1nemotron) * [unsloth/Llama-3.2-3B-Instruct](https://huggingface.co/unsloth/Llama-3.2-3B-Instruct) + [marcuscedricridia/talonp12-helpsteer](https://huggingface.co/marcuscedricridia/talonp12-helpsteer) * [unsloth/Llama-3.2-3B-Instruct](https://huggingface.co/unsloth/Llama-3.2-3B-Instruct) + [marcuscedricridia/talonp5](https://huggingface.co/marcuscedricridia/talonp5) * [unsloth/Llama-3.2-3B-Instruct](https://huggingface.co/unsloth/Llama-3.2-3B-Instruct) + [marcuscedricridia/talonp13-debug](https://huggingface.co/marcuscedricridia/talonp13-debug) * [unsloth/Llama-3.2-3B-Instruct](https://huggingface.co/unsloth/Llama-3.2-3B-Instruct) + [marcuscedricridia/talonp2lora](https://huggingface.co/marcuscedricridia/talonp2lora) * [unsloth/Llama-3.2-3B-Instruct](https://huggingface.co/unsloth/Llama-3.2-3B-Instruct) + [marcuscedricridia/talonp7-mathv1.1nemotron](https://huggingface.co/marcuscedricridia/talonp7-mathv1.1nemotron) * [unsloth/Llama-3.2-3B-Instruct](https://huggingface.co/unsloth/Llama-3.2-3B-Instruct) + [marcuscedricridia/talon-merged-run1](https://huggingface.co/marcuscedricridia/talon-merged-run1) * [unsloth/Llama-3.2-3B-Instruct](https://huggingface.co/unsloth/Llama-3.2-3B-Instruct) + [marcuscedricridia/talonp11-fleschprose](https://huggingface.co/marcuscedricridia/talonp11-fleschprose) * [unsloth/Llama-3.2-3B-Instruct](https://huggingface.co/unsloth/Llama-3.2-3B-Instruct) + [marcuscedricridia/talonp9-scinemotron](https://huggingface.co/marcuscedricridia/talonp9-scinemotron) ### Configuration The following YAML configuration was used to produce this model: ```yaml merge_method: ties base_model: unsloth/Llama-3.2-3B-Instruct parameters: normalize: true int8_mask: true dtype: bfloat16 models: - model: unsloth/Llama-3.2-3B-Instruct # No parameters needed for the base model itself in this list format - model: unsloth/Llama-3.2-3B-Instruct+marcuscedricridia/talon-merged-run1 parameters: density: 0.5 weight: 0.0461538462 - model: unsloth/Llama-3.2-3B-Instruct+marcuscedricridia/talonp13-debug parameters: density: 0.5 weight: 0.1 - model: unsloth/Llama-3.2-3B-Instruct+marcuscedricridia/talonp12-helpsteer parameters: density: 0.5 weight: 0.0461538462 - model: unsloth/Llama-3.2-3B-Instruct+marcuscedricridia/talonp11-fleschprose parameters: density: 0.5 weight: 0.0461538462 - model: unsloth/Llama-3.2-3B-Instruct+marcuscedricridia/talonp10-safe parameters: density: 0.5 weight: 0.1 - model: unsloth/Llama-3.2-3B-Instruct+marcuscedricridia/talonp9-scinemotron parameters: density: 0.5 weight: 0.1 - model: unsloth/Llama-3.2-3B-Instruct+marcuscedricridia/talonp8-code.1nemotron parameters: density: 0.5 weight: 0.1 - model: unsloth/Llama-3.2-3B-Instruct+marcuscedricridia/talonp7-mathv1.1nemotron parameters: density: 0.5 weight: 0.1 - model: unsloth/Llama-3.2-3B-Instruct+marcuscedricridia/talonp6-mathv1nemotron parameters: density: 0.5 weight: 0.0461538462 - model: unsloth/Llama-3.2-3B-Instruct+marcuscedricridia/talonp5 parameters: density: 0.5 weight: 0.0461538462 - model: unsloth/Llama-3.2-3B-Instruct+marcuscedricridia/talonp4 parameters: density: 0.5 weight: 0.0461538462 - model: unsloth/Llama-3.2-3B-Instruct+marcuscedricridia/talonp1lora parameters: density: 0.5 weight: 0.1 - model: unsloth/Llama-3.2-3B-Instruct+marcuscedricridia/talonp2lora parameters: density: 0.5 weight: 0.1 ```