--- base_model: [] library_name: transformers tags: - mergekit - merge --- # merge This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). Thank you to [MatchaSpeed](https://huggingface.co/matchaaaaa) for picking out the models to go into this merge. You're awesome! 🤗 The models in this merge were chosen for their attention to detail and/or their creativity. Based on each model's individual strength, weights were applied to prioritize the model's intelligence (self-attention weights) or style/creativity (multi-layer perceptron weights). GGUF versions available courtesy of BackyardAI https://huggingface.co/backyardai/BlackMagic-7B-GGUF # Usage Recommended prompt format is Alpaca, as this is most widely accepted by the component models. ## Merge Details ### Merge Method This model was merged using the [task arithmetic](https://arxiv.org/abs/2212.04089) merge method using Mistral-7B-v0.1 as a base. ### Models Merged The following models were included in the merge: * fearlessdots/WizardLM-2-7B-abliterated * Gryphe/MythoMist-7b * Sao10K/Frostwind-v2.1-m7 * senseable/Westlake-7b-v2 * maywell/PiVoT-0.1-Evil-a * Undi95/Toppy-M-7B ### Configuration The following YAML configuration was used to produce this model: ```yaml models: - model: fearlessdots/WizardLM-2-7B-abliterated parameters: weight: 1.0 - model: Undi95/Toppy-M-7B parameters: weight: - filter: self_attn value: 0.8 - value: 0.5 - model: senseable/Westlake-7b-v2 parameters: weight: - filter: self_attn value: 0.6 - value: 0.4 - model: maywell/PiVoT-0.1-Evil-a parameters: weight: - filter: mlp value: 0.2 - value: 0.0 - model: Sao10K/Frostwind-v2.1-m7 parameters: weight: - filter: self_attn value: 0.2 - filter: mlp value: 0.8 - value: 0.5 - model: Gryphe/MythoMist-7b parameters: weight: - filter: mlp value: 0.6 - value: 0.0 base_model: TeeZee/Mistral-7B-v0.1-fp32 merge_method: task_arithmetic dtype: float32 ```