--- base_model: - mistralai/Mistral-7B-v0.3 library_name: transformers tags: - mergekit - merge license: apache-2.0 language: - en --- # Mistral-7B-v0.3-prune6 This is a layer-pruned pre-trained language model sliced with [mergekit](https://github.com/cg123/mergekit). No additional training. ## layer selection ![plot](mistral-7b-v0_3-prune6.png) ## Quick eval Quick eval for: pszemraj/Mistral-7B-v0.3-prune6 hf (pretrained=pszemraj/Mistral-7B-v0.3-prune6,trust_remote_code=True,dtype=bfloat16), gen_kwargs: (None), limit: None, num_fewshot: None, batch_size: 2 | Tasks |Version|Filter|n-shot| Metric | Value | |Stderr| |--------------|------:|------|-----:|----------|------:|---|-----:| |arc_easy | 1|none | 0|acc | 0.6393|± |0.0099| | | |none | 0|acc_norm | 0.6309|± |0.0099| |boolq | 2|none | 0|acc | 0.7599|± |0.0075| |lambada_openai| 1|none | 0|perplexity|10.1184|± |0.2771| | | |none | 0|acc | 0.5507|± |0.0069| |openbookqa | 1|none | 0|acc | 0.2200|± |0.0185| | | |none | 0|acc_norm | 0.3580|± |0.0215| |piqa | 1|none | 0|acc | 0.7203|± |0.0105| | | |none | 0|acc_norm | 0.7350|± |0.0103| |winogrande | 1|none | 0|acc | 0.6906|± |0.0130| ### original bootstrapping for stddev: perplexity hf (pretrained=mistralai/Mistral-7B-v0.3,trust_remote_code=True,dtype=bfloat16), gen_kwargs: (None), limit: None, num_fewshot: None, batch_size: 2 | Tasks |Version|Filter|n-shot| Metric |Value | |Stderr| |--------------|------:|------|-----:|----------|-----:|---|-----:| |arc_easy | 1|none | 0|acc |0.7959|± |0.0083| | | |none | 0|acc_norm |0.7832|± |0.0085| |boolq | 2|none | 0|acc |0.8202|± |0.0067| |lambada_openai| 1|none | 0|perplexity|3.2578|± |0.0601| | | |none | 0|acc |0.7518|± |0.0060| |openbookqa | 1|none | 0|acc |0.3340|± |0.0211| | | |none | 0|acc_norm |0.4420|± |0.0222| |piqa | 1|none | 0|acc |0.8009|± |0.0093| | | |none | 0|acc_norm |0.8215|± |0.0089| |winogrande | 1|none | 0|acc |0.7380|± |0.0124| ## Merge Details ### Merge Method This model was merged using the passthrough merge method. ### Models Merged The following models were included in the merge: * [mistralai/Mistral-7B-v0.3](https://huggingface.co/mistralai/Mistral-7B-v0.3) ### Configuration The following YAML configuration was used to produce this model: ```yaml dtype: bfloat16 merge_method: passthrough slices: - sources: - layer_range: [0, 22] model: mistralai/Mistral-7B-v0.3 - sources: - layer_range: [28, 32] model: mistralai/Mistral-7B-v0.3 ```