metadata
language:
- ko
- en
license: mit
Model Card for free-evo-qwen72b-v0.8
1st place : 2024 4th May - avg. 81.28 Open Llm Leaderboard
but this kicked away. maybe the explanation was not enough.
Method
- We were inspired by this Sakana project
Process
- two models with the same architecture are needed so fine-tune a model to create a gap between the two of them.
- merge original one and fine-tuned one
- evaluate the merged model
- merge again it with original model
- evaluate again
- keep going until evaluate avg is higher then original one
that's it. simple.
Base Architecture
- QWEN2
Base Models
- several QWEN2 based models