File size: 1,365 Bytes
cf0cb67 883561d 001dcac cba330f b4c04a4 15adf72 cf0cb67 cba330f 718d574 7c69ba1 718d574 cba330f 6789319 09ca66d 6789319 1db9085 6b2b215 6789319 5326ee0 15adf72 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 |
---
base_model:
- Qwen/Qwen3-4B-Instruct-2507
- Qwen/Qwen3-4B-Thinking-2507
library_name: transformers
datasets:
- openai/gsm8k
tags:
- evoluation
- math
- merge
---
# 📑 Model Card
[💻Github Repo](https://github.com/Hert4/Evolution-Merge) • [🤗Model Collections](https://huggingface.co/collections/beyoru/evolution-model)
## Model Details
This model is a merged version of two Qwen base models:
- **Qwen/Qwen3-4B-Instruct-2507**
- **Qwen/Qwen3-4B-Thinking-2507**
## Notations:
- **Evoluation dataset**: `openai/gsm8k` (subset of 100 samples, not trained)
- **Generation runs**: 50
- **Population size**: 10
- This model design for instruct model not reasoning model with same function like Qwen3-Instruct-2507
- **A good start for SFT or GRPO training.**
## Evaluation
- For my evaluation in my agent benchmark is not surpass too much but only 3% with instruct model.
- Surpass `openfree/Darwin-Qwen3-4B` (Evolution model) and base model in ACEBench.
```bibtex
@misc{nafy_qwen_merge_2025,
title = {Merged Qwen3 4B Instruct + Thinking Models},
author = {Beyoru},
year = {2025},
howpublished = {\url{https://huggingface.co/beyoru/EvolLLM}},
note = {Merged model combining instruction-tuned and reasoning Qwen3 variants.},
base_models = {Qwen/Qwen3-4B-Instruct-2507, Qwen/Qwen3-4B-Thinking-2507}
} |