File size: 1,365 Bytes
cf0cb67
883561d
 
 
001dcac
 
 
cba330f
 
b4c04a4
15adf72
cf0cb67
 
cba330f
 
718d574
7c69ba1
718d574
cba330f
 
 
 
 
 
 
 
 
6789319
09ca66d
 
6789319
 
 
1db9085
6b2b215
6789319
 
5326ee0
 
 
 
 
 
 
 
15adf72
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
---
base_model:
- Qwen/Qwen3-4B-Instruct-2507
- Qwen/Qwen3-4B-Thinking-2507
library_name: transformers
datasets:
- openai/gsm8k
tags:
- evoluation
- math
- merge
---

# 📑 Model Card


[💻Github Repo](https://github.com/Hert4/Evolution-Merge) • [🤗Model Collections](https://huggingface.co/collections/beyoru/evolution-model)

## Model Details
This model is a merged version of two Qwen base models:  
- **Qwen/Qwen3-4B-Instruct-2507**  
- **Qwen/Qwen3-4B-Thinking-2507**  


## Notations:
- **Evoluation dataset**: `openai/gsm8k` (subset of 100 samples, not trained)  
- **Generation runs**: 50  
- **Population size**: 10
- This model design for instruct model not reasoning model with same function like Qwen3-Instruct-2507
- **A good start for SFT or GRPO training.**

## Evaluation

- For my evaluation in my agent benchmark is not surpass too much but only 3% with instruct model.
- Surpass `openfree/Darwin-Qwen3-4B` (Evolution model) and base model in ACEBench.


```bibtex
@misc{nafy_qwen_merge_2025,
  title        = {Merged Qwen3 4B Instruct + Thinking Models},
  author       = {Beyoru},
  year         = {2025},
  howpublished = {\url{https://huggingface.co/beyoru/EvolLLM}},
  note         = {Merged model combining instruction-tuned and reasoning Qwen3 variants.},
  base_models  = {Qwen/Qwen3-4B-Instruct-2507, Qwen/Qwen3-4B-Thinking-2507}
}