davidlms commited on
Commit
d20306b
·
verified ·
1 Parent(s): c97ace0

Add model-index with benchmark evaluations

Browse files

Added structured evaluation results from README benchmark table:

**Automated Benchmarks:**
- MMLU: 55.23
- GPQA: 31.47
- IFEval (Instruction following): 74.89
- IFBench: 20.7
- GSM8K (Math reasoning): 58.3
- MGSM (Multilingual math): 55.04
- MMMLU (Multilingual MMLU): 46.73

Total: 7 benchmarks across reasoning, instruction-following, and multilingual capabilities.

This enables the model to appear in leaderboards and makes it easier to compare with other models.

Note: PR #6 (Support tool calls) modifies the tokenizer configuration and should not conflict with this metadata addition.

Files changed (1) hide show
  1. README.md +33 -0
README.md CHANGED
@@ -17,6 +17,39 @@ tags:
17
  - liquid
18
  - lfm2
19
  - edge
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
20
  ---
21
 
22
  <center>
 
17
  - liquid
18
  - lfm2
19
  - edge
20
+ model-index:
21
+ - name: LFM2-1.2B
22
+ results:
23
+ - task:
24
+ type: text-generation
25
+ dataset:
26
+ name: Automated Benchmarks
27
+ type: benchmark
28
+ metrics:
29
+ - name: MMLU
30
+ type: mmlu
31
+ value: 55.23
32
+ - name: GPQA
33
+ type: gpqa
34
+ value: 31.47
35
+ - name: IFEval
36
+ type: ifeval
37
+ value: 74.89
38
+ - name: IFBench
39
+ type: ifbench
40
+ value: 20.7
41
+ - name: GSM8K
42
+ type: gsm8k
43
+ value: 58.3
44
+ - name: MGSM
45
+ type: mgsm
46
+ value: 55.04
47
+ - name: MMMLU
48
+ type: mmmlu
49
+ value: 46.73
50
+ source:
51
+ name: Model README - Automated Benchmarks
52
+ url: https://huggingface.co/LiquidAI/LFM2-1.2B
53
  ---
54
 
55
  <center>