56m commited on
Commit
d8ac915
·
verified ·
1 Parent(s): ee5eb1b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +33 -40
README.md CHANGED
@@ -4,17 +4,10 @@ task_categories:
4
  - text-generation
5
  language:
6
  - en
7
- tags:
8
- - Reasoning
9
  size_categories:
10
  - 1K<n<10K
11
  ---
12
 
13
- ## about Eval.py
14
- > **How to Run Eval:**
15
- > 1. Prepare a CSV with two columns: `target` (the truth) and `prediction` (the model's full output).
16
- > 2. Run: `python eval.py --csv results.csv`
17
- > 3. The script will automatically clean spaces, extract the final 9 digits, and calculate the Exact Match score.
18
 
19
  # 🧩 NumPuzzle-Easy: The 1,500-Step Spatial Reasoning Nightmare
20
 
@@ -68,39 +61,6 @@ We invite all researchers and AI enthusiasts to post their evaluation logs in th
68
 
69
  ### 🛠 Evaluation Protocol: The Strict Mandate
70
 
71
- This benchmark is not a test of linguistic creativity or "vibes." It is a cold, binary assessment of a model's ability to maintain a deterministic world model. To maintain the integrity of NumPuzzle-Easy, all evaluators must adhere to the following rigorous protocol.
72
-
73
- #### 1. The "Exact Match" (EM) Mandate
74
- We reject the use of fuzzy matching, BLEU, ROUGE, or any other semantic similarity metrics. In the realm of spatial reasoning, "almost correct" is fundamentally "wrong." If a model calculates 14 out of 15 moves correctly, its 15th move and the subsequent 4-chain reaction will be catastrophic. Therefore, the **only valid metric is Exact Match (EM).**
75
- - **Criteria:** The output must be exactly the 9-digit sequence representing Row 1.
76
- - **Fail Case:** `086738476` vs `086730476` (Result: **0**).
77
-
78
- #### 2. The Leading Zero Trap (The "No Mercy" Policy)
79
- One of the most common failures in LLMs is the loss of numeric context. In this puzzle, Row 1 can frequently begin with one or more zeros (e.g., `001234567`).
80
- - **The Rule:** The model must output exactly **9 digits**.
81
- - **The Rationale:** If a model outputs 8 digits (e.g., `12345678`), it indicates that the LLM has collapsed the grid representation into a standard integer, losing the spatial awareness of the 9-column architecture. This is an immediate failure. No post-processing or "fixing" of the output is allowed.
82
-
83
- #### 3. Zero-Shot vs. Reasoning Trace
84
- To observe the pure architectural limits, we recommend **Zero-Shot evaluation**. However, for "Reasoning" models (o1, DeepSeek-R1, Qwen 3.6 Plus), the internal "Thinking Process" or "Chain of Thought" (CoT) should be preserved in logs.
85
- - **What to look for:** Does the model's internal grid representation "drift" over time? Most models will lose the position of the active orb around Move 7 or 8. We encourage evaluators to share these logs in the Community tab to visualize the "moment of collapse."
86
-
87
- #### 4. Community-Driven Verification (Transparency Protocol)
88
- As the creator, I do not provide an "Official Leaderboard" based on my own tests. To prevent "Benchmark Laundering" (where providers cherry-pick successful runs), the truth must be crowdsourced.
89
- - **How to Submit:** Please open a thread in the **[Community Tab]** with your model's name.
90
- - **Requirements:**
91
- - Full transparency on the prompt used.
92
- - If testing a subset (e.g., 100/1500), specify the exact row indices.
93
- - Post a screenshot or a JSON dump of the failed responses.
94
- - **Verification:** Only results that can be replicated by other community members will be recognized as the definitive score for a given model version.
95
-
96
- #### 5. Cost-Effective Testing
97
- We recognize that 1,500 samples of o1-level reasoning are expensive.
98
- - **Sampling:** For quick validation, a random sample of **100 rows** is acceptable. Given the 1% failure rate, a 100-sample test is statistically sufficient to prove a model's inability to solve this benchmark.
99
- - **Stop-Loss:** If your model fails the first 50 consecutive puzzles, it is safe to assume an accuracy of <2%. You may stop there and report the result as "Complete Failure."
100
-
101
- **"The grid does not care about your reasoning. It only cares about the truth."**
102
-
103
- ---
104
 
105
  # 📂 File Structure & Usage
106
 
@@ -129,6 +89,39 @@ If you are running your own Python script, you can use either file.
129
  1. **Statistical Power:** With 1,500 problems, any "lucky guess" from an LLM is statistically irrelevant.
130
  2. **Cost-Efficiency:** Small enough to be processed via OpenAI/Anthropic APIs within a few hours (overnight), keeping benchmarking costs reasonable while ensuring total failure for the model.
131
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
132
 
133
  ---
134
 
 
4
  - text-generation
5
  language:
6
  - en
 
 
7
  size_categories:
8
  - 1K<n<10K
9
  ---
10
 
 
 
 
 
 
11
 
12
  # 🧩 NumPuzzle-Easy: The 1,500-Step Spatial Reasoning Nightmare
13
 
 
61
 
62
  ### 🛠 Evaluation Protocol: The Strict Mandate
63
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
64
 
65
  # 📂 File Structure & Usage
66
 
 
89
  1. **Statistical Power:** With 1,500 problems, any "lucky guess" from an LLM is statistically irrelevant.
90
  2. **Cost-Efficiency:** Small enough to be processed via OpenAI/Anthropic APIs within a few hours (overnight), keeping benchmarking costs reasonable while ensuring total failure for the model.
91
 
92
+ ## 📊 Call for Evaluation: Test your model on the NumPuzzle-Easy Benchmark!
93
+
94
+ We are crowdsourcing evaluation results for **NumPuzzle-Easy**!
95
+ This benchmark evaluates how accurately an LLM can identify and extract target numbers of varying lengths from complex contexts.
96
+
97
+ Instead of waiting in long queues on centralized leaderboards, you can evaluate your model locally using Hugging Face's official `Lighteval` tool. By running the command below, the results will be automatically verified and submitted to your model's repository as a Pull Request!
98
+
99
+ ### 🚀 How to evaluate your model
100
+
101
+ **1. Install Lighteval**
102
+ ```bash
103
+ pip install lighteval
104
+ ```
105
+
106
+ **2. Download the task script**
107
+ Make sure you have the `numpuzzle_task.py` script in your working directory.
108
+
109
+ **3. Run the evaluation**
110
+ Run the following command. Replace `[YOUR_MODEL_ID]` with the Hugging Face ID of the model you want to test (e.g., `meta-llama/Meta-Llama-3-8B-Instruct`).
111
+
112
+ ```bash
113
+ lighteval accelerate \
114
+ --model_args "pretrained=[YOUR_MODEL_ID]" \
115
+ --custom_tasks "Evai.py" \
116
+ --tasks "custom|numpuzzle-easy|0|0" \
117
+ --push_to_hub \
118
+ --results_org "[YOUR_HF_USERNAME]"
119
+ ```
120
+
121
+ ### 💡 What happens next?
122
+ By using the `--push_to_hub` flag, Lighteval will automatically generate a JSON/YAML report with your score and push it to the Hub. You can then show off your verified NumPuzzle score on your model card!
123
+
124
+ We look forward to seeing your results! Let's find out which model is the true master of numbers. 🧩🔢
125
 
126
  ---
127