KexianTang commited on
Commit
5b59d49
·
verified ·
1 Parent(s): 174be01

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -51,7 +51,7 @@ In this work, we introduce **LEGO-Puzzles**, a scalable and systematic benchmark
51
 
52
  To comprehensively assess spatial reasoning capabilities, LEGO-Puzzles is structured into three core task categories: **Spatial Understanding**, **Single-Step Sequential Reasoning**, and **Multi-Step Sequential Reasoning**. Each task requires models to understand visual inputs, perform step-by-step logical deduction, and maintain spatial consistency across sequences.
53
 
54
- In addition to traditional Visual Question Answering (VQA), LEGO-Puzzles also incorporates **image generation tasks**, assessing whether MLLMs can visually simulate structural transformations and anticipate future assembly states.
55
 
56
  We further introduce **LEGO-Puzzles-Lite**, a distilled subset tailored for human-model comparison, and a fine-grained evaluation suite named **Next-k-Step** to test reasoning scalability under increasing complexity.
57
 
@@ -99,7 +99,7 @@ We design **5 LEGO-based image generation tasks** testing a model's ability to s
99
  - 🎯 **Appearance Similarity**
100
  - 🎯 **Instruction Following**
101
 
102
- Only **Gemini-2.0-Flash** and **GPT-4o** show partial success. Open-source models typically fail to produce structurally valid or instruction-aligned images.
103
 
104
  <div align="center">
105
  <img src="https://tangkexian.github.io/LEGO-Puzzles/static/images/Generation_results.png" width="100%">
 
51
 
52
  To comprehensively assess spatial reasoning capabilities, LEGO-Puzzles is structured into three core task categories: **Spatial Understanding**, **Single-Step Sequential Reasoning**, and **Multi-Step Sequential Reasoning**. Each task requires models to understand visual inputs, perform step-by-step logical deduction, and maintain spatial consistency across sequences.
53
 
54
+ Furthermore, based on LEGO-Puzzzles, we design **image generation tasks** to investigate whether MLLMs can transfer their spatial understanding and reasoning abilities to image generation.
55
 
56
  We further introduce **LEGO-Puzzles-Lite**, a distilled subset tailored for human-model comparison, and a fine-grained evaluation suite named **Next-k-Step** to test reasoning scalability under increasing complexity.
57
 
 
99
  - 🎯 **Appearance Similarity**
100
  - 🎯 **Instruction Following**
101
 
102
+ Only **GPT-4o** and **Gemini-2.0-Flash** demonstrate partial success, while open-source models generally fail to produce structurally valid or instruction-aligned images. We evaluate GPT-4o, Gemini-2.0-Flash, GPT-4o* (referring to the version released prior to March 6, 2025), Emu2, GILL, and Anole using a scoring scale from 0 to 3 for both ***Appearance*** and ***Instruction Following*** dimensions.
103
 
104
  <div align="center">
105
  <img src="https://tangkexian.github.io/LEGO-Puzzles/static/images/Generation_results.png" width="100%">