|
--- |
|
license: apache-2.0 |
|
task_categories: |
|
- text-generation |
|
--- |
|
Official implementation of the paper ["Order Matters: Investigate the Position Bias in Multi-constraint Instruction Following"](https://huggingface.co/papers/2502.17204). |
|
|
|
Code: https://github.com/meowpass/PBIF |
|
|
|
We systematically study **the position bias problem in multi-constraint |
|
instruction following**. Through our experiments, we have the following findings: |
|
|
|
- **LLMs prefer to "hard-to-easy" constraint order** |
|
- existing LLMs can achieve a better following accuracy in multi-constraint instructions when presented with constraints in “hard-to-easy” orders. |
|
- This finding can be generalized in both single-round and multi-round scenarios, regardless of the architecture of LLM, the size of LLM’s parameters and the number of constraints. |
|
- **Constraints order affect how the LLMs handle a specific constraint** |
|
- The "Hard-to-easy" constraint order induces the LLM to pay more attention to the constraint part in the multi-constraint instructions. |
|
- The LLM’s performance on various constraints is strongly correlated with its attention patterns. |
|
|
|
# PBIF Dataset |
|
The dataset consists of single_round inference data and multi_round inference data. |
|
For each of the data, there are 5 fields: |
|
- `prompt`: Synthesized multi-constraint instructions. |
|
- `constraint`: The constraints contained in the instructions. |
|
- `instruction_id_list`: The id of the constraints in the instructions. |
|
- `kwargs`: Corresponding parameters for the constraints, which are only used for evaluation. |
|
- `ranking`: The constraint order of the instruction. (0 indicates the hardest constraint) |
|
It is worth noting that, in multi_round inference data, the `prompt` is the initial instruction, which is more convenient for the user to construct the multi-round dialog data for themselves. |