Add pipeline tag, library name and link to project page

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +41 -6
README.md CHANGED
@@ -1,13 +1,48 @@
1
  ---
2
- license: apache-2.0
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  ---
4
 
5
- # RoboMaster
6
 
7
- It synthesizes realistic robotic manipulation video given an initial frame, a prompt, a user-defined object mask, and a collaborative trajectory describing the motion of both robotic arm and manipulated object in decomposed interaction phases. It supports diverse manipulation skills and can generalize to in-the-wild scenarios.
 
 
8
 
9
- ## Usage
10
 
11
- This is the implementation based on CogVideoX-5B. Please refer to our [github](https://github.com/KwaiVGI/RoboMaster) for details on usage.
12
 
13
- <video controls autoplay src="https://cdn-uploads.huggingface.co/production/uploads/63aef2cafcca84593e6682db/M7xBPv-NmqZeCvLRoDlu6.mp4"></video>
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ license: cc-by-4.0
3
+ task_categories:
4
+ - image-text-to-text
5
+ configs:
6
+ - config_name: default
7
+ data_files:
8
+ - split: HCMAS_train
9
+ path: version_v4/HCMAS-train.json
10
+ - split: HCMAS_test
11
+ path: version_v4/HCMAS-test.json
12
+ - split: HCSHR_train
13
+ path: version_v4/HCSHR-train.json
14
+ - split: HCSHR_test
15
+ path: version_v4/HCSHR-test.json
16
  ---
17
 
18
+ # Aligning VLM Assistants with Personalized Situated Cognition (ACL 2025 main)
19
 
20
+ [![GitHub Stars](https://img.shields.io/github/stars/your-username/PCogAlign?style=social)](https://github.com/liyongqi2002/PCogAlign)
21
+ [![Hugging Face Dataset](https://img.shields.io/badge/dataset-PCogAlignBench-blue)](https://huggingface.co/datasets/YongqiLi/PCogAlignBench)
22
+ [![arXiv](https://img.shields.io/badge/arXiv-2506.00930-orange)](https://arxiv.org/abs/2506.00930)
23
 
24
+ This repository contains the constructed benchmark in our ACL 2025 main paper **"Aligning VLM Assistants with Personalized Situated Cognition"**.
25
 
26
+ > ⚠️ This project is for academic research only and not intended for commercial use.
27
 
28
+ ## Abstract
29
+
30
+ Vision-language models (VLMs) aligned with general human objectives, such as being harmless and hallucination-free, have become valuable assistants of humans in managing visual tasks.
31
+ However, people with diversified backgrounds have different cognition even in the same situation. Consequently, they may have personalized expectations for VLM assistants.
32
+ This highlights the urgent need to align VLM assistants with personalized situated cognition for real-world assistance.
33
+ To study this problem, we first simplify it by characterizing individuals based on the sociological concept of Role-Set. Then, we propose to evaluate the individuals' actions to examine whether the personalized alignment is achieved.
34
+ Further, we construct a benchmark named PCogAlignBench, which includes 18k instances and 20 individuals with different Role-Sets.
35
+ Finally, we present a framework called PCogAlign, which constructs a cognition-aware and action-based reward model for personalized alignment.
36
+ Experimental results and human evaluations demonstrate the reliability of the PCogAlignBench and the effectiveness of our proposed PCogAlign.
37
+
38
+ ## 🙌 Acknowledgments
39
+
40
+ All datasets and models used are obtained through legal and ethical means. For detailed ethical considerations, please refer to our paper's Ethics Statement section.
41
+
42
+ ## 📬 Contact
43
+
44
+ For any questions or feedback, feel free to reach out to us at [liyongqi@whu.edu.cn].
45
+
46
+ ---
47
+
48
+ ✨ Thank you for your interest in PCogAlign! Stay tuned for more updates.