in2IN / README.md
pabloruizponce's picture
Update README.md
f8b2446 verified
|
raw
history blame
3.46 kB
---
library_name: transformers
pipeline_tag: text-to-3d
---
# in2IN: Leveraging individual Information to Generate Human INteractions
<p style="display:flex; gap:5px" align="center">
<a href="https://pabloruizponce.github.io/in2IN/"><img alt="Project" src="https://img.shields.io/badge/-Project%20Page-lightgrey?logo=Google%20Chrome&color=informational&logoColor=white"></a>
<a href="https://arxiv.org/abs/2404.09988"><img alt="arXiv" src="https://img.shields.io/badge/arXiv-2404.09988-b31b1b.svg"></a>
<a href="https://paperswithcode.com/sota/motion-synthesis-on-interhuman?p=in2in-leveraging-individual-information-to-1"><img src="https://img.shields.io/endpoint.svg?url=https://paperswithcode.com/badge/in2in-leveraging-individual-information-to-1/motion-synthesis-on-interhuman"/></a>
</p>
<div style="text-align: center;">
    <img src="cover.png" align="center" width=100% >
</div>
</br>
Generating human-human motion interactions conditioned on textual descriptions is a very useful application in many areas such as robotics, gaming, animation, and the metaverse. Alongside this utility also comes a great difficulty in modeling the highly dimensional inter-personal dynamics. In addition, properly capturing the intra-personal diversity of interactions has a lot of challenges. Current methods generate interactions with limited diversity of intra-person dynamics due to the limitations of the available datasets and conditioning strategies. For this, we introduce <b>in2IN</b>, a novel diffusion model for human-human motion generation which is conditioned not only on the textual description of the overall interaction but also on the individual descriptions of the actions performed by each person involved in the interaction. To train this model, we use a large language model to extend the InterHuman dataset with individual descriptions. As a result, <b>in2IN</b> achieves state-of-the-art performance in the InterHuman dataset. Furthermore, in order to increase the intra-personal diversity on the existing interaction datasets, we propose <b>DualMDM</b>, a model composition technique that combines the motions generated with <b>in2IN</b> and the motions generated by a single-person motion prior pre-trained on HumanML3D. As a result, <b>DualMDM</b> generates motions with higher individual diversity and improves control over the intra-person dynamics while maintaining inter-personal coherence.
## 🕹️ Usage
**Input**: The model gets as input the textual description of the overall interaction and the two individual descriptions from the interactants
**Output** (2,T,N,3): the model returns an array with the coordinates of the N joints of each interactant during a motion of T timesteps of duration,  
 
```python
from transformers import AutoModel        
model = AutoModel.from_pretrained("pabloruizponce/in2IN", trust_remote_code=True)
model(textI, texti1, texti2)
```
## 📚 Citation
If you find our work helpful, please cite:
```bibtex
@InProceedings{Ruiz-Ponce_2024_CVPR,
author = {Ruiz-Ponce, Pablo and Barquero, German and Palmero, Cristina and Escalera, Sergio and Garc{\'\i}a-Rodr{\'\i}guez, Jos\'e},
title = {in2IN: Leveraging Individual Information to Generate Human INteractions},
booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops},
month = {June},
year = {2024},
pages = {1941-1951}
}
```