--- license: apache-2.0 datasets: - multi-train/coco_captions_1107 - visual_genome language: - en pipeline_tag: text-to-image tags: - scene_graph - transformers - laplacian - autoregressive - vqvae --- # trf-sg2im Model card for the paper __"[Transformer-Based Image Generation from Scene Graphs](https://arxiv.org/abs/2303.04634)"__. Original GitHub implementation [here](https://github.com/perceivelab/trf-sg2im). ![teaser](docs/teaser.gif) ## Model This model is a two-stage scene-graph-to-image approach. It takes a scene graph as input and generates a layout using a transformer-based architecture with Laplacian Positional Encoding. Then, it uses this estimated layout to condition an autoregressive GPT-like transformer to compose the image in the latent, discrete space, converted into the final image by a VQVAE. ![architecture](docs/architecture.png) ## Usage For usage instructions, please refer to the original [GitHub repo](https://github.com/perceivelab/trf-sg2im). ## Results Comparison with other state-of-the-art approaches ![results](docs/results.png)