leloy commited on
Commit
493b0db
1 Parent(s): f4498f2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -7,7 +7,7 @@ pipeline_tag: image-text-to-text
7
  ---
8
  ## Anole: An Open, Autoregressive and Native Multimodal Models for Interleaved Image-Text Generation
9
 
10
- Note: This is a huggingface-compatible version of the model converted by [leloy](https://huggingface.co/leloy). Currently only works [this branch of the Transformers library](https://github.com/leloykun/transformers/tree/fc--anole).
11
 
12
  Anole is the first open-source, autoregressive, and natively trained large multimodal model capable of interleaved image-text generation (without using stable diffusion). While it builds upon the strengths of Chameleon, Anole excels at the complex task of generating coherent sequences of alternating text and images. Through an innovative fine-tuning process using a carefully curated dataset of approximately 6,000 images, Anole achieves remarkable image generation and understanding capabilities with minimal additional training. This efficient approach, combined with its open-source nature, positions Anole as a catalyst for accelerated research and development in multimodal AI. Preliminary tests demonstrate Anole's exceptional ability to follow nuanced instructions, producing high-quality images and interleaved text-image content that closely aligns with user prompts.
13
 
 
7
  ---
8
  ## Anole: An Open, Autoregressive and Native Multimodal Models for Interleaved Image-Text Generation
9
 
10
+ Note: This is a huggingface-compatible version of the model converted by [leloy](https://huggingface.co/leloy). Currently only works on [this branch of the Transformers library](https://github.com/leloykun/transformers/tree/fc--anole).
11
 
12
  Anole is the first open-source, autoregressive, and natively trained large multimodal model capable of interleaved image-text generation (without using stable diffusion). While it builds upon the strengths of Chameleon, Anole excels at the complex task of generating coherent sequences of alternating text and images. Through an innovative fine-tuning process using a carefully curated dataset of approximately 6,000 images, Anole achieves remarkable image generation and understanding capabilities with minimal additional training. This efficient approach, combined with its open-source nature, positions Anole as a catalyst for accelerated research and development in multimodal AI. Preliminary tests demonstrate Anole's exceptional ability to follow nuanced instructions, producing high-quality images and interleaved text-image content that closely aligns with user prompts.
13