📣 Important Announcement: FFUSION Ai Public Beta Release is Here!
🔭 We are thrilled to launch the public beta release of FFUSION Ai, though we want to clarify that it's currently limited in its breadth. Having been trained on just a fraction of our full image collection (20%), the capabilities of the model are not yet fully realized. This early version is primarily intended for experimentation with various prompt combinations and initial testing.
💡 While we're committed to delivering the highest level of excellence, we want to highlight that our model, notably the Unet component, is still developing its proficiency with certain objects and faces. But fear not, we're actively fine-tuning these areas as we progress towards the final release.
🙏 A huge shout out to our Reddit community for their support in alpha testing and for helping the text encoder respond to some exciting fuse ideas. We couldn't have come this far without you!
💡 Your contribution in this beta testing phase is extremely crucial to us. We invite you to explore the model extensively, experiment with it, and do not hesitate to report any prompts that don't meet your expectations. Your feedback is our guiding light in refining the performance and overall quality of FFUSION Ai.
⚠️ Attention: The model is based on Stable Diffusion 2.1 - 512 and is designed for optimal performance up to a resolution of approximately 600-700 pixels. For larger image sizes, we recommend upscaling them independently or patiently waiting for our final release that's just around the corner. This forthcoming release will enhance performance and support for higher resolutions.
👥 Thank you for being part of the FFUSION Ai beta testing community. Your support, feedback, and passion inspire us to continually develop a pioneering tool that is set to revolutionize creativity and visualization. Together, we can shape the future of storytelling and creativity.
🔮 Why not add some effects to your favorite prompts or fuse them together for a surreal twist? (Please note, Pen Pineapple Apple Pan effects and FUSIONS are excluded in this beta version)
🔒 With over 730.9449 hours of dedicated training sessions, our Fusion AI model offers a wealth of data subsets and robust datasets developed in collaboration with two enterprise corporate accounts for Mid Journey. We also pride ourselves in having an effective utilization of GPU usage, making the most out of our partnership with Idle Stoev, Source Code Bulgaria, Praesidium CX & BlackSwan Technologies. 🚀
Full transparency on our extensive 700,000-image dataset, training methodologies, classifications, and successful experiments is on its way. This information will be released shortly after the final version, further establishing FFUSION Ai as a trusted tool in the world of AI-powered creativity. Let's continue to imagine, create and explore together!
Introducing FFUSION AI - a groundbreaking tool for image generation and transformation, crafted around the cutting-edge Latent Diffusion Model. We build on the impressive capabilities of Stability AI's Stable Diffusion v2.1 512 & 768 models, harnessing the power of a fixed, pre-trained text encoder (OpenCLIP-ViT/H). Let's delve into a universe where creativity knows no bounds!
Developed by: Idle Stoev, Source Code Bulgaria, Praesidium CX & BlackSwan Technologies
Model type: Diffusion-based text-to-image generation model
License: CreativeML Open RAIL++-M License
FFUSION AI is a multi-faceted tool that shines in various applications. Primarily envisioned for research, FFUSION AI has potential to:
- Securely deploy models that could generate sensitive content, making AI safer.
- Examine the limitations and inherent biases in generative models.
- Unleash the artist within, aiding in creative processes or artistic endeavours.
- Reinvent educational or creative utilities with AI-driven innovations.
- Propel the research in the fascinating domain of generative models.
However, it's crucial to note that certain uses of FFUSION AI are strictly prohibited, as outlined below.
Our policy, adopted from the principles of the Stable Diffusion v2.1 model card, ensures the responsible use of Fusion AI beta and final releases. We expressly prohibit the utilization of our model for generating or distributing images that might incite hostility or exclusion. This includes:
- Content that is distressing, offensive, or perpetuates harmful stereotypes.
- Misuse or malicious use that harms individuals or communities, including creating demeaning or harmful representations, or promoting discriminatory content.
- Using the model for impersonation without consent or creating non-consensual explicit content.
- Generating or spreading mis- and disinformation, violent, gory imagery, or violating copyright terms.
While our model leaps toward the future of AI-driven creativity, it's essential to recognize its current limitations:
- The quest for perfect photorealism continues.
- Rendering legible text remains a challenge.
- Even more complex tasks, such as depicting "A red cube on top of a blue sphere in the middle of the ocean in a desert" may pose difficulty (but still processable).
- Human figures, particularly faces, may not be accurately generated.
The power of generative models also brings with it the potential for bias. As Stable Diffusion v2 trains primarily on subsets of LAION-2B(en) with English descriptions, the representations of non-English communities may be insufficient, often defaulting to white-biased and western cultures. As such, discretion is advised, recognizing that the model may unintentionally amplify biases, irrespective of input or intent. At FFUSION AI, we're committed to bringing your wildest imaginations to life, while maintaining a safe, inclusive, and responsible use of AI. Together, let's revolutionize the world of creativity! 🌟
We are excited to unveil the following versions:
BaSE and FUSION models will soon come with enhanced training capabilities including LoRa, LyCORIS, Dylora & Kohya-ss/sd-scripts. More information will be revealed upon release.
- di.FFUSION.ai-tXe-FXAA: Trained on "121361" images. Enhance your model's quality and sharpness using the pre-trained Unet.
- di.FFUSION.ai-tXe-fX: Trained on "211924" images. Amplify your model's surrealism and effects.
Our dedication to sustainable development is reflected in the model's carbon footprint. The CO2 emissions, calculated using the Machine Learning Impact calculator, stand at 124.95 kg for a total of 1190 hours of usage with an A100 PCIe 40GB GPU.
Hardware Type: A100 PCIe 40GB
Hours used: 1190
Cloud Provider: CoreWeave & Runpod (official partner)
Compute Region: US Cyxtera Chicago Data Center - ORD1 / EU - CZ & EU - RO
- Carbon Emitted (Power consumption x Time x Carbon produced based on the location of the power grid): 124.95 kg of CO2 emitted.
- Power consumption x Time x Carbon Produced Based on the Local Power Grid: 250W x 1190h = 297.5 kWh x 0.42 kg eq. CO2/kWh = 124.95 kg eq. CO2
- Local Hardware Storage 4x16TB Raid5 WD Gold Optimizer: AdamW & Dadaptation
This model card was written by: Idle Stoev and is based on the Stability AI - Stable Diffusion 2.1 model card.
- Downloads last month