--- license: cc-by-sa-4.0 language: - en pipeline_tag: text-generation tags: - code --- # MagicPrompt_SD_V1 This is a Prompt Generator likes [Gustavosta/MagicPrompt-Stable-Diffusion](https://huggingface.co/Gustavosta/MagicPrompt-Stable-Diffusion)! But I wash the origin prompts data, and trains a powerful model to generate prompt for [魔导绪论](https://magic-tag.netlify.app/) It's using Paddle to handle the training and other things. Not PyTorch or Tensorsflow. There's the result I get form this model: - You can use CPU to run the model! But GPU 10x faster then CPU 🚀. - CPU (about 300ms/per ) | GPU (about 90ms/per 🚀 ) V2-10 Model - You can add some change easier passing some params. ## 📕 Using example is here [飞桨仓库](https://aistudio.baidu.com/aistudio/projectdetail/5116158?contributionType=1) You can wrapper a FastAPI or Flask to easily deploy it to your server