docker run -d -p 5000:5000 --gpus=all r8.im/xarty8932/dream@sha256:5e3c45aa9c9896f86634175309490225e5a379a6a81c39abbf55eab2cd16b657
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support
HF Inference deployability: The model has no library tag.