File size: 1,328 Bytes
cb2ac60
 
 
 
ffed138
 
cb2ac60
 
 
0e8338d
ffed138
0e8338d
482963e
 
 
0e8338d
 
482963e
 
0e8338d
 
cb2ac60
ffed138
 
 
 
482963e
 
 
 
 
ffed138
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
#!/usr/bin/env python
# coding: utf-8

import random
from dalle_mini.backend import ServiceError, get_images_from_backend
from dalle_mini.helpers import captioned_strip

import streamlit as st


st.sidebar.markdown('Visit [our report](https://wandb.ai/dalle-mini/dalle-mini/reports/DALL-E-mini--Vmlldzo4NjIxODA)')

st.header('DALL-E mini')
st.subheader('Generate images from a text prompt')

prompt = st.text_input("What do you want to see?")

#TODO: I think there's an issue where we can't run twice the same inference (not due to caching) - may need to use st.form

if prompt != "":
    st.write(f"Generating candidates for: {prompt}")

    try:
        backend_url = st.secrets["BACKEND_SERVER"]
        print(f"Getting selections: {prompt}")
        selected = get_images_from_backend(prompt, backend_url)

        cols = st.beta_columns(4)
        for i, img in enumerate(selected):
            cols[i%4].image(img)
        
    except ServiceError as error:
        st.write(f"Service unavailable, status: {error.status_code}")
    except KeyError:
        st.write("""
        **Error: BACKEND_SERVER unset**

        Please, create a file called `.streamlit/secrets.toml` inside the app's folder and include a line to configure the server URL:
        ```
        BACKEND_SERVER="<server url>"
        ```
        """)