File size: 4,165 Bytes
3ca0269
 
cdb6f86
e52146b
 
fd6fd81
e52146b
3ca0269
 
 
 
879455c
a2c0551
e015622
e52146b
 
8cd2153
bca383d
719ed9c
ae2cf82
 
 
0ed5b20
a296341
 
 
3b81d2d
 
a296341
 
 
 
0ed5b20
e52146b
879455c
3ca0269
879455c
 
3ca0269
879455c
 
e52146b
879455c
fd6fd81
e52146b
879455c
 
e52146b
a2c0551
 
 
e015622
 
 
e52146b
 
e015622
e52146b
 
 
e015622
9cfce0d
879455c
e015622
2f64630
 
e015622
2f64630
879455c
faf4ba4
 
 
879455c
 
c20780a
fd6fd81
 
 
e52146b
 
a2c0551
 
e015622
879455c
ca2922b
3ca0269
e015622
 
 
 
 
879455c
 
e015622
1c1e6e9
9052a89
24b3b53
 
 
 
879455c
24b3b53
 
 
879455c
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
# Supported values:
# - VIDEOCHAIN
# - REPLICATE
# - INFERENCE_ENDPOINT
# - INFERENCE_API
# - OPENAI
RENDERING_ENGINE="INFERENCE_API"

# Supported values:
# - INFERENCE_ENDPOINT
# - INFERENCE_API
# - OPENAI
# - GROQ
# - ANTHROPIC
LLM_ENGINE="INFERENCE_API"

# set this to control the number of pages
MAX_NB_PAGES=6

# Set to "true" to create artificial delays and smooth out traffic
NEXT_PUBLIC_ENABLE_RATE_LIMITER="false"

# ------------- HUGGING FACE OAUTH -------------
ENABLE_HUGGING_FACE_OAUTH=
ENABLE_HUGGING_FACE_OAUTH_WALL=
HUGGING_FACE_OAUTH_CLIENT_ID=

# in production this should be the space's domain and/or URL
HUGGING_FACE_OAUTH_REDIRECT_URL=

# this one must be kept secret (and is unused for now)
HUGGING_FACE_OAUTH_SECRET=

# ------------- PROVIDER AUTH ------------
# You only need to configure the access token(s) for the provider(s) you want to use

# HuggingFace.co token: available for the LLM engine and the RENDERING engine
AUTH_HF_API_TOKEN=

# Replicate.com token: available for the RENDERING engine
AUTH_REPLICATE_API_TOKEN=

# OpenAI.dom token: available for the LLM engine and the RENDERING engine
AUTH_OPENAI_API_KEY=

# An experimental RENDERING engine (sorry it is not very documented yet, so you can use one of the other engines)
AUTH_VIDEOCHAIN_API_TOKEN=

# Groq.com key: available for the LLM engine
AUTH_GROQ_API_KEY=

# Anthropic.com key: available for the LLM engine
AUTH_ANTHROPIC_API_KEY=

# ------------- RENDERING API CONFIG --------------

# If you decide to use Replicate for the RENDERING engine
RENDERING_REPLICATE_API_MODEL="stabilityai/sdxl"
RENDERING_REPLICATE_API_MODEL_VERSION="da77bc59ee60423279fd632efb4795ab731d9e3ca9705ef3341091fb989b7eaf"

# If you decide to use a private Hugging Face Inference Endpoint for the RENDERING engine
RENDERING_HF_INFERENCE_ENDPOINT_URL="https://XXXXXXXXXX.endpoints.huggingface.cloud"

# If you decide to use a Hugging Face Inference API model for the RENDERING engine
RENDERING_HF_INFERENCE_API_BASE_MODEL="stabilityai/stable-diffusion-xl-base-1.0"

# If you decide to use a Hugging Face Inference API model for the RENDERING engine
RENDERING_HF_INFERENCE_API_REFINER_MODEL="stabilityai/stable-diffusion-xl-refiner-1.0"

# If your model returns a different file type (eg. jpg or webp) change it here
RENDERING_HF_INFERENCE_API_FILE_TYPE="image/png"

# An experimental RENDERING engine (sorry it is not very documented yet, so you can use one of the other engines)
RENDERING_VIDEOCHAIN_API_URL="http://localhost:7860"

RENDERING_OPENAI_API_BASE_URL="https://api.openai.com/v1"
RENDERING_OPENAI_API_MODEL="dall-e-3"

# ------------- LLM API CONFIG ----------------

LLM_GROQ_API_MODEL="mixtral-8x7b-32768"

# If you decide to use OpenAI for the LLM engine
LLM_OPENAI_API_BASE_URL="https://api.openai.com/v1"
LLM_OPENAI_API_MODEL="gpt-4-turbo"

# If you decide to use Anthropic (eg. Claude) for the LLM engine
# https://docs.anthropic.com/claude/docs/models-overview
LLM_ANTHROPIC_API_MODEL="claude-3-opus-20240229"

# If you decide to use a private Hugging Face Inference Endpoint for the LLM engine
LLM_HF_INFERENCE_ENDPOINT_URL=""

# If you decide to use a Hugging Face Inference API model for the LLM engine
# LLM_HF_INFERENCE_API_MODEL="HuggingFaceH4/zephyr-7b-beta"
LLM_HF_INFERENCE_API_MODEL="HuggingFaceH4/zephyr-7b-beta"

# ----------- COMMUNITY SHARING (OPTIONAL) -----------
# You don't need those community sharing options to run the AI Comic Factory
# locally or on your own server (they are meant to be used by the Hugging Face team)
NEXT_PUBLIC_ENABLE_COMMUNITY_SHARING="false"
COMMUNITY_API_URL=
COMMUNITY_API_TOKEN=
COMMUNITY_API_ID=

# ----------- CENSORSHIP (OPTIONAL) -----------
# censorship is currently disabled, but will be required when we create a "community roll"
# (a public repositoruy of user-generated comic strips)
ENABLE_CENSORSHIP="false"

# Due to the sensitive nature of some of keywords we want to ban (users try all kind of crazy illegal things)
# the words are are not put in clear in the source code, but behind an encryption key
# (I don't want the project to be flagged by an AI robot police on GitHub or something)
SECRET_FINGERPRINT=""