Commit
•
879455c
1
Parent(s):
274b152
update env variables
Browse files- .env +37 -19
- CONTRIBUTORS.md +9 -0
- README.md +6 -5
- package-lock.json +240 -0
- package.json +1 -0
- src/app/engine/caption.ts +1 -1
- src/app/engine/censorship.ts +175 -30
- src/app/engine/render.ts +21 -62
- src/app/queries/getStory.ts +4 -4
- src/app/queries/getStyle.ts +4 -4
- src/app/queries/predict.ts +3 -135
- src/app/queries/predictWithHuggingFace.ts +90 -0
- src/app/queries/predictWithOpenAI.ts +33 -0
.env
CHANGED
@@ -8,47 +8,65 @@ RENDERING_ENGINE="INFERENCE_API"
|
|
8 |
# Supported values:
|
9 |
# - INFERENCE_ENDPOINT
|
10 |
# - INFERENCE_API
|
|
|
11 |
LLM_ENGINE="INFERENCE_API"
|
12 |
|
13 |
# ------------- PROVIDER AUTH ------------
|
|
|
14 |
|
15 |
-
#
|
16 |
-
|
17 |
-
HF_API_TOKEN=
|
18 |
|
19 |
-
# Replicate token
|
20 |
-
|
21 |
|
22 |
-
# OpenAI
|
23 |
-
|
24 |
|
25 |
-
#
|
26 |
-
|
27 |
|
28 |
# ------------- RENDERING API CONFIG --------------
|
29 |
|
30 |
-
|
31 |
-
|
32 |
RENDERING_REPLICATE_API_MODEL="stabilityai/sdxl"
|
33 |
RENDERING_REPLICATE_API_MODEL_VERSION="da77bc59ee60423279fd632efb4795ab731d9e3ca9705ef3341091fb989b7eaf"
|
34 |
|
35 |
-
|
36 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
37 |
|
38 |
# ------------- LLM API CONFIG ----------------
|
39 |
|
40 |
-
#
|
41 |
-
|
42 |
-
|
43 |
|
44 |
-
#
|
45 |
-
|
|
|
|
|
46 |
LLM_HF_INFERENCE_API_MODEL="codellama/CodeLlama-7b-hf"
|
47 |
|
48 |
# ----------- COMMUNITY SHARING (OPTIONAL) -----------
|
49 |
-
NEXT_PUBLIC_ENABLE_COMMUNITY_SHARING="false"
|
50 |
# You don't need those community sharing options to run the AI Comic Factory
|
51 |
# locally or on your own server (they are meant to be used by the Hugging Face team)
|
|
|
52 |
COMMUNITY_API_URL=
|
53 |
COMMUNITY_API_TOKEN=
|
54 |
COMMUNITY_API_ID=
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
8 |
# Supported values:
|
9 |
# - INFERENCE_ENDPOINT
|
10 |
# - INFERENCE_API
|
11 |
+
# - OPENAI
|
12 |
LLM_ENGINE="INFERENCE_API"
|
13 |
|
14 |
# ------------- PROVIDER AUTH ------------
|
15 |
+
# You only need to configure the access token(s) for the provider(s) you want to use
|
16 |
|
17 |
+
# HuggingFace.co token: available for the LLM engine and the RENDERING engine
|
18 |
+
AUTH_HF_API_TOKEN=
|
|
|
19 |
|
20 |
+
# Replicate.com token: available for the RENDERING engine
|
21 |
+
AUTH_REPLICATE_API_TOKEN=
|
22 |
|
23 |
+
# OpenAI.dom token: available for the LLM engine and the RENDERING engine
|
24 |
+
AUTH_OPENAI_TOKEN=
|
25 |
|
26 |
+
# An experimental RENDERING engine (sorry it is not very documented yet, so you can use one of the other engines)
|
27 |
+
AUTH_VIDEOCHAIN_API_TOKEN=
|
28 |
|
29 |
# ------------- RENDERING API CONFIG --------------
|
30 |
|
31 |
+
# If you decided to use Replicate for the RENDERING engine
|
|
|
32 |
RENDERING_REPLICATE_API_MODEL="stabilityai/sdxl"
|
33 |
RENDERING_REPLICATE_API_MODEL_VERSION="da77bc59ee60423279fd632efb4795ab731d9e3ca9705ef3341091fb989b7eaf"
|
34 |
|
35 |
+
# If you decided to use a private Hugging Face Inference Endpoint for the RENDERING engine
|
36 |
+
RENDERING_HF_RENDERING_INFERENCE_ENDPOINT_URL="https://XXXXXXXXXX.endpoints.huggingface.cloud"
|
37 |
+
|
38 |
+
# If you decided to use a Hugging Face Inference API model for the RENDERING engine
|
39 |
+
RENDERING_HF_RENDERING_INFERENCE_API_MODEL="stabilityai/stable-diffusion-xl-base-1.0"
|
40 |
+
|
41 |
+
# An experimental RENDERING engine (sorry it is not very documented yet, so you can use one of the other engines)
|
42 |
+
RENDERING_VIDEOCHAIN_API_URL="http://localhost:7860"
|
43 |
|
44 |
# ------------- LLM API CONFIG ----------------
|
45 |
|
46 |
+
# If you decided to use OpenAI for the LLM engine
|
47 |
+
LLM_OPENAI_API_BASE_URL="https://api.openai.com/v1"
|
48 |
+
LLM_OPENAI_API_MODEL="gpt-3.5-turbo"
|
49 |
|
50 |
+
# If you decided to use a private Hugging Face Inference Endpoint for the LLM engine
|
51 |
+
LLM_HF_INFERENCE_ENDPOINT_URL=""
|
52 |
+
|
53 |
+
# If you decided to use a Hugging Face Inference API model for the LLM engine
|
54 |
LLM_HF_INFERENCE_API_MODEL="codellama/CodeLlama-7b-hf"
|
55 |
|
56 |
# ----------- COMMUNITY SHARING (OPTIONAL) -----------
|
|
|
57 |
# You don't need those community sharing options to run the AI Comic Factory
|
58 |
# locally or on your own server (they are meant to be used by the Hugging Face team)
|
59 |
+
NEXT_PUBLIC_ENABLE_COMMUNITY_SHARING="false"
|
60 |
COMMUNITY_API_URL=
|
61 |
COMMUNITY_API_TOKEN=
|
62 |
COMMUNITY_API_ID=
|
63 |
+
|
64 |
+
# ----------- CENSORSHIP (OPTIONAL) -----------
|
65 |
+
# censorship is currently disabled, but will be required when we create a "community roll"
|
66 |
+
# (a public repositoruy of user-generated comic strips)
|
67 |
+
ENABLE_CENSORSHIP="false"
|
68 |
+
|
69 |
+
# Due to the sensitive nature of some of keywords we want to ban (users try all kind of crazy illegal things)
|
70 |
+
# the words are are not put in clear in the source code, but behind an encryption key
|
71 |
+
# (I don't want the project to be flagged by an AI robot police on GitHub or something)
|
72 |
+
SECRET_FINGERPRINT=""
|
CONTRIBUTORS.md
ADDED
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
This project was developed Julian Bilcke (@jbilcke-hf), as part of his work at Hugging Face.
|
2 |
+
|
3 |
+
------------------------------------------
|
4 |
+
|
5 |
+
A huge thanks to external developers for their contributions!
|
6 |
+
|
7 |
+
艾逗笔 (@idoubi):
|
8 |
+
- Added support for OpenAI: https://github.com/jbilcke-hf/ai-comic-factory/pull/6
|
9 |
+
|
README.md
CHANGED
@@ -20,13 +20,14 @@ it requires various components to run for the frontend, backend, LLM, SDXL etc.
|
|
20 |
If you try to duplicate the project, open the `.env` you will see it requires some variables.
|
21 |
|
22 |
Provider config:
|
23 |
-
- `LLM_ENGINE`: can be
|
24 |
-
- `RENDERING_ENGINE`: can
|
25 |
|
26 |
Auth config:
|
27 |
-
- `
|
28 |
-
- `
|
29 |
-
- `
|
|
|
30 |
|
31 |
Rendering config:
|
32 |
- `RENDERING_HF_INFERENCE_ENDPOINT_URL`: necessary if you decide to use a custom inference endpoint
|
|
|
20 |
If you try to duplicate the project, open the `.env` you will see it requires some variables.
|
21 |
|
22 |
Provider config:
|
23 |
+
- `LLM_ENGINE`: can be one of: "INFERENCE_API", "INFERENCE_ENDPOINT", "OPENAI"
|
24 |
+
- `RENDERING_ENGINE`: can be one of: "INFERENCE_API", "INFERENCE_ENDPOINT", "REPLICATE", "VIDEOCHAIN" for now, unless you code your custom solution
|
25 |
|
26 |
Auth config:
|
27 |
+
- `AUTH_HF_API_TOKEN`: only if you decide to use OpenAI for the LLM engine necessary if you decide to use an inference api model or a custom inference endpoint
|
28 |
+
- `AUTH_OPENAI_TOKEN`: only if you decide to use OpenAI for the LLM engine
|
29 |
+
- `AITH_VIDEOCHAIN_API_TOKEN`: secret token to access the VideoChain API server
|
30 |
+
- `AUTH_REPLICATE_API_TOKEN`: in case you want to use Replicate.com
|
31 |
|
32 |
Rendering config:
|
33 |
- `RENDERING_HF_INFERENCE_ENDPOINT_URL`: necessary if you decide to use a custom inference endpoint
|
package-lock.json
CHANGED
@@ -37,11 +37,13 @@
|
|
37 |
"cmdk": "^0.2.0",
|
38 |
"cookies-next": "^2.1.2",
|
39 |
"date-fns": "^2.30.0",
|
|
|
40 |
"eslint": "8.45.0",
|
41 |
"eslint-config-next": "13.4.10",
|
42 |
"html2canvas": "^1.4.1",
|
43 |
"lucide-react": "^0.260.0",
|
44 |
"next": "13.4.10",
|
|
|
45 |
"pick": "^0.0.1",
|
46 |
"postcss": "8.4.26",
|
47 |
"react": "18.2.0",
|
@@ -3691,6 +3693,15 @@
|
|
3691 |
"resolved": "https://registry.npmjs.org/@types/node/-/node-20.4.2.tgz",
|
3692 |
"integrity": "sha512-Dd0BYtWgnWJKwO1jkmTrzofjK2QXXcai0dmtzvIBhcA+RsG5h8R3xlyta0kGOZRNfL9GuRtb1knmPEhQrePCEw=="
|
3693 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
3694 |
"node_modules/@types/prop-types": {
|
3695 |
"version": "15.7.5",
|
3696 |
"resolved": "https://registry.npmjs.org/@types/prop-types/-/prop-types-15.7.5.tgz",
|
@@ -3847,6 +3858,17 @@
|
|
3847 |
"url": "https://opencollective.com/typescript-eslint"
|
3848 |
}
|
3849 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
3850 |
"node_modules/abs-svg-path": {
|
3851 |
"version": "0.1.1",
|
3852 |
"resolved": "https://registry.npmjs.org/abs-svg-path/-/abs-svg-path-0.1.1.tgz",
|
@@ -3879,6 +3901,17 @@
|
|
3879 |
"node": ">=0.4.0"
|
3880 |
}
|
3881 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
3882 |
"node_modules/ajv": {
|
3883 |
"version": "6.12.6",
|
3884 |
"resolved": "https://registry.npmjs.org/ajv/-/ajv-6.12.6.tgz",
|
@@ -4096,6 +4129,11 @@
|
|
4096 |
"has-symbols": "^1.0.3"
|
4097 |
}
|
4098 |
},
|
|
|
|
|
|
|
|
|
|
|
4099 |
"node_modules/autoprefixer": {
|
4100 |
"version": "10.4.14",
|
4101 |
"resolved": "https://registry.npmjs.org/autoprefixer/-/autoprefixer-10.4.14.tgz",
|
@@ -4209,6 +4247,11 @@
|
|
4209 |
"resolved": "https://registry.npmjs.org/balanced-match/-/balanced-match-1.0.2.tgz",
|
4210 |
"integrity": "sha512-3oSeUO0TMV67hN1AmbXsK4yaqU7tjiHlbxRDZOpH0KW9+CeX4bRAaX0Anxt0tx2MrpRpWwQaPwIlISEJhYU5Pw=="
|
4211 |
},
|
|
|
|
|
|
|
|
|
|
|
4212 |
"node_modules/base64-arraybuffer": {
|
4213 |
"version": "1.0.2",
|
4214 |
"resolved": "https://registry.npmjs.org/base64-arraybuffer/-/base64-arraybuffer-1.0.2.tgz",
|
@@ -4430,6 +4473,14 @@
|
|
4430 |
"url": "https://github.com/chalk/chalk?sponsor=1"
|
4431 |
}
|
4432 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
4433 |
"node_modules/chokidar": {
|
4434 |
"version": "3.5.3",
|
4435 |
"resolved": "https://registry.npmjs.org/chokidar/-/chokidar-3.5.3.tgz",
|
@@ -4795,6 +4846,17 @@
|
|
4795 |
"simple-swizzle": "^0.2.2"
|
4796 |
}
|
4797 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
4798 |
"node_modules/command-score": {
|
4799 |
"version": "0.1.2",
|
4800 |
"resolved": "https://registry.npmjs.org/command-score/-/command-score-0.1.2.tgz",
|
@@ -4879,6 +4941,14 @@
|
|
4879 |
"node": ">= 8"
|
4880 |
}
|
4881 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
4882 |
"node_modules/crypto-js": {
|
4883 |
"version": "4.1.1",
|
4884 |
"resolved": "https://registry.npmjs.org/crypto-js/-/crypto-js-4.1.1.tgz",
|
@@ -5012,6 +5082,14 @@
|
|
5012 |
"url": "https://github.com/sponsors/ljharb"
|
5013 |
}
|
5014 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
5015 |
"node_modules/dequal": {
|
5016 |
"version": "2.0.3",
|
5017 |
"resolved": "https://registry.npmjs.org/dequal/-/dequal-2.0.3.tgz",
|
@@ -5051,6 +5129,15 @@
|
|
5051 |
"node": ">=0.3.1"
|
5052 |
}
|
5053 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
5054 |
"node_modules/dir-glob": {
|
5055 |
"version": "3.0.1",
|
5056 |
"resolved": "https://registry.npmjs.org/dir-glob/-/dir-glob-3.0.1.tgz",
|
@@ -5139,6 +5226,14 @@
|
|
5139 |
"resolved": "https://registry.npmjs.org/emoji-regex/-/emoji-regex-10.2.1.tgz",
|
5140 |
"integrity": "sha512-97g6QgOk8zlDRdgq1WxwgTMgEWGVAQvB5Fdpgc1MkNy56la5SKP9GsMXKDOdqwn90/41a8yPwIGk1Y6WVbeMQA=="
|
5141 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
5142 |
"node_modules/end-of-stream": {
|
5143 |
"version": "1.4.4",
|
5144 |
"resolved": "https://registry.npmjs.org/end-of-stream/-/end-of-stream-1.4.4.tgz",
|
@@ -5697,6 +5792,14 @@
|
|
5697 |
"node": ">=0.10.0"
|
5698 |
}
|
5699 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
5700 |
"node_modules/events": {
|
5701 |
"version": "3.3.0",
|
5702 |
"resolved": "https://registry.npmjs.org/events/-/events-3.3.0.tgz",
|
@@ -5846,6 +5949,36 @@
|
|
5846 |
"is-callable": "^1.1.3"
|
5847 |
}
|
5848 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
5849 |
"node_modules/fraction.js": {
|
5850 |
"version": "4.2.1",
|
5851 |
"resolved": "https://registry.npmjs.org/fraction.js/-/fraction.js-4.2.1.tgz",
|
@@ -6197,11 +6330,30 @@
|
|
6197 |
"entities": "^4.4.0"
|
6198 |
}
|
6199 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
6200 |
"node_modules/hyphen": {
|
6201 |
"version": "1.6.6",
|
6202 |
"resolved": "https://registry.npmjs.org/hyphen/-/hyphen-1.6.6.tgz",
|
6203 |
"integrity": "sha512-XtqmnT+b9n5MX+MsqluFAVTIenbtC25iskW0Z+jLd+awfhA+ZbWKWQMIvLJccGoa2bM1R6juWJ27cZxIFOmkWw=="
|
6204 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
6205 |
"node_modules/idb-keyval": {
|
6206 |
"version": "6.2.1",
|
6207 |
"resolved": "https://registry.npmjs.org/idb-keyval/-/idb-keyval-6.2.1.tgz",
|
@@ -6366,6 +6518,11 @@
|
|
6366 |
"url": "https://github.com/sponsors/ljharb"
|
6367 |
}
|
6368 |
},
|
|
|
|
|
|
|
|
|
|
|
6369 |
"node_modules/is-callable": {
|
6370 |
"version": "1.2.7",
|
6371 |
"resolved": "https://registry.npmjs.org/is-callable/-/is-callable-1.2.7.tgz",
|
@@ -6851,6 +7008,16 @@
|
|
6851 |
"resolved": "https://registry.npmjs.org/make-error/-/make-error-1.3.6.tgz",
|
6852 |
"integrity": "sha512-s8UhlNe7vPKomQhC1qFelMokr/Sc3AgNbso3n74mVPA5LTZwkB9NlXf4XPamLxJE8h0gh73rM94xvwRT2CVInw=="
|
6853 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
6854 |
"node_modules/media-engine": {
|
6855 |
"version": "1.0.3",
|
6856 |
"resolved": "https://registry.npmjs.org/media-engine/-/media-engine-1.0.3.tgz",
|
@@ -6876,6 +7043,25 @@
|
|
6876 |
"node": ">=8.6"
|
6877 |
}
|
6878 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
6879 |
"node_modules/mimic-response": {
|
6880 |
"version": "3.1.0",
|
6881 |
"resolved": "https://registry.npmjs.org/mimic-response/-/mimic-response-3.1.0.tgz",
|
@@ -7050,6 +7236,24 @@
|
|
7050 |
"resolved": "https://registry.npmjs.org/node-addon-api/-/node-addon-api-6.1.0.tgz",
|
7051 |
"integrity": "sha512-+eawOlIgy680F0kBzPUNFhMZGtJ1YmqM6l4+Crf4IkImjYrO/mqPwRMh352g23uIaQKFItcQ64I7KMaJxHgAVA=="
|
7052 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
7053 |
"node_modules/node-fetch": {
|
7054 |
"version": "2.7.0",
|
7055 |
"resolved": "https://registry.npmjs.org/node-fetch/-/node-fetch-2.7.0.tgz",
|
@@ -7223,6 +7427,29 @@
|
|
7223 |
"wrappy": "1"
|
7224 |
}
|
7225 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
7226 |
"node_modules/opencollective-postinstall": {
|
7227 |
"version": "2.0.3",
|
7228 |
"resolved": "https://registry.npmjs.org/opencollective-postinstall/-/opencollective-postinstall-2.0.3.tgz",
|
@@ -8043,6 +8270,11 @@
|
|
8043 |
"url": "https://github.com/sponsors/ljharb"
|
8044 |
}
|
8045 |
},
|
|
|
|
|
|
|
|
|
|
|
8046 |
"node_modules/sanitize-html": {
|
8047 |
"version": "2.11.0",
|
8048 |
"resolved": "https://registry.npmjs.org/sanitize-html/-/sanitize-html-2.11.0.tgz",
|
@@ -9067,6 +9299,14 @@
|
|
9067 |
"node": ">=10.13.0"
|
9068 |
}
|
9069 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
9070 |
"node_modules/webidl-conversions": {
|
9071 |
"version": "3.0.1",
|
9072 |
"resolved": "https://registry.npmjs.org/webidl-conversions/-/webidl-conversions-3.0.1.tgz",
|
|
|
37 |
"cmdk": "^0.2.0",
|
38 |
"cookies-next": "^2.1.2",
|
39 |
"date-fns": "^2.30.0",
|
40 |
+
"encoding": "^0.1.13",
|
41 |
"eslint": "8.45.0",
|
42 |
"eslint-config-next": "13.4.10",
|
43 |
"html2canvas": "^1.4.1",
|
44 |
"lucide-react": "^0.260.0",
|
45 |
"next": "13.4.10",
|
46 |
+
"openai": "^4.10.0",
|
47 |
"pick": "^0.0.1",
|
48 |
"postcss": "8.4.26",
|
49 |
"react": "18.2.0",
|
|
|
3693 |
"resolved": "https://registry.npmjs.org/@types/node/-/node-20.4.2.tgz",
|
3694 |
"integrity": "sha512-Dd0BYtWgnWJKwO1jkmTrzofjK2QXXcai0dmtzvIBhcA+RsG5h8R3xlyta0kGOZRNfL9GuRtb1knmPEhQrePCEw=="
|
3695 |
},
|
3696 |
+
"node_modules/@types/node-fetch": {
|
3697 |
+
"version": "2.6.6",
|
3698 |
+
"resolved": "https://registry.npmjs.org/@types/node-fetch/-/node-fetch-2.6.6.tgz",
|
3699 |
+
"integrity": "sha512-95X8guJYhfqiuVVhRFxVQcf4hW/2bCuoPwDasMf/531STFoNoWTT7YDnWdXHEZKqAGUigmpG31r2FE70LwnzJw==",
|
3700 |
+
"dependencies": {
|
3701 |
+
"@types/node": "*",
|
3702 |
+
"form-data": "^4.0.0"
|
3703 |
+
}
|
3704 |
+
},
|
3705 |
"node_modules/@types/prop-types": {
|
3706 |
"version": "15.7.5",
|
3707 |
"resolved": "https://registry.npmjs.org/@types/prop-types/-/prop-types-15.7.5.tgz",
|
|
|
3858 |
"url": "https://opencollective.com/typescript-eslint"
|
3859 |
}
|
3860 |
},
|
3861 |
+
"node_modules/abort-controller": {
|
3862 |
+
"version": "3.0.0",
|
3863 |
+
"resolved": "https://registry.npmjs.org/abort-controller/-/abort-controller-3.0.0.tgz",
|
3864 |
+
"integrity": "sha512-h8lQ8tacZYnR3vNQTgibj+tODHI5/+l06Au2Pcriv/Gmet0eaj4TwWH41sO9wnHDiQsEj19q0drzdWdeAHtweg==",
|
3865 |
+
"dependencies": {
|
3866 |
+
"event-target-shim": "^5.0.0"
|
3867 |
+
},
|
3868 |
+
"engines": {
|
3869 |
+
"node": ">=6.5"
|
3870 |
+
}
|
3871 |
+
},
|
3872 |
"node_modules/abs-svg-path": {
|
3873 |
"version": "0.1.1",
|
3874 |
"resolved": "https://registry.npmjs.org/abs-svg-path/-/abs-svg-path-0.1.1.tgz",
|
|
|
3901 |
"node": ">=0.4.0"
|
3902 |
}
|
3903 |
},
|
3904 |
+
"node_modules/agentkeepalive": {
|
3905 |
+
"version": "4.5.0",
|
3906 |
+
"resolved": "https://registry.npmjs.org/agentkeepalive/-/agentkeepalive-4.5.0.tgz",
|
3907 |
+
"integrity": "sha512-5GG/5IbQQpC9FpkRGsSvZI5QYeSCzlJHdpBQntCsuTOxhKD8lqKhrleg2Yi7yvMIf82Ycmmqln9U8V9qwEiJew==",
|
3908 |
+
"dependencies": {
|
3909 |
+
"humanize-ms": "^1.2.1"
|
3910 |
+
},
|
3911 |
+
"engines": {
|
3912 |
+
"node": ">= 8.0.0"
|
3913 |
+
}
|
3914 |
+
},
|
3915 |
"node_modules/ajv": {
|
3916 |
"version": "6.12.6",
|
3917 |
"resolved": "https://registry.npmjs.org/ajv/-/ajv-6.12.6.tgz",
|
|
|
4129 |
"has-symbols": "^1.0.3"
|
4130 |
}
|
4131 |
},
|
4132 |
+
"node_modules/asynckit": {
|
4133 |
+
"version": "0.4.0",
|
4134 |
+
"resolved": "https://registry.npmjs.org/asynckit/-/asynckit-0.4.0.tgz",
|
4135 |
+
"integrity": "sha512-Oei9OH4tRh0YqU3GxhX79dM/mwVgvbZJaSNaRk+bshkj0S5cfHcgYakreBjrHwatXKbz+IoIdYLxrKim2MjW0Q=="
|
4136 |
+
},
|
4137 |
"node_modules/autoprefixer": {
|
4138 |
"version": "10.4.14",
|
4139 |
"resolved": "https://registry.npmjs.org/autoprefixer/-/autoprefixer-10.4.14.tgz",
|
|
|
4247 |
"resolved": "https://registry.npmjs.org/balanced-match/-/balanced-match-1.0.2.tgz",
|
4248 |
"integrity": "sha512-3oSeUO0TMV67hN1AmbXsK4yaqU7tjiHlbxRDZOpH0KW9+CeX4bRAaX0Anxt0tx2MrpRpWwQaPwIlISEJhYU5Pw=="
|
4249 |
},
|
4250 |
+
"node_modules/base-64": {
|
4251 |
+
"version": "0.1.0",
|
4252 |
+
"resolved": "https://registry.npmjs.org/base-64/-/base-64-0.1.0.tgz",
|
4253 |
+
"integrity": "sha512-Y5gU45svrR5tI2Vt/X9GPd3L0HNIKzGu202EjxrXMpuc2V2CiKgemAbUUsqYmZJvPtCXoUKjNZwBJzsNScUbXA=="
|
4254 |
+
},
|
4255 |
"node_modules/base64-arraybuffer": {
|
4256 |
"version": "1.0.2",
|
4257 |
"resolved": "https://registry.npmjs.org/base64-arraybuffer/-/base64-arraybuffer-1.0.2.tgz",
|
|
|
4473 |
"url": "https://github.com/chalk/chalk?sponsor=1"
|
4474 |
}
|
4475 |
},
|
4476 |
+
"node_modules/charenc": {
|
4477 |
+
"version": "0.0.2",
|
4478 |
+
"resolved": "https://registry.npmjs.org/charenc/-/charenc-0.0.2.tgz",
|
4479 |
+
"integrity": "sha512-yrLQ/yVUFXkzg7EDQsPieE/53+0RlaWTs+wBrvW36cyilJ2SaDWfl4Yj7MtLTXleV9uEKefbAGUPv2/iWSooRA==",
|
4480 |
+
"engines": {
|
4481 |
+
"node": "*"
|
4482 |
+
}
|
4483 |
+
},
|
4484 |
"node_modules/chokidar": {
|
4485 |
"version": "3.5.3",
|
4486 |
"resolved": "https://registry.npmjs.org/chokidar/-/chokidar-3.5.3.tgz",
|
|
|
4846 |
"simple-swizzle": "^0.2.2"
|
4847 |
}
|
4848 |
},
|
4849 |
+
"node_modules/combined-stream": {
|
4850 |
+
"version": "1.0.8",
|
4851 |
+
"resolved": "https://registry.npmjs.org/combined-stream/-/combined-stream-1.0.8.tgz",
|
4852 |
+
"integrity": "sha512-FQN4MRfuJeHf7cBbBMJFXhKSDq+2kAArBlmRBvcvFE5BB1HZKXtSFASDhdlz9zOYwxh8lDdnvmMOe/+5cdoEdg==",
|
4853 |
+
"dependencies": {
|
4854 |
+
"delayed-stream": "~1.0.0"
|
4855 |
+
},
|
4856 |
+
"engines": {
|
4857 |
+
"node": ">= 0.8"
|
4858 |
+
}
|
4859 |
+
},
|
4860 |
"node_modules/command-score": {
|
4861 |
"version": "0.1.2",
|
4862 |
"resolved": "https://registry.npmjs.org/command-score/-/command-score-0.1.2.tgz",
|
|
|
4941 |
"node": ">= 8"
|
4942 |
}
|
4943 |
},
|
4944 |
+
"node_modules/crypt": {
|
4945 |
+
"version": "0.0.2",
|
4946 |
+
"resolved": "https://registry.npmjs.org/crypt/-/crypt-0.0.2.tgz",
|
4947 |
+
"integrity": "sha512-mCxBlsHFYh9C+HVpiEacem8FEBnMXgU9gy4zmNC+SXAZNB/1idgp/aulFJ4FgCi7GPEVbfyng092GqL2k2rmow==",
|
4948 |
+
"engines": {
|
4949 |
+
"node": "*"
|
4950 |
+
}
|
4951 |
+
},
|
4952 |
"node_modules/crypto-js": {
|
4953 |
"version": "4.1.1",
|
4954 |
"resolved": "https://registry.npmjs.org/crypto-js/-/crypto-js-4.1.1.tgz",
|
|
|
5082 |
"url": "https://github.com/sponsors/ljharb"
|
5083 |
}
|
5084 |
},
|
5085 |
+
"node_modules/delayed-stream": {
|
5086 |
+
"version": "1.0.0",
|
5087 |
+
"resolved": "https://registry.npmjs.org/delayed-stream/-/delayed-stream-1.0.0.tgz",
|
5088 |
+
"integrity": "sha512-ZySD7Nf91aLB0RxL4KGrKHBXl7Eds1DAmEdcoVawXnLD7SDhpNgtuII2aAkg7a7QS41jxPSZ17p4VdGnMHk3MQ==",
|
5089 |
+
"engines": {
|
5090 |
+
"node": ">=0.4.0"
|
5091 |
+
}
|
5092 |
+
},
|
5093 |
"node_modules/dequal": {
|
5094 |
"version": "2.0.3",
|
5095 |
"resolved": "https://registry.npmjs.org/dequal/-/dequal-2.0.3.tgz",
|
|
|
5129 |
"node": ">=0.3.1"
|
5130 |
}
|
5131 |
},
|
5132 |
+
"node_modules/digest-fetch": {
|
5133 |
+
"version": "1.3.0",
|
5134 |
+
"resolved": "https://registry.npmjs.org/digest-fetch/-/digest-fetch-1.3.0.tgz",
|
5135 |
+
"integrity": "sha512-CGJuv6iKNM7QyZlM2T3sPAdZWd/p9zQiRNS9G+9COUCwzWFTs0Xp8NF5iePx7wtvhDykReiRRrSeNb4oMmB8lA==",
|
5136 |
+
"dependencies": {
|
5137 |
+
"base-64": "^0.1.0",
|
5138 |
+
"md5": "^2.3.0"
|
5139 |
+
}
|
5140 |
+
},
|
5141 |
"node_modules/dir-glob": {
|
5142 |
"version": "3.0.1",
|
5143 |
"resolved": "https://registry.npmjs.org/dir-glob/-/dir-glob-3.0.1.tgz",
|
|
|
5226 |
"resolved": "https://registry.npmjs.org/emoji-regex/-/emoji-regex-10.2.1.tgz",
|
5227 |
"integrity": "sha512-97g6QgOk8zlDRdgq1WxwgTMgEWGVAQvB5Fdpgc1MkNy56la5SKP9GsMXKDOdqwn90/41a8yPwIGk1Y6WVbeMQA=="
|
5228 |
},
|
5229 |
+
"node_modules/encoding": {
|
5230 |
+
"version": "0.1.13",
|
5231 |
+
"resolved": "https://registry.npmjs.org/encoding/-/encoding-0.1.13.tgz",
|
5232 |
+
"integrity": "sha512-ETBauow1T35Y/WZMkio9jiM0Z5xjHHmJ4XmjZOq1l/dXz3lr2sRn87nJy20RupqSh1F2m3HHPSp8ShIPQJrJ3A==",
|
5233 |
+
"dependencies": {
|
5234 |
+
"iconv-lite": "^0.6.2"
|
5235 |
+
}
|
5236 |
+
},
|
5237 |
"node_modules/end-of-stream": {
|
5238 |
"version": "1.4.4",
|
5239 |
"resolved": "https://registry.npmjs.org/end-of-stream/-/end-of-stream-1.4.4.tgz",
|
|
|
5792 |
"node": ">=0.10.0"
|
5793 |
}
|
5794 |
},
|
5795 |
+
"node_modules/event-target-shim": {
|
5796 |
+
"version": "5.0.1",
|
5797 |
+
"resolved": "https://registry.npmjs.org/event-target-shim/-/event-target-shim-5.0.1.tgz",
|
5798 |
+
"integrity": "sha512-i/2XbnSz/uxRCU6+NdVJgKWDTM427+MqYbkQzD321DuCQJUqOuJKIA0IM2+W2xtYHdKOmZ4dR6fExsd4SXL+WQ==",
|
5799 |
+
"engines": {
|
5800 |
+
"node": ">=6"
|
5801 |
+
}
|
5802 |
+
},
|
5803 |
"node_modules/events": {
|
5804 |
"version": "3.3.0",
|
5805 |
"resolved": "https://registry.npmjs.org/events/-/events-3.3.0.tgz",
|
|
|
5949 |
"is-callable": "^1.1.3"
|
5950 |
}
|
5951 |
},
|
5952 |
+
"node_modules/form-data": {
|
5953 |
+
"version": "4.0.0",
|
5954 |
+
"resolved": "https://registry.npmjs.org/form-data/-/form-data-4.0.0.tgz",
|
5955 |
+
"integrity": "sha512-ETEklSGi5t0QMZuiXoA/Q6vcnxcLQP5vdugSpuAyi6SVGi2clPPp+xgEhuMaHC+zGgn31Kd235W35f7Hykkaww==",
|
5956 |
+
"dependencies": {
|
5957 |
+
"asynckit": "^0.4.0",
|
5958 |
+
"combined-stream": "^1.0.8",
|
5959 |
+
"mime-types": "^2.1.12"
|
5960 |
+
},
|
5961 |
+
"engines": {
|
5962 |
+
"node": ">= 6"
|
5963 |
+
}
|
5964 |
+
},
|
5965 |
+
"node_modules/form-data-encoder": {
|
5966 |
+
"version": "1.7.2",
|
5967 |
+
"resolved": "https://registry.npmjs.org/form-data-encoder/-/form-data-encoder-1.7.2.tgz",
|
5968 |
+
"integrity": "sha512-qfqtYan3rxrnCk1VYaA4H+Ms9xdpPqvLZa6xmMgFvhO32x7/3J/ExcTd6qpxM0vH2GdMI+poehyBZvqfMTto8A=="
|
5969 |
+
},
|
5970 |
+
"node_modules/formdata-node": {
|
5971 |
+
"version": "4.4.1",
|
5972 |
+
"resolved": "https://registry.npmjs.org/formdata-node/-/formdata-node-4.4.1.tgz",
|
5973 |
+
"integrity": "sha512-0iirZp3uVDjVGt9p49aTaqjk84TrglENEDuqfdlZQ1roC9CWlPk6Avf8EEnZNcAqPonwkG35x4n3ww/1THYAeQ==",
|
5974 |
+
"dependencies": {
|
5975 |
+
"node-domexception": "1.0.0",
|
5976 |
+
"web-streams-polyfill": "4.0.0-beta.3"
|
5977 |
+
},
|
5978 |
+
"engines": {
|
5979 |
+
"node": ">= 12.20"
|
5980 |
+
}
|
5981 |
+
},
|
5982 |
"node_modules/fraction.js": {
|
5983 |
"version": "4.2.1",
|
5984 |
"resolved": "https://registry.npmjs.org/fraction.js/-/fraction.js-4.2.1.tgz",
|
|
|
6330 |
"entities": "^4.4.0"
|
6331 |
}
|
6332 |
},
|
6333 |
+
"node_modules/humanize-ms": {
|
6334 |
+
"version": "1.2.1",
|
6335 |
+
"resolved": "https://registry.npmjs.org/humanize-ms/-/humanize-ms-1.2.1.tgz",
|
6336 |
+
"integrity": "sha512-Fl70vYtsAFb/C06PTS9dZBo7ihau+Tu/DNCk/OyHhea07S+aeMWpFFkUaXRa8fI+ScZbEI8dfSxwY7gxZ9SAVQ==",
|
6337 |
+
"dependencies": {
|
6338 |
+
"ms": "^2.0.0"
|
6339 |
+
}
|
6340 |
+
},
|
6341 |
"node_modules/hyphen": {
|
6342 |
"version": "1.6.6",
|
6343 |
"resolved": "https://registry.npmjs.org/hyphen/-/hyphen-1.6.6.tgz",
|
6344 |
"integrity": "sha512-XtqmnT+b9n5MX+MsqluFAVTIenbtC25iskW0Z+jLd+awfhA+ZbWKWQMIvLJccGoa2bM1R6juWJ27cZxIFOmkWw=="
|
6345 |
},
|
6346 |
+
"node_modules/iconv-lite": {
|
6347 |
+
"version": "0.6.3",
|
6348 |
+
"resolved": "https://registry.npmjs.org/iconv-lite/-/iconv-lite-0.6.3.tgz",
|
6349 |
+
"integrity": "sha512-4fCk79wshMdzMp2rH06qWrJE4iolqLhCUH+OiuIgU++RB0+94NlDL81atO7GX55uUKueo0txHNtvEyI6D7WdMw==",
|
6350 |
+
"dependencies": {
|
6351 |
+
"safer-buffer": ">= 2.1.2 < 3.0.0"
|
6352 |
+
},
|
6353 |
+
"engines": {
|
6354 |
+
"node": ">=0.10.0"
|
6355 |
+
}
|
6356 |
+
},
|
6357 |
"node_modules/idb-keyval": {
|
6358 |
"version": "6.2.1",
|
6359 |
"resolved": "https://registry.npmjs.org/idb-keyval/-/idb-keyval-6.2.1.tgz",
|
|
|
6518 |
"url": "https://github.com/sponsors/ljharb"
|
6519 |
}
|
6520 |
},
|
6521 |
+
"node_modules/is-buffer": {
|
6522 |
+
"version": "1.1.6",
|
6523 |
+
"resolved": "https://registry.npmjs.org/is-buffer/-/is-buffer-1.1.6.tgz",
|
6524 |
+
"integrity": "sha512-NcdALwpXkTm5Zvvbk7owOUSvVvBKDgKP5/ewfXEznmQFfs4ZRmanOeKBTjRVjka3QFoN6XJ+9F3USqfHqTaU5w=="
|
6525 |
+
},
|
6526 |
"node_modules/is-callable": {
|
6527 |
"version": "1.2.7",
|
6528 |
"resolved": "https://registry.npmjs.org/is-callable/-/is-callable-1.2.7.tgz",
|
|
|
7008 |
"resolved": "https://registry.npmjs.org/make-error/-/make-error-1.3.6.tgz",
|
7009 |
"integrity": "sha512-s8UhlNe7vPKomQhC1qFelMokr/Sc3AgNbso3n74mVPA5LTZwkB9NlXf4XPamLxJE8h0gh73rM94xvwRT2CVInw=="
|
7010 |
},
|
7011 |
+
"node_modules/md5": {
|
7012 |
+
"version": "2.3.0",
|
7013 |
+
"resolved": "https://registry.npmjs.org/md5/-/md5-2.3.0.tgz",
|
7014 |
+
"integrity": "sha512-T1GITYmFaKuO91vxyoQMFETst+O71VUPEU3ze5GNzDm0OWdP8v1ziTaAEPUr/3kLsY3Sftgz242A1SetQiDL7g==",
|
7015 |
+
"dependencies": {
|
7016 |
+
"charenc": "0.0.2",
|
7017 |
+
"crypt": "0.0.2",
|
7018 |
+
"is-buffer": "~1.1.6"
|
7019 |
+
}
|
7020 |
+
},
|
7021 |
"node_modules/media-engine": {
|
7022 |
"version": "1.0.3",
|
7023 |
"resolved": "https://registry.npmjs.org/media-engine/-/media-engine-1.0.3.tgz",
|
|
|
7043 |
"node": ">=8.6"
|
7044 |
}
|
7045 |
},
|
7046 |
+
"node_modules/mime-db": {
|
7047 |
+
"version": "1.52.0",
|
7048 |
+
"resolved": "https://registry.npmjs.org/mime-db/-/mime-db-1.52.0.tgz",
|
7049 |
+
"integrity": "sha512-sPU4uV7dYlvtWJxwwxHD0PuihVNiE7TyAbQ5SWxDCB9mUYvOgroQOwYQQOKPJ8CIbE+1ETVlOoK1UC2nU3gYvg==",
|
7050 |
+
"engines": {
|
7051 |
+
"node": ">= 0.6"
|
7052 |
+
}
|
7053 |
+
},
|
7054 |
+
"node_modules/mime-types": {
|
7055 |
+
"version": "2.1.35",
|
7056 |
+
"resolved": "https://registry.npmjs.org/mime-types/-/mime-types-2.1.35.tgz",
|
7057 |
+
"integrity": "sha512-ZDY+bPm5zTTF+YpCrAU9nK0UgICYPT0QtT1NZWFv4s++TNkcgVaT0g6+4R2uI4MjQjzysHB1zxuWL50hzaeXiw==",
|
7058 |
+
"dependencies": {
|
7059 |
+
"mime-db": "1.52.0"
|
7060 |
+
},
|
7061 |
+
"engines": {
|
7062 |
+
"node": ">= 0.6"
|
7063 |
+
}
|
7064 |
+
},
|
7065 |
"node_modules/mimic-response": {
|
7066 |
"version": "3.1.0",
|
7067 |
"resolved": "https://registry.npmjs.org/mimic-response/-/mimic-response-3.1.0.tgz",
|
|
|
7236 |
"resolved": "https://registry.npmjs.org/node-addon-api/-/node-addon-api-6.1.0.tgz",
|
7237 |
"integrity": "sha512-+eawOlIgy680F0kBzPUNFhMZGtJ1YmqM6l4+Crf4IkImjYrO/mqPwRMh352g23uIaQKFItcQ64I7KMaJxHgAVA=="
|
7238 |
},
|
7239 |
+
"node_modules/node-domexception": {
|
7240 |
+
"version": "1.0.0",
|
7241 |
+
"resolved": "https://registry.npmjs.org/node-domexception/-/node-domexception-1.0.0.tgz",
|
7242 |
+
"integrity": "sha512-/jKZoMpw0F8GRwl4/eLROPA3cfcXtLApP0QzLmUT/HuPCZWyB7IY9ZrMeKw2O/nFIqPQB3PVM9aYm0F312AXDQ==",
|
7243 |
+
"funding": [
|
7244 |
+
{
|
7245 |
+
"type": "github",
|
7246 |
+
"url": "https://github.com/sponsors/jimmywarting"
|
7247 |
+
},
|
7248 |
+
{
|
7249 |
+
"type": "github",
|
7250 |
+
"url": "https://paypal.me/jimmywarting"
|
7251 |
+
}
|
7252 |
+
],
|
7253 |
+
"engines": {
|
7254 |
+
"node": ">=10.5.0"
|
7255 |
+
}
|
7256 |
+
},
|
7257 |
"node_modules/node-fetch": {
|
7258 |
"version": "2.7.0",
|
7259 |
"resolved": "https://registry.npmjs.org/node-fetch/-/node-fetch-2.7.0.tgz",
|
|
|
7427 |
"wrappy": "1"
|
7428 |
}
|
7429 |
},
|
7430 |
+
"node_modules/openai": {
|
7431 |
+
"version": "4.10.0",
|
7432 |
+
"resolved": "https://registry.npmjs.org/openai/-/openai-4.10.0.tgz",
|
7433 |
+
"integrity": "sha512-II4b5/7qzwYkqA9MSjgqdofCc798EW+dtF2h6qNaVLet+qO7FShAJTWnoyzb50J4ZH1rPxRFAsmDLIhY3PT6DQ==",
|
7434 |
+
"dependencies": {
|
7435 |
+
"@types/node": "^18.11.18",
|
7436 |
+
"@types/node-fetch": "^2.6.4",
|
7437 |
+
"abort-controller": "^3.0.0",
|
7438 |
+
"agentkeepalive": "^4.2.1",
|
7439 |
+
"digest-fetch": "^1.3.0",
|
7440 |
+
"form-data-encoder": "1.7.2",
|
7441 |
+
"formdata-node": "^4.3.2",
|
7442 |
+
"node-fetch": "^2.6.7"
|
7443 |
+
},
|
7444 |
+
"bin": {
|
7445 |
+
"openai": "bin/cli"
|
7446 |
+
}
|
7447 |
+
},
|
7448 |
+
"node_modules/openai/node_modules/@types/node": {
|
7449 |
+
"version": "18.17.19",
|
7450 |
+
"resolved": "https://registry.npmjs.org/@types/node/-/node-18.17.19.tgz",
|
7451 |
+
"integrity": "sha512-+pMhShR3Or5GR0/sp4Da7FnhVmTalWm81M6MkEldbwjETSaPalw138Z4KdpQaistvqQxLB7Cy4xwYdxpbSOs9Q=="
|
7452 |
+
},
|
7453 |
"node_modules/opencollective-postinstall": {
|
7454 |
"version": "2.0.3",
|
7455 |
"resolved": "https://registry.npmjs.org/opencollective-postinstall/-/opencollective-postinstall-2.0.3.tgz",
|
|
|
8270 |
"url": "https://github.com/sponsors/ljharb"
|
8271 |
}
|
8272 |
},
|
8273 |
+
"node_modules/safer-buffer": {
|
8274 |
+
"version": "2.1.2",
|
8275 |
+
"resolved": "https://registry.npmjs.org/safer-buffer/-/safer-buffer-2.1.2.tgz",
|
8276 |
+
"integrity": "sha512-YZo3K82SD7Riyi0E1EQPojLz7kpepnSQI9IyPbHHg1XXXevb5dJI7tpyN2ADxGcQbHG7vcyRHk0cbwqcQriUtg=="
|
8277 |
+
},
|
8278 |
"node_modules/sanitize-html": {
|
8279 |
"version": "2.11.0",
|
8280 |
"resolved": "https://registry.npmjs.org/sanitize-html/-/sanitize-html-2.11.0.tgz",
|
|
|
9299 |
"node": ">=10.13.0"
|
9300 |
}
|
9301 |
},
|
9302 |
+
"node_modules/web-streams-polyfill": {
|
9303 |
+
"version": "4.0.0-beta.3",
|
9304 |
+
"resolved": "https://registry.npmjs.org/web-streams-polyfill/-/web-streams-polyfill-4.0.0-beta.3.tgz",
|
9305 |
+
"integrity": "sha512-QW95TCTaHmsYfHDybGMwO5IJIM93I/6vTRk+daHTWFPhwh+C8Cg7j7XyKrwrj8Ib6vYXe0ocYNrmzY4xAAN6ug==",
|
9306 |
+
"engines": {
|
9307 |
+
"node": ">= 14"
|
9308 |
+
}
|
9309 |
+
},
|
9310 |
"node_modules/webidl-conversions": {
|
9311 |
"version": "3.0.1",
|
9312 |
"resolved": "https://registry.npmjs.org/webidl-conversions/-/webidl-conversions-3.0.1.tgz",
|
package.json
CHANGED
@@ -38,6 +38,7 @@
|
|
38 |
"cmdk": "^0.2.0",
|
39 |
"cookies-next": "^2.1.2",
|
40 |
"date-fns": "^2.30.0",
|
|
|
41 |
"eslint": "8.45.0",
|
42 |
"eslint-config-next": "13.4.10",
|
43 |
"html2canvas": "^1.4.1",
|
|
|
38 |
"cmdk": "^0.2.0",
|
39 |
"cookies-next": "^2.1.2",
|
40 |
"date-fns": "^2.30.0",
|
41 |
+
"encoding": "^0.1.13",
|
42 |
"eslint": "8.45.0",
|
43 |
"eslint-config-next": "13.4.10",
|
44 |
"html2canvas": "^1.4.1",
|
src/app/engine/caption.ts
CHANGED
@@ -33,7 +33,7 @@ export async function see({
|
|
33 |
headers: {
|
34 |
Accept: "application/json",
|
35 |
"Content-Type": "application/json",
|
36 |
-
// Authorization: `Bearer ${
|
37 |
},
|
38 |
body: JSON.stringify(request),
|
39 |
cache: 'no-store',
|
|
|
33 |
headers: {
|
34 |
Accept: "application/json",
|
35 |
"Content-Type": "application/json",
|
36 |
+
// Authorization: `Bearer ${videochainApi}`,
|
37 |
},
|
38 |
body: JSON.stringify(request),
|
39 |
cache: 'no-store',
|
src/app/engine/censorship.ts
CHANGED
@@ -1,39 +1,184 @@
|
|
1 |
|
2 |
-
|
3 |
-
// unfortunately due to abuse by some users, I have to add this NSFW filter
|
4 |
-
const secretSalt = `${process.env.SECRET_CENSORSHIP_KEY || ""}`
|
5 |
-
|
6 |
-
// TODO the censorship is not implement yet actually
|
7 |
-
|
8 |
// I don't want to be banned by Replicate because bad actors are asking
|
9 |
// for some naked anime stuff or whatever
|
10 |
// I also want to avoid a PR scandal due to some bad user generated content
|
11 |
|
12 |
-
|
13 |
-
|
14 |
-
|
15 |
-
|
16 |
-
|
17 |
-
|
18 |
-
"
|
19 |
-
"
|
20 |
-
"
|
21 |
-
"
|
22 |
-
"
|
23 |
-
"
|
24 |
-
"
|
25 |
-
|
26 |
-
|
27 |
-
|
28 |
-
"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
29 |
]
|
30 |
|
31 |
-
|
32 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
33 |
export const filterOutBadWords = (sentence: string) => {
|
34 |
-
|
35 |
-
|
36 |
-
|
37 |
-
|
38 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
39 |
}
|
|
|
1 |
|
|
|
|
|
|
|
|
|
|
|
|
|
2 |
// I don't want to be banned by Replicate because bad actors are asking
|
3 |
// for some naked anime stuff or whatever
|
4 |
// I also want to avoid a PR scandal due to some bad user generated content
|
5 |
|
6 |
+
import { computeSecretFingerprint } from "@/lib/computeSecretFingerprint"
|
7 |
+
|
8 |
+
// those keywords have been generated by looking at the logs of the panorama and the AI Comic Factory
|
9 |
+
// those are real requests some users tried to attempt.. :|
|
10 |
+
|
11 |
+
const chickens = [
|
12 |
+
"fcb4dacbd99b21368c50f29c1d47071c87cf2225ab9192282c785460391cd365",
|
13 |
+
"68840b60ac27eacaa7afe17e898d3c4a2dc71acff8c74d6782c1bcaafd14963d",
|
14 |
+
"67f745224fd6e1a7a3a244514d5807fcc994cbb62ca4ec8fa44cd14244a515ae",
|
15 |
+
"681fea565117808c6dbe002520d2cfeeb3e5c67e68630afb4a453449a9da587b",
|
16 |
+
"2f3d913b3db9e15a930aac43eb2d6fe8817db8e4bcf37794bf0227b06b718d1b",
|
17 |
+
"922a700b807e4994df82eba2b48a6ac131fe8d8d1035d06b3592d622fb232161",
|
18 |
+
"cb69ee6774eafcc720adb1f689d28acbb9f47998cbea0299ec66a58dedf91c37"
|
19 |
+
]
|
20 |
+
|
21 |
+
const ducks = [
|
22 |
+
"1c52cb20c0cbc76349fa63232b982bd394cf0850ebc17240dcf33c19fb15a26d",
|
23 |
+
"e1d4de9b8d464d7da07c276b63a42c1c9922224f0a6cab6b0826427ce4a7461a",
|
24 |
+
"0be3174bfb1a48a65875c2f035b1ae14fbc8f232f55785018de0cfe2132fa952",
|
25 |
+
"0f174769641b2e5d2c79b5a83e8ef91e004f6f3e62531cd70cfdff02159268cb",
|
26 |
+
"e9fb8ae8ff720acd91025229478a21e43e8e976e30119a76c293201adf572736",
|
27 |
+
"f65a0dc0e07b5d084ff24c69dcdb953f7b57101d2ebb716d4dfb5963076ef807",
|
28 |
+
"2bf38af1646489c2c086f811d082054cd29e23fa7bb5c525396bec01b3ab688e"
|
29 |
+
]
|
30 |
+
|
31 |
+
const cats = [
|
32 |
+
"fcffc3e997d952007d1b902a9cf40b750ba4a410ac65bfd95475996bf51359e4",
|
33 |
+
"3172a5fa159754d703489dfba5af520b8ace107cdf170f4c4cb38a6797aa163f",
|
34 |
+
"500012dbff4498a9c4513369d6b9b373fab9330ffd2cb1e622294043cc21b610",
|
35 |
+
"84e3a8d34ee7d0c8e7a2926dd1acad46a0b66b9d27725b3a7e5053550f490301"
|
36 |
+
]
|
37 |
+
|
38 |
+
const roasted = [
|
39 |
+
"a2bfbce0046c9a52a0eabf98f73e0f8e09959970431fc892ebdb4e1c97031b50",
|
40 |
+
"6eca1adf06851f99e9cdfbb496c27d46ff81106903d11f3346a146e96082b016",
|
41 |
+
"49a124c9ed6fbbad4105b3657dc25de369bcafb9d6787f610c08f584cd607d0f",
|
42 |
+
"c3afb59420c812cbc7c8f57ad3e8d79407f10106a99f829aa65316c99d0b29c4",
|
43 |
+
"2b808858836a5c205080f5b93201ef92e098cff931d8de6d9f20dc722997d077",
|
44 |
+
"07bef89d1a7d63c9c5ed64ba0f73d6cff689811847c2e20c8b3fbfb060e1d64e",
|
45 |
+
"baeb994922d5473f534aa54322d83effe74c6c4dac807e6b523a677d7acdc17b",
|
46 |
+
"ea4735a879edd5cc94ca7db26edd5a970df69a41f0009d3444486647e44175af",
|
47 |
+
"f2412249030454cd13ac6f7965871d924c16daacda0123de81892adb19ce49ac",
|
48 |
+
"9958c56e12bab8549cf752bcd8bec4ac36cf79c404b1faf5611f057bb71bc0e1",
|
49 |
+
"76cdade0b3d4caf0888f60318a5cbca00f830a3b0bf37735fc64fdaeb67c34d3",
|
50 |
+
"1bf53c97869e1ea89bda19da64a9173d48fe4ec823e949e2c898f8abb3fbf457",
|
51 |
+
"1bf53c97869e1ea89bda19da64a9173d48fe4ec823e949e2c898f8abb3fbf457",
|
52 |
+
"3d7f973fab8f4a19c0a3e59efe970ed7bd55a1cb795752d9cbe3c19e8a7d81ec"
|
53 |
]
|
54 |
|
55 |
+
const banned = [
|
56 |
+
"8a05d4869d9d6ce388c6cd2db13ca12b88097b90f9be027d5ffaaa467c7a6e5e",
|
57 |
+
"0c475212a608138244c5fc150b1563e5ef79c516234fd78dcd5993f726c359a0",
|
58 |
+
"df17388805f99f2ff3e5ae97a0f55e5c927eb47f17ca65822bf8c88f02bac3dd",
|
59 |
+
"86c3355d1bd581cdf7306729d8dd0ee9b7a317b9cfd6d7a6f5fad9c0dafe2167",
|
60 |
+
"23a2484cd420c9ffbfcc2c0075a9b330664450ced1fc64ab6a65e278086b8c6e",
|
61 |
+
"fb4cabe709b62eea1b4cc0030c76f5e4a43ee677ce19124e8e7bafa86c78ab66",
|
62 |
+
"d99c26daee85f7dc81c46c061a5874cff7179ed72d884d2316d664d36ffe7ab5",
|
63 |
+
"b93c38af5aa221d76c60ee3eb762efee0cdb0daf29ceb235b7dda6d46c06490d",
|
64 |
+
"8cf6c8765dc757319461dd9a785e77c201b8e5a604d36b817cd987c6a5e62500",
|
65 |
+
"f4a1cb290745717f86c3cee30fc324c0d80a9945fcbc7bbeb010579f58792f1e",
|
66 |
+
"7c87c47c42fc983119551342be9ddd5b32e530c0504ccdbbaa1e12b1d9f1bbcb",
|
67 |
+
"d04fad4f21d030da7a1301afbf480ef6246eb7bbf0f26e31865b2e015a25f747",
|
68 |
+
"d685ff22fb9da01ee949db212770729603989850864ef7a7085e1f086cfa7deb",
|
69 |
+
"533b90588d9ccf7967da54691f575e9fd4926c6e0b5fd94a47b932bcea270bee",
|
70 |
+
"9c2d61f28f5bb7f3f1dc9122be64cda8a428b46ce68b70120da4c41dba96ba4c",
|
71 |
+
"5d4b1a3eebe64dfa631d0e3b084bd96ee9364c3669269f838ca17a4900276264",
|
72 |
+
"d56f56413b9679fc0820a2c0237224ded8554c61fab8959c174123c8b68ba029",
|
73 |
+
"323a9ab60739726070d615ff3a05d7ff6bb6e3c4dd9ff16ce24f253ecd7b8851",
|
74 |
+
"975c6739de7d4999db15972f707f5f4e95649275f1c0c48e895b8c537e8638ec",
|
75 |
+
"67ee26eb9e1c1c7124797321b02bca90a19c18171782917cd4a487b722484dce",
|
76 |
+
"6df5aa7b72a4e6e3fb726489ff1437daa5752047507f4da912680b1d6647c7d6",
|
77 |
+
"b0864805364359e8c5810c233b1bf2c74dedce9055ae5f7680ba05b4e39db8e2",
|
78 |
+
"a8f841472ecffdd6266151148320c8e36847a24ead9d3338e0313b075c16649d",
|
79 |
+
"f9b127cd90e85b0ff68dd220361671663f0154b2b827f1f7ea797b020ca0018c",
|
80 |
+
"d5c20e9a1ecf01c82da24c514d867498b3e5f522adc1523ce29404a6563641d5",
|
81 |
+
"241022b49d7c0aba24a61eea1137a804f36e4bcb47af42950275baac9b4e7aac",
|
82 |
+
"fc99a70e17b6c86ef1b537654b0f50353567a7b59912c3ba955f3fca4d1ea696",
|
83 |
+
"255306e968009003d295cb2a7256f27bfcdb5d1743bf4d9f2aa4b8adf1a7734d",
|
84 |
+
"048c7b709763dd9c43794d241c369f0abcb079d546ddcbbba9968a1ed1da7ed7",
|
85 |
+
"520cbfeef3e4c405d79478eedccb97a4d476be585626dd2b1c53292797491bc7",
|
86 |
+
"f9f28a7ae7e8b1719b350a04dc087a4b8e33478d109ceeef6ba892b32d1105c9",
|
87 |
+
"d177f1bfe603647ef4c1c0e6f1a7172081fb9bbc2ea859705949f2c5aa5d4f22",
|
88 |
+
"302feef2c09247fbd23789581f7f5e2219f88ae0a937880954938573c2a52a84",
|
89 |
+
"99edd6f57b864873835f16f19c805dd94bed9da8967b84e3a62782f106d9ebcc",
|
90 |
+
"e75e5f01dcd8351c9553e89558085bd68e6feb295dee5d8da0c9b43ee303ce36",
|
91 |
+
"135e52a026aea9d2e12de358a85e05cf21121a18269269b7c62678c3bc846f5b",
|
92 |
+
"28e5b2d3eb5f1ef4cc7b570878b03acf303a6ca4ca95893591e0fb943b0beab0",
|
93 |
+
"a26b26340f8d0363633490556d20bcc250726d10e1431eb8c22d6b1ff3f2b14a",
|
94 |
+
"27e4ddde96ec6a1dbe1cf12d79448b3e72f144944c15b299629542d1b65fbabf",
|
95 |
+
"efd9c0a391ee93251046a58326d1b21b33fe21d71a3fb1855b9048ade53df77c",
|
96 |
+
"6d505fcce416c26a606878aab4d249a034ba2a9846cb1f883e0f9e3fb76ba6da",
|
97 |
+
"3a37b8a1b72f9bca51233536d50f9c8d33a787434684787871e0049c82347cda",
|
98 |
+
"16f9b451184a7c3148344c7d0315f5312ca20553d2271912ecaad91810d977e6",
|
99 |
+
"7406537eb74d1885bd05e191228de313b13702a64d90ae1736c6377b25ab579a",
|
100 |
+
"7e4d1395ae18980015cab16c85ffa20b4cb90a2db594126e893d0f7ac6eecaa8",
|
101 |
+
"ba813ee6c25698f0f68a07121d38bb47c9aa404c1ab0a6e767595cb75e1747b8",
|
102 |
+
"6586c93f3ece83e01ecc1eb84a7711e7975826a388d478a009468ea0ed9dc03e",
|
103 |
+
"8960174c74d86e03ae88fb6774580170e49952f2286d960be08c556bbd0dda95",
|
104 |
+
"4d611454369aa1a4e2b7eed1734fac5d480f08fb86b87a162967e416370f2a8e",
|
105 |
+
"59d48440f85eabf565fe8d3bc6b973ba64c70df3b36b0511e0e67ceca91762b3",
|
106 |
+
"cd926926e2af74e43d1a6a420a7e1933b78662320477a3c018b2711d8765e339",
|
107 |
+
"80e90057df6a59823f51aafac36ed5bc4e5ac26d675d9c1467501590c82f12d4",
|
108 |
+
"a9cf28b869b70e258adde5639a048f866ec86f8f3f3d53bfc960b86aa6da9239",
|
109 |
+
"cc2adbf8ac0cddeefa304d7b20f14a7e047a4b2299cc5e8f898f5c59660bd964",
|
110 |
+
"92a150a46146e9d3f84899cf15e12514af684e7ee18d7add782ddd4f4a15ef18",
|
111 |
+
"d9b2e84ef6dc0ce449357d52c9095f69b173a1b848ea2921199d33b0ec10024a",
|
112 |
+
"a9329a7e4d367a0135c1ca86c6ce5ecabcc26529235229d71b6bf991f7689e21",
|
113 |
+
"8f160c6fd8ccc3fb2a371a4b52748f0bd030766627c4322e2911fe82f6b10497",
|
114 |
+
"620e96eae4f3e88cbe0770292b33724c5df3866d83f39df6380441f7271c80e2",
|
115 |
+
"cafa3481fa3c45ed1e55cd0129c12b477eeab5aa3d6da20cae6d6292f19b0e6d",
|
116 |
+
"be07994e9a83aa3689e79b6e96123676ccc4fa29f523c28c750c6d60505531ee",
|
117 |
+
"f6498069768cd3aa79b2b0c91879694f05a259c8ee4a6bb343f0435f74eb1b53",
|
118 |
+
"c9b6b26cb3a694eb78fcac0a14ad18d46d50907186a9add41022d31d191b2b65"
|
119 |
+
]
|
120 |
+
|
121 |
+
const young = [
|
122 |
+
"ffdf66787b4a33b78b18c18822e334cfe2c8406caf442851deef451bd43140a1",
|
123 |
+
"858f22219afc4b32a7ba9a27a213d7f495e77c3cceed8147eae5282bf3e23d39",
|
124 |
+
"8c3c46df84ace3d58d4ce0fbc513017986b33c6002ae369d9f7dd1f892a898cb",
|
125 |
+
"66caa22b9483fdf026ce67de61067d81535a7c9b3169cbc5c2a455ac8dcc7bec",
|
126 |
+
"76893047b1eff9fadc7be07b13adb5aaed9c73bcdeea46ee07098605e2c7ff76",
|
127 |
+
"526cb848754e2baaa17376a5693d90ba3f69f71fd2a866f22876ac8a075849a7",
|
128 |
+
"f59c38e31d0f64dc1bfcdf34451723bc1a65570e209e5496c8d1d7f6d3d649db",
|
129 |
+
"e013a67e275c62c1402ccbbb11ad14afb8b8a82318a44c07d67599ed5ac874de",
|
130 |
+
"3bef34219fb07f867ecbff4d6748f598d6cc0761e17dd0d431ee1f4ec3281374",
|
131 |
+
"8211bf5f613fac06cd5d074d34c16dfacc9367c8afaa6ad3aff99d145e5221be"
|
132 |
+
]
|
133 |
+
|
134 |
+
const getFingerprint = (word: string) => {
|
135 |
+
return computeSecretFingerprint(
|
136 |
+
word.toLocaleLowerCase().replaceAll(/[^a-zA-Z0-9]/gi, "")
|
137 |
+
)
|
138 |
+
}
|
139 |
+
|
140 |
+
const encode = (list: string[]) => {
|
141 |
+
console.log(JSON.stringify(
|
142 |
+
list.sort((a, b) => (b.length - a.length))
|
143 |
+
.map(item => getFingerprint(item)), null, 2))
|
144 |
+
}
|
145 |
+
|
146 |
+
// encode([ "badword" ])
|
147 |
+
|
148 |
export const filterOutBadWords = (sentence: string) => {
|
149 |
+
if (process.env.ENABLE_CENSORSHIP !== "true") { return sentence }
|
150 |
+
|
151 |
+
let requireCensorship = false
|
152 |
+
|
153 |
+
const words = sentence.replaceAll(/[^a-zA-Z0-9]/gi, " ").replaceAll(/\s+/gi, " ").trim().split(" ")
|
154 |
+
|
155 |
+
const sanitized = words.map(word => {
|
156 |
+
const fingerprint = getFingerprint(word)
|
157 |
+
|
158 |
+
let result: string = word
|
159 |
+
// some users want to play it smart and bypass our system so let's play too
|
160 |
+
if (chickens.includes(fingerprint)) {
|
161 |
+
result = "large chicken"
|
162 |
+
} else if (ducks.includes(fingerprint)) {
|
163 |
+
result = "big duck"
|
164 |
+
} else if (cats.includes(fingerprint)) {
|
165 |
+
result = "cat"
|
166 |
+
} else if (roasted.includes(fingerprint)) {
|
167 |
+
result = "roasted chicken"
|
168 |
+
} else if (young.includes(fingerprint)) {
|
169 |
+
result = "adult"
|
170 |
+
} else if (banned.includes(fingerprint)) {
|
171 |
+
result = "_BANNED_"
|
172 |
+
}
|
173 |
+
|
174 |
+
if (result !== word) {
|
175 |
+
requireCensorship = true
|
176 |
+
}
|
177 |
+
return result
|
178 |
+
}).filter(item => item !== "_BANNED_").join(" ")
|
179 |
+
|
180 |
+
// if the user didn't try to use a bad word, we leave it untouched
|
181 |
+
// he words array has been degraded by the replace operation, but it removes commas etc which isn't great
|
182 |
+
// so if the request was genuine and SFW, it's best to return the original prompt
|
183 |
+
return requireCensorship ? sanitized : sentence
|
184 |
}
|
src/app/engine/render.ts
CHANGED
@@ -9,16 +9,18 @@ import { sleep } from "@/lib/sleep"
|
|
9 |
|
10 |
const renderingEngine = `${process.env.RENDERING_ENGINE || ""}` as RenderingEngine
|
11 |
|
12 |
-
|
13 |
-
|
14 |
-
const huggingFaceToken = `${process.env.HF_API_TOKEN || ""}`
|
15 |
const huggingFaceInferenceEndpointUrl = `${process.env.RENDERING_HF_INFERENCE_ENDPOINT_URL || ""}`
|
16 |
const huggingFaceInferenceApiModel = `${process.env.RENDERING_HF_INFERENCE_API_MODEL || ""}`
|
17 |
|
18 |
-
const replicateToken = `${process.env.
|
19 |
const replicateModel = `${process.env.RENDERING_REPLICATE_API_MODEL || ""}`
|
20 |
const replicateModelVersion = `${process.env.RENDERING_REPLICATE_API_MODEL_VERSION || ""}`
|
21 |
|
|
|
|
|
|
|
22 |
export async function newRender({
|
23 |
prompt,
|
24 |
// negativePrompt,
|
@@ -30,10 +32,10 @@ export async function newRender({
|
|
30 |
width: number
|
31 |
height: number
|
32 |
}) {
|
33 |
-
// console.log(`newRender(${prompt})`)
|
34 |
if (!prompt) {
|
35 |
-
|
36 |
-
|
|
|
37 |
}
|
38 |
|
39 |
let defaulResult: RenderedScene = {
|
@@ -69,7 +71,7 @@ export async function newRender({
|
|
69 |
|
70 |
// console.log("prediction:", prediction)
|
71 |
|
72 |
-
// no need to reply straight away
|
73 |
// also our friends at Replicate won't like it if we spam them with requests
|
74 |
await sleep(4000)
|
75 |
|
@@ -93,8 +95,6 @@ export async function newRender({
|
|
93 |
throw new Error(`you need to configure your RENDERING_HF_INFERENCE_API_MODEL in order to use the INFERENCE_API rendering engine`)
|
94 |
}
|
95 |
|
96 |
-
const seed = generateSeed()
|
97 |
-
|
98 |
const url = renderingEngine === "INFERENCE_ENDPOINT"
|
99 |
? huggingFaceInferenceEndpointUrl
|
100 |
: `https://api-inference.huggingface.co/models/${huggingFaceInferenceApiModel}`
|
@@ -102,7 +102,6 @@ export async function newRender({
|
|
102 |
const res = await fetch(url, {
|
103 |
method: "POST",
|
104 |
headers: {
|
105 |
-
// Accept: "application/json",
|
106 |
"Content-Type": "application/json",
|
107 |
Authorization: `Bearer ${huggingFaceToken}`,
|
108 |
},
|
@@ -148,13 +147,12 @@ export async function newRender({
|
|
148 |
segments: []
|
149 |
} as RenderedScene
|
150 |
} else {
|
151 |
-
// console.log(`calling POST ${apiUrl}/render with prompt: ${prompt}`)
|
152 |
const res = await fetch(`${videochainApiUrl}/render`, {
|
153 |
method: "POST",
|
154 |
headers: {
|
155 |
Accept: "application/json",
|
156 |
"Content-Type": "application/json",
|
157 |
-
Authorization: `Bearer ${
|
158 |
},
|
159 |
body: JSON.stringify({
|
160 |
prompt,
|
@@ -182,14 +180,7 @@ export async function newRender({
|
|
182 |
// next: { revalidate: 1 }
|
183 |
})
|
184 |
|
185 |
-
|
186 |
-
// console.log("res:", res)
|
187 |
-
// The return value is *not* serialized
|
188 |
-
// You can return Date, Map, Set, etc.
|
189 |
-
|
190 |
-
// Recommendation: handle errors
|
191 |
if (res.status !== 200) {
|
192 |
-
// This will activate the closest `error.js` Error Boundary
|
193 |
throw new Error('Failed to fetch data')
|
194 |
}
|
195 |
|
@@ -204,8 +195,9 @@ export async function newRender({
|
|
204 |
|
205 |
export async function getRender(renderId: string) {
|
206 |
if (!renderId) {
|
207 |
-
|
208 |
-
|
|
|
209 |
}
|
210 |
|
211 |
let defaulResult: RenderedScene = {
|
@@ -221,24 +213,15 @@ export async function getRender(renderId: string) {
|
|
221 |
try {
|
222 |
if (renderingEngine === "REPLICATE") {
|
223 |
if (!replicateToken) {
|
224 |
-
throw new Error(`you need to configure your
|
225 |
}
|
226 |
if (!replicateModel) {
|
227 |
-
throw new Error(`you need to configure your
|
228 |
}
|
229 |
|
230 |
-
// const replicate = new Replicate({ auth: replicateToken })
|
231 |
-
|
232 |
-
// console.log("Calling replicate..")
|
233 |
-
// const prediction = await replicate.predictions.get(renderId)
|
234 |
-
// console.log("Prediction:", prediction)
|
235 |
-
|
236 |
-
// console.log(`calling GET https://api.replicate.com/v1/predictions/${renderId}`)
|
237 |
const res = await fetch(`https://api.replicate.com/v1/predictions/${renderId}`, {
|
238 |
method: "GET",
|
239 |
headers: {
|
240 |
-
// Accept: "application/json",
|
241 |
-
// "Content-Type": "application/json",
|
242 |
Authorization: `Token ${replicateToken}`,
|
243 |
},
|
244 |
cache: 'no-store',
|
@@ -246,10 +229,6 @@ export async function getRender(renderId: string) {
|
|
246 |
// next: { revalidate: 1 }
|
247 |
})
|
248 |
|
249 |
-
// console.log("res:", res)
|
250 |
-
// The return value is *not* serialized
|
251 |
-
// You can return Date, Map, Set, etc.
|
252 |
-
|
253 |
// Recommendation: handle errors
|
254 |
if (res.status !== 200) {
|
255 |
// This will activate the closest `error.js` Error Boundary
|
@@ -257,7 +236,6 @@ export async function getRender(renderId: string) {
|
|
257 |
}
|
258 |
|
259 |
const response = (await res.json()) as any
|
260 |
-
// console.log("response:", response)
|
261 |
|
262 |
return {
|
263 |
renderId,
|
@@ -275,36 +253,26 @@ export async function getRender(renderId: string) {
|
|
275 |
headers: {
|
276 |
Accept: "application/json",
|
277 |
"Content-Type": "application/json",
|
278 |
-
Authorization: `Bearer ${
|
279 |
},
|
280 |
cache: 'no-store',
|
281 |
// we can also use this (see https://vercel.com/blog/vercel-cache-api-nextjs-cache)
|
282 |
// next: { revalidate: 1 }
|
283 |
})
|
284 |
-
|
285 |
-
// console.log("res:", res)
|
286 |
-
// The return value is *not* serialized
|
287 |
-
// You can return Date, Map, Set, etc.
|
288 |
|
289 |
-
// Recommendation: handle errors
|
290 |
if (res.status !== 200) {
|
291 |
-
// This will activate the closest `error.js` Error Boundary
|
292 |
throw new Error('Failed to fetch data')
|
293 |
}
|
294 |
|
295 |
const response = (await res.json()) as RenderedScene
|
296 |
-
// console.log("response:", response)
|
297 |
return response
|
298 |
}
|
299 |
} catch (err) {
|
300 |
console.error(err)
|
301 |
defaulResult.status = "error"
|
302 |
defaulResult.error = `${err}`
|
303 |
-
// Gorgon.clear(cacheKey)
|
304 |
return defaulResult
|
305 |
}
|
306 |
-
|
307 |
-
// }, cacheDurationInSec * 1000)
|
308 |
}
|
309 |
|
310 |
export async function upscaleImage(image: string): Promise<{
|
@@ -312,8 +280,9 @@ export async function upscaleImage(image: string): Promise<{
|
|
312 |
error: string
|
313 |
}> {
|
314 |
if (!image) {
|
315 |
-
|
316 |
-
|
|
|
317 |
}
|
318 |
|
319 |
let defaulResult = {
|
@@ -328,7 +297,7 @@ export async function upscaleImage(image: string): Promise<{
|
|
328 |
headers: {
|
329 |
Accept: "application/json",
|
330 |
"Content-Type": "application/json",
|
331 |
-
Authorization: `Bearer ${
|
332 |
},
|
333 |
cache: 'no-store',
|
334 |
body: JSON.stringify({ image, factor: 3 })
|
@@ -336,13 +305,7 @@ export async function upscaleImage(image: string): Promise<{
|
|
336 |
// next: { revalidate: 1 }
|
337 |
})
|
338 |
|
339 |
-
// console.log("res:", res)
|
340 |
-
// The return value is *not* serialized
|
341 |
-
// You can return Date, Map, Set, etc.
|
342 |
-
|
343 |
-
// Recommendation: handle errors
|
344 |
if (res.status !== 200) {
|
345 |
-
// This will activate the closest `error.js` Error Boundary
|
346 |
throw new Error('Failed to fetch data')
|
347 |
}
|
348 |
|
@@ -350,13 +313,9 @@ export async function upscaleImage(image: string): Promise<{
|
|
350 |
assetUrl: string
|
351 |
error: string
|
352 |
}
|
353 |
-
// console.log("response:", response)
|
354 |
return response
|
355 |
} catch (err) {
|
356 |
console.error(err)
|
357 |
-
// Gorgon.clear(cacheKey)
|
358 |
return defaulResult
|
359 |
}
|
360 |
-
|
361 |
-
// }, cacheDurationInSec * 1000)
|
362 |
}
|
|
|
9 |
|
10 |
const renderingEngine = `${process.env.RENDERING_ENGINE || ""}` as RenderingEngine
|
11 |
|
12 |
+
// TODO: we should split Hugging Face and Replicate backends into separate files
|
13 |
+
const huggingFaceToken = `${process.env.AUTH_HF_API_TOKEN || ""}`
|
|
|
14 |
const huggingFaceInferenceEndpointUrl = `${process.env.RENDERING_HF_INFERENCE_ENDPOINT_URL || ""}`
|
15 |
const huggingFaceInferenceApiModel = `${process.env.RENDERING_HF_INFERENCE_API_MODEL || ""}`
|
16 |
|
17 |
+
const replicateToken = `${process.env.AUTH_REPLICATE_API_TOKEN || ""}`
|
18 |
const replicateModel = `${process.env.RENDERING_REPLICATE_API_MODEL || ""}`
|
19 |
const replicateModelVersion = `${process.env.RENDERING_REPLICATE_API_MODEL_VERSION || ""}`
|
20 |
|
21 |
+
const videochainToken = `${process.env.AUTH_VIDEOCHAIN_API_TOKEN || ""}`
|
22 |
+
const videochainApiUrl = `${process.env.RENDERING_VIDEOCHAIN_API_URL || ""}`
|
23 |
+
|
24 |
export async function newRender({
|
25 |
prompt,
|
26 |
// negativePrompt,
|
|
|
32 |
width: number
|
33 |
height: number
|
34 |
}) {
|
|
|
35 |
if (!prompt) {
|
36 |
+
const error = `cannot call the rendering API without a prompt, aborting..`
|
37 |
+
console.error(error)
|
38 |
+
throw new Error(error)
|
39 |
}
|
40 |
|
41 |
let defaulResult: RenderedScene = {
|
|
|
71 |
|
72 |
// console.log("prediction:", prediction)
|
73 |
|
74 |
+
// no need to reply straight away as images take time to generate, this isn't instantaneous
|
75 |
// also our friends at Replicate won't like it if we spam them with requests
|
76 |
await sleep(4000)
|
77 |
|
|
|
95 |
throw new Error(`you need to configure your RENDERING_HF_INFERENCE_API_MODEL in order to use the INFERENCE_API rendering engine`)
|
96 |
}
|
97 |
|
|
|
|
|
98 |
const url = renderingEngine === "INFERENCE_ENDPOINT"
|
99 |
? huggingFaceInferenceEndpointUrl
|
100 |
: `https://api-inference.huggingface.co/models/${huggingFaceInferenceApiModel}`
|
|
|
102 |
const res = await fetch(url, {
|
103 |
method: "POST",
|
104 |
headers: {
|
|
|
105 |
"Content-Type": "application/json",
|
106 |
Authorization: `Bearer ${huggingFaceToken}`,
|
107 |
},
|
|
|
147 |
segments: []
|
148 |
} as RenderedScene
|
149 |
} else {
|
|
|
150 |
const res = await fetch(`${videochainApiUrl}/render`, {
|
151 |
method: "POST",
|
152 |
headers: {
|
153 |
Accept: "application/json",
|
154 |
"Content-Type": "application/json",
|
155 |
+
Authorization: `Bearer ${videochainToken}`,
|
156 |
},
|
157 |
body: JSON.stringify({
|
158 |
prompt,
|
|
|
180 |
// next: { revalidate: 1 }
|
181 |
})
|
182 |
|
|
|
|
|
|
|
|
|
|
|
|
|
183 |
if (res.status !== 200) {
|
|
|
184 |
throw new Error('Failed to fetch data')
|
185 |
}
|
186 |
|
|
|
195 |
|
196 |
export async function getRender(renderId: string) {
|
197 |
if (!renderId) {
|
198 |
+
const error = `cannot call the rendering API without a renderId, aborting..`
|
199 |
+
console.error(error)
|
200 |
+
throw new Error(error)
|
201 |
}
|
202 |
|
203 |
let defaulResult: RenderedScene = {
|
|
|
213 |
try {
|
214 |
if (renderingEngine === "REPLICATE") {
|
215 |
if (!replicateToken) {
|
216 |
+
throw new Error(`you need to configure your AUTH_REPLICATE_API_TOKEN in order to use the REPLICATE rendering engine`)
|
217 |
}
|
218 |
if (!replicateModel) {
|
219 |
+
throw new Error(`you need to configure your RENDERING_REPLICATE_API_MODEL in order to use the REPLICATE rendering engine`)
|
220 |
}
|
221 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
222 |
const res = await fetch(`https://api.replicate.com/v1/predictions/${renderId}`, {
|
223 |
method: "GET",
|
224 |
headers: {
|
|
|
|
|
225 |
Authorization: `Token ${replicateToken}`,
|
226 |
},
|
227 |
cache: 'no-store',
|
|
|
229 |
// next: { revalidate: 1 }
|
230 |
})
|
231 |
|
|
|
|
|
|
|
|
|
232 |
// Recommendation: handle errors
|
233 |
if (res.status !== 200) {
|
234 |
// This will activate the closest `error.js` Error Boundary
|
|
|
236 |
}
|
237 |
|
238 |
const response = (await res.json()) as any
|
|
|
239 |
|
240 |
return {
|
241 |
renderId,
|
|
|
253 |
headers: {
|
254 |
Accept: "application/json",
|
255 |
"Content-Type": "application/json",
|
256 |
+
Authorization: `Bearer ${videochainToken}`,
|
257 |
},
|
258 |
cache: 'no-store',
|
259 |
// we can also use this (see https://vercel.com/blog/vercel-cache-api-nextjs-cache)
|
260 |
// next: { revalidate: 1 }
|
261 |
})
|
|
|
|
|
|
|
|
|
262 |
|
|
|
263 |
if (res.status !== 200) {
|
|
|
264 |
throw new Error('Failed to fetch data')
|
265 |
}
|
266 |
|
267 |
const response = (await res.json()) as RenderedScene
|
|
|
268 |
return response
|
269 |
}
|
270 |
} catch (err) {
|
271 |
console.error(err)
|
272 |
defaulResult.status = "error"
|
273 |
defaulResult.error = `${err}`
|
|
|
274 |
return defaulResult
|
275 |
}
|
|
|
|
|
276 |
}
|
277 |
|
278 |
export async function upscaleImage(image: string): Promise<{
|
|
|
280 |
error: string
|
281 |
}> {
|
282 |
if (!image) {
|
283 |
+
const error = `cannot call the rendering API without an image, aborting..`
|
284 |
+
console.error(error)
|
285 |
+
throw new Error(error)
|
286 |
}
|
287 |
|
288 |
let defaulResult = {
|
|
|
297 |
headers: {
|
298 |
Accept: "application/json",
|
299 |
"Content-Type": "application/json",
|
300 |
+
Authorization: `Bearer ${videochainToken}`,
|
301 |
},
|
302 |
cache: 'no-store',
|
303 |
body: JSON.stringify({ image, factor: 3 })
|
|
|
305 |
// next: { revalidate: 1 }
|
306 |
})
|
307 |
|
|
|
|
|
|
|
|
|
|
|
308 |
if (res.status !== 200) {
|
|
|
309 |
throw new Error('Failed to fetch data')
|
310 |
}
|
311 |
|
|
|
313 |
assetUrl: string
|
314 |
error: string
|
315 |
}
|
|
|
316 |
return response
|
317 |
} catch (err) {
|
318 |
console.error(err)
|
|
|
319 |
return defaulResult
|
320 |
}
|
|
|
|
|
321 |
}
|
src/app/queries/getStory.ts
CHANGED
@@ -37,15 +37,15 @@ export const getStory = async ({
|
|
37 |
let result = ""
|
38 |
|
39 |
try {
|
40 |
-
result = await predict(query)
|
41 |
-
if (!result.
|
42 |
throw new Error("empty result!")
|
43 |
}
|
44 |
} catch (err) {
|
45 |
console.log(`prediction of the story failed, trying again..`)
|
46 |
try {
|
47 |
-
result = await predict(query+".")
|
48 |
-
if (!result.
|
49 |
throw new Error("empty result!")
|
50 |
}
|
51 |
} catch (err) {
|
|
|
37 |
let result = ""
|
38 |
|
39 |
try {
|
40 |
+
result = `${await predict(query) || ""}`.trim()
|
41 |
+
if (!result.length) {
|
42 |
throw new Error("empty result!")
|
43 |
}
|
44 |
} catch (err) {
|
45 |
console.log(`prediction of the story failed, trying again..`)
|
46 |
try {
|
47 |
+
result = `${await predict(query+".") || ""}`.trim()
|
48 |
+
if (!result.length) {
|
49 |
throw new Error("empty result!")
|
50 |
}
|
51 |
} catch (err) {
|
src/app/queries/getStyle.ts
CHANGED
@@ -30,15 +30,15 @@ export const getStory = async ({
|
|
30 |
|
31 |
let result = ""
|
32 |
try {
|
33 |
-
result = await predict(query)
|
34 |
-
if (!result.
|
35 |
throw new Error("empty result!")
|
36 |
}
|
37 |
} catch (err) {
|
38 |
console.log(`prediction of the story failed, trying again..`)
|
39 |
try {
|
40 |
-
result = await predict(query+".")
|
41 |
-
if (!result.
|
42 |
throw new Error("empty result!")
|
43 |
}
|
44 |
} catch (err) {
|
|
|
30 |
|
31 |
let result = ""
|
32 |
try {
|
33 |
+
result = `${await predict(query) || ""}`.trim()
|
34 |
+
if (!result.length) {
|
35 |
throw new Error("empty result!")
|
36 |
}
|
37 |
} catch (err) {
|
38 |
console.log(`prediction of the story failed, trying again..`)
|
39 |
try {
|
40 |
+
result = `${await predict(query+".") || ""}`.trim()
|
41 |
+
if (!result.length) {
|
42 |
throw new Error("empty result!")
|
43 |
}
|
44 |
} catch (err) {
|
src/app/queries/predict.ts
CHANGED
@@ -1,141 +1,9 @@
|
|
1 |
"use server"
|
2 |
|
3 |
-
import { HfInference, HfInferenceEndpoint } from "@huggingface/inference"
|
4 |
-
|
5 |
-
import type { ChatCompletionMessage } from "openai/resources/chat"
|
6 |
import { LLMEngine } from "@/types"
|
7 |
-
import
|
8 |
-
|
9 |
-
const hf = new HfInference(process.env.HF_API_TOKEN)
|
10 |
|
11 |
-
|
12 |
-
// note: we always try "inference endpoint" first
|
13 |
const llmEngine = `${process.env.LLM_ENGINE || ""}` as LLMEngine
|
14 |
-
const inferenceEndpoint = `${process.env.LLM_HF_INFERENCE_ENDPOINT_URL || ""}`
|
15 |
-
const inferenceModel = `${process.env.LLM_HF_INFERENCE_API_MODEL || ""}`
|
16 |
-
const openaiApiKey = `${process.env.LLM_OPENAI_API_KEY || ""}`
|
17 |
-
|
18 |
-
|
19 |
-
let hfie: HfInferenceEndpoint
|
20 |
-
|
21 |
-
switch (llmEngine) {
|
22 |
-
case "INFERENCE_ENDPOINT":
|
23 |
-
if (inferenceEndpoint) {
|
24 |
-
console.log("Using a custom HF Inference Endpoint")
|
25 |
-
hfie = hf.endpoint(inferenceEndpoint)
|
26 |
-
} else {
|
27 |
-
const error = "No Inference Endpoint URL defined"
|
28 |
-
console.error(error)
|
29 |
-
throw new Error(error)
|
30 |
-
}
|
31 |
-
break;
|
32 |
-
|
33 |
-
case "INFERENCE_API":
|
34 |
-
if (inferenceModel) {
|
35 |
-
console.log("Using an HF Inference API Model")
|
36 |
-
} else {
|
37 |
-
const error = "No Inference API model defined"
|
38 |
-
console.error(error)
|
39 |
-
throw new Error(error)
|
40 |
-
}
|
41 |
-
break;
|
42 |
-
|
43 |
-
case "OPENAI":
|
44 |
-
if (openaiApiKey) {
|
45 |
-
console.log("Using an OpenAI API Key")
|
46 |
-
} else {
|
47 |
-
const error = "No OpenAI API key defined"
|
48 |
-
console.error(error)
|
49 |
-
throw new Error(error)
|
50 |
-
}
|
51 |
-
break;
|
52 |
-
|
53 |
-
default:
|
54 |
-
const error = "No Inference Endpoint URL or Inference API Model defined"
|
55 |
-
console.error(error)
|
56 |
-
throw new Error(error)
|
57 |
-
}
|
58 |
-
|
59 |
-
export async function predict(inputs: string) {
|
60 |
-
|
61 |
-
console.log(`predict: `, inputs)
|
62 |
-
|
63 |
-
if (llmEngine==="OPENAI") {
|
64 |
-
return predictWithOpenAI(inputs)
|
65 |
-
}
|
66 |
-
|
67 |
-
const api = llmEngine ==="INFERENCE_ENDPOINT" ? hfie : hf
|
68 |
-
|
69 |
-
let instructions = ""
|
70 |
-
try {
|
71 |
-
for await (const output of api.textGenerationStream({
|
72 |
-
model: llmEngine ==="INFERENCE_ENDPOINT" ? undefined : (inferenceModel || undefined),
|
73 |
-
inputs,
|
74 |
-
parameters: {
|
75 |
-
do_sample: true,
|
76 |
-
// we don't require a lot of token for our task
|
77 |
-
// but to be safe, let's count ~110 tokens per panel
|
78 |
-
max_new_tokens: 450, // 1150,
|
79 |
-
return_full_text: false,
|
80 |
-
}
|
81 |
-
})) {
|
82 |
-
instructions += output.token.text
|
83 |
-
process.stdout.write(output.token.text)
|
84 |
-
if (
|
85 |
-
instructions.includes("</s>") ||
|
86 |
-
instructions.includes("<s>") ||
|
87 |
-
instructions.includes("[INST]") ||
|
88 |
-
instructions.includes("[/INST]") ||
|
89 |
-
instructions.includes("<SYS>") ||
|
90 |
-
instructions.includes("</SYS>") ||
|
91 |
-
instructions.includes("<|end|>") ||
|
92 |
-
instructions.includes("<|assistant|>")
|
93 |
-
) {
|
94 |
-
break
|
95 |
-
}
|
96 |
-
}
|
97 |
-
} catch (err) {
|
98 |
-
console.error(`error during generation: ${err}`)
|
99 |
-
}
|
100 |
-
|
101 |
-
// need to do some cleanup of the garbage the LLM might have gave us
|
102 |
-
return (
|
103 |
-
instructions
|
104 |
-
.replaceAll("<|end|>", "")
|
105 |
-
.replaceAll("<s>", "")
|
106 |
-
.replaceAll("</s>", "")
|
107 |
-
.replaceAll("[INST]", "")
|
108 |
-
.replaceAll("[/INST]", "")
|
109 |
-
.replaceAll("<SYS>", "")
|
110 |
-
.replaceAll("</SYS>", "")
|
111 |
-
.replaceAll("<|assistant|>", "")
|
112 |
-
.replaceAll('""', '"')
|
113 |
-
)
|
114 |
-
}
|
115 |
-
|
116 |
-
async function predictWithOpenAI(inputs: string) {
|
117 |
-
const openaiApiBaseUrl = `${process.env.OPENAI_API_BASE_URL || "https://api.openai.com/v1"}`
|
118 |
-
const openaiApiModel = `${process.env.OPENAI_API_MODEL || "gpt-3.5-turbo"}`
|
119 |
-
|
120 |
-
const openai = new OpenAI({
|
121 |
-
apiKey: openaiApiKey,
|
122 |
-
baseURL: openaiApiBaseUrl,
|
123 |
-
})
|
124 |
-
|
125 |
-
const messages: ChatCompletionMessage[] = [
|
126 |
-
{ role: "system", content: inputs },
|
127 |
-
]
|
128 |
-
|
129 |
-
try {
|
130 |
-
const res = await openai.chat.completions.create({
|
131 |
-
messages: messages,
|
132 |
-
stream: false,
|
133 |
-
model: openaiApiModel,
|
134 |
-
temperature: 0.8
|
135 |
-
})
|
136 |
|
137 |
-
|
138 |
-
} catch (err) {
|
139 |
-
console.error(`error during generation: ${err}`)
|
140 |
-
}
|
141 |
-
}
|
|
|
1 |
"use server"
|
2 |
|
|
|
|
|
|
|
3 |
import { LLMEngine } from "@/types"
|
4 |
+
import { predictWithHuggingFace } from "./predictWithHuggingFace"
|
5 |
+
import { predictWithOpenAI } from "./predictWithOpenAI"
|
|
|
6 |
|
|
|
|
|
7 |
const llmEngine = `${process.env.LLM_ENGINE || ""}` as LLMEngine
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
8 |
|
9 |
+
export const predict = llmEngine === "OPENAI" ? predictWithOpenAI : predictWithHuggingFace
|
|
|
|
|
|
|
|
src/app/queries/predictWithHuggingFace.ts
ADDED
@@ -0,0 +1,90 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"use server"
|
2 |
+
|
3 |
+
import { HfInference, HfInferenceEndpoint } from "@huggingface/inference"
|
4 |
+
import { LLMEngine } from "@/types"
|
5 |
+
|
6 |
+
const hf = new HfInference(process.env.AUTH_HF_API_TOKEN)
|
7 |
+
|
8 |
+
const llmEngine = `${process.env.LLM_ENGINE || ""}` as LLMEngine
|
9 |
+
const inferenceEndpoint = `${process.env.LLM_HF_INFERENCE_ENDPOINT_URL || ""}`
|
10 |
+
const inferenceModel = `${process.env.LLM_HF_INFERENCE_API_MODEL || ""}`
|
11 |
+
|
12 |
+
let hfie: HfInferenceEndpoint = hf
|
13 |
+
|
14 |
+
switch (llmEngine) {
|
15 |
+
case "INFERENCE_ENDPOINT":
|
16 |
+
if (inferenceEndpoint) {
|
17 |
+
console.log("Using a custom HF Inference Endpoint")
|
18 |
+
hfie = hf.endpoint(inferenceEndpoint)
|
19 |
+
} else {
|
20 |
+
const error = "No Inference Endpoint URL defined"
|
21 |
+
console.error(error)
|
22 |
+
throw new Error(error)
|
23 |
+
}
|
24 |
+
break;
|
25 |
+
|
26 |
+
case "INFERENCE_API":
|
27 |
+
if (inferenceModel) {
|
28 |
+
console.log("Using an HF Inference API Model")
|
29 |
+
} else {
|
30 |
+
const error = "No Inference API model defined"
|
31 |
+
console.error(error)
|
32 |
+
throw new Error(error)
|
33 |
+
}
|
34 |
+
break;
|
35 |
+
|
36 |
+
default:
|
37 |
+
const error = "Please check your Hugging Face Inference API or Inference Endpoint settings"
|
38 |
+
console.error(error)
|
39 |
+
throw new Error(error)
|
40 |
+
}
|
41 |
+
|
42 |
+
const api = llmEngine === "INFERENCE_ENDPOINT" ? hfie : hf
|
43 |
+
|
44 |
+
export async function predictWithHuggingFace(inputs: string) {
|
45 |
+
let instructions = ""
|
46 |
+
try {
|
47 |
+
for await (const output of api.textGenerationStream({
|
48 |
+
model: llmEngine === "INFERENCE_ENDPOINT" ? undefined : (inferenceModel || undefined),
|
49 |
+
inputs,
|
50 |
+
parameters: {
|
51 |
+
do_sample: true,
|
52 |
+
// we don't require a lot of token for our task
|
53 |
+
// but to be safe, let's count ~110 tokens per panel
|
54 |
+
max_new_tokens: 450, // 1150,
|
55 |
+
return_full_text: false,
|
56 |
+
}
|
57 |
+
})) {
|
58 |
+
instructions += output.token.text
|
59 |
+
process.stdout.write(output.token.text)
|
60 |
+
if (
|
61 |
+
instructions.includes("</s>") ||
|
62 |
+
instructions.includes("<s>") ||
|
63 |
+
instructions.includes("[INST]") ||
|
64 |
+
instructions.includes("[/INST]") ||
|
65 |
+
instructions.includes("<SYS>") ||
|
66 |
+
instructions.includes("</SYS>") ||
|
67 |
+
instructions.includes("<|end|>") ||
|
68 |
+
instructions.includes("<|assistant|>")
|
69 |
+
) {
|
70 |
+
break
|
71 |
+
}
|
72 |
+
}
|
73 |
+
} catch (err) {
|
74 |
+
console.error(`error during generation: ${err}`)
|
75 |
+
}
|
76 |
+
|
77 |
+
// need to do some cleanup of the garbage the LLM might have gave us
|
78 |
+
return (
|
79 |
+
instructions
|
80 |
+
.replaceAll("<|end|>", "")
|
81 |
+
.replaceAll("<s>", "")
|
82 |
+
.replaceAll("</s>", "")
|
83 |
+
.replaceAll("[INST]", "")
|
84 |
+
.replaceAll("[/INST]", "")
|
85 |
+
.replaceAll("<SYS>", "")
|
86 |
+
.replaceAll("</SYS>", "")
|
87 |
+
.replaceAll("<|assistant|>", "")
|
88 |
+
.replaceAll('""', '"')
|
89 |
+
)
|
90 |
+
}
|
src/app/queries/predictWithOpenAI.ts
ADDED
@@ -0,0 +1,33 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"use server"
|
2 |
+
|
3 |
+
import type { ChatCompletionMessage } from "openai/resources/chat"
|
4 |
+
import OpenAI from "openai"
|
5 |
+
|
6 |
+
const openaiApiKey = `${process.env.AUTH_OPENAI_API_KEY || ""}`
|
7 |
+
|
8 |
+
export async function predictWithOpenAI(inputs: string) {
|
9 |
+
const openaiApiBaseUrl = `${process.env.LLM_OPENAI_API_BASE_URL || "https://api.openai.com/v1"}`
|
10 |
+
const openaiApiModel = `${process.env.LLM_OPENAI_API_MODEL || "gpt-3.5-turbo"}`
|
11 |
+
|
12 |
+
const openai = new OpenAI({
|
13 |
+
apiKey: openaiApiKey,
|
14 |
+
baseURL: openaiApiBaseUrl,
|
15 |
+
})
|
16 |
+
|
17 |
+
const messages: ChatCompletionMessage[] = [
|
18 |
+
{ role: "system", content: inputs },
|
19 |
+
]
|
20 |
+
|
21 |
+
try {
|
22 |
+
const res = await openai.chat.completions.create({
|
23 |
+
messages: messages,
|
24 |
+
stream: false,
|
25 |
+
model: openaiApiModel,
|
26 |
+
temperature: 0.8
|
27 |
+
})
|
28 |
+
|
29 |
+
return res.choices[0].message.content
|
30 |
+
} catch (err) {
|
31 |
+
console.error(`error during generation: ${err}`)
|
32 |
+
}
|
33 |
+
}
|