TobDeBer
commited on
Commit
·
19822a8
1
Parent(s):
52af04b
normal files
Browse files- .gitattributes +3 -0
- README.md +13 -0
- build_diffusion.sh +5 -0
- build_gguf.sh +5 -0
- diffusion/Dockerfile +6 -0
- gguf/Dockerfile +10 -0
.gitattributes
CHANGED
@@ -33,3 +33,6 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
llama-cli filter=lfs diff=lfs merge=lfs -text
|
37 |
+
llama-server filter=lfs diff=lfs merge=lfs -text
|
38 |
+
sd_cuda filter=lfs diff=lfs merge=lfs -text
|
README.md
ADDED
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Container Repository for CPU adaptations of Inference code
|
2 |
+
|
3 |
+
## Variants
|
4 |
+
|
5 |
+
### CPUdiffusion
|
6 |
+
|
7 |
+
- inference diffusion models on CPU
|
8 |
+
- include CUDAonCPU stack
|
9 |
+
|
10 |
+
### CPUgguf
|
11 |
+
|
12 |
+
- inference gguf models on CPU
|
13 |
+
- include GUI libraries
|
build_diffusion.sh
ADDED
@@ -0,0 +1,5 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
podman build --squash-all --tag bookworm:diffusion diffusion
|
2 |
+
podman image prune -f
|
3 |
+
podman save localhost/bookworm:diffusion >test.tar
|
4 |
+
#time xz -9e -T0 test.tar
|
5 |
+
|
build_gguf.sh
ADDED
@@ -0,0 +1,5 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
podman build --squash-all --tag bookworm:gguf gguf
|
2 |
+
podman image prune -f
|
3 |
+
podman save localhost/bookworm:gguf >gguf.tar
|
4 |
+
time xz -6 -T0 gguf.tar
|
5 |
+
|
diffusion/Dockerfile
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
FROM debian:bookworm-slim
|
2 |
+
|
3 |
+
RUN apt-get update && apt-get upgrade
|
4 |
+
RUN apt-get install -y git git-lfs pip cmake python3
|
5 |
+
|
6 |
+
CMD ["sleep", " infinity"]
|
gguf/Dockerfile
ADDED
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
FROM debian:bookworm-slim
|
2 |
+
RUN apt-get update && apt-get upgrade
|
3 |
+
RUN apt-get install -y git git-lfs pip cmake python3
|
4 |
+
|
5 |
+
RUN git clone https://github.com/ggerganov/llama.cpp.git
|
6 |
+
RUN cd llama.cpp && make -j 32 llama-server
|
7 |
+
RUN cp llama.cpp/llama-server .
|
8 |
+
RUN rm -rf llama.cpp/
|
9 |
+
|
10 |
+
CMD ["sleep", " infinity"]
|