| | <!DOCTYPE html> |
| | <html lang="en"> |
| | <head> |
| | <meta charset="UTF-8"> |
| | <meta name="viewport" content="width=device-width, initial-scale=1.0"> |
| | <title>GLM-4.7 Flash Deployment Hub</title> |
| | <meta name="description" content="Deployment interface for GLM-4.7-Flash-Uncensored model"> |
| | |
| | |
| | <script src="https://cdn.tailwindcss.com"></script> |
| | |
| | |
| | <link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/font-awesome/6.4.0/css/all.min.css"> |
| | |
| | |
| | <link href="https://fonts.googleapis.com/css2?family=Inter:wght@300;400;600;700&family=Fira+Code:wght@400;600&display=swap" rel="stylesheet"> |
| |
|
| | <style> |
| | :root { |
| | --primary-color: #6366f1; |
| | --accent-color: #8b5cf6; |
| | --dark-bg: #0f172a; |
| | --panel-bg: #1e293b; |
| | --text-color: #e2e8f0; |
| | --success-color: #10b981; |
| | --terminal-font: 'Fira Code', monospace; |
| | } |
| | |
| | body { |
| | background-color: var(--dark-bg); |
| | color: var(--text-color); |
| | font-family: 'Inter', sans-serif; |
| | line-height: 1.6; |
| | } |
| | |
| | .glass-panel { |
| | background: rgba(30, 41, 59, 0.7); |
| | backdrop-filter: blur(10px); |
| | border: 1px solid rgba(255, 255, 255, 0.1); |
| | box-shadow: 0 4px 6px -1px rgba(0, 0, 0, 0.5); |
| | } |
| | |
| | .text-gradient { |
| | background: linear-gradient(to right, #818cf8, #c084fc); |
| | -webkit-background-clip: text; |
| | -webkit-text-fill-color: transparent; |
| | } |
| | |
| | .code-block { |
| | font-family: var(--terminal-font); |
| | background-color: #000; |
| | border-radius: 8px; |
| | padding: 1rem; |
| | position: relative; |
| | overflow-x: auto; |
| | } |
| | |
| | .code-header { |
| | background-color: #2d3748; |
| | padding: 0.5rem 1rem; |
| | border-radius: 8px 8px 0 0; |
| | font-size: 0.85rem; |
| | color: #a0aec0; |
| | display: flex; |
| | justify-content: space-between; |
| | align-items: center; |
| | } |
| | |
| | .terminal-window { |
| | background-color: #0d1117; |
| | border: 1px solid #30363d; |
| | border-radius: 12px; |
| | overflow: hidden; |
| | box-shadow: 0 10px 15px -3px rgba(0, 0, 0, 0.7); |
| | } |
| | |
| | .terminal-header { |
| | background-color: #161b22; |
| | padding: 0.75rem; |
| | border-bottom: 1px solid #30363d; |
| | display: flex; |
| | gap: 8px; |
| | } |
| | |
| | .dot { width: 12px; height: 12px; border-radius: 50%; } |
| | .red { background-color: #ff5f56; } |
| | .yellow { background-color: #ffbd2e; } |
| | .green { background-color: #27c93f; } |
| | |
| | .terminal-body { |
| | padding: 1rem; |
| | font-family: var(--terminal-font); |
| | font-size: 0.9rem; |
| | color: #c9d1d9; |
| | min-height: 200px; |
| | } |
| | |
| | .cmd-text { color: #7ee787; } |
| | .path-text { color: #58a6ff; } |
| | .output-text { color: #d2a8ff; } |
| | .error-text { color: #ff7b72; } |
| | .success-text { color: #3fb950; } |
| | |
| | .loader { |
| | border: 4px solid #f3f3f3; |
| | border-top: 4px solid var(--primary-color); |
| | border-radius: 50%; |
| | width: 24px; |
| | height: 24px; |
| | animation: spin 1s linear infinite; |
| | } |
| | |
| | @keyframes spin { |
| | 0% { transform: rotate(0deg); } |
| | 100% { transform: rotate(360deg); } |
| | } |
| | |
| | .feature-card:hover { |
| | transform: translateY(-5px); |
| | transition: all 0.3s ease; |
| | } |
| | </style> |
| | </head> |
| | <body class="min-h-screen flex flex-col"> |
| |
|
| | |
| | <nav class="glass-panel sticky top-0 z-50 border-b border-indigo-500/30"> |
| | <div class="max-w-7xl mx-auto px-4 sm:px-6 lg:px-8"> |
| | <div class="flex items-center justify-between h-16"> |
| | <div class="flex items-center gap-3"> |
| | <i class="fa-solid fa-robot text-2xl text-indigo-400"></i> |
| | <div> |
| | <h1 class="text-xl font-bold tracking-tight"> |
| | GLM-4.7 <span class="text-indigo-400">Flash</span> Deployer |
| | </h1> |
| | <p class="text-xs text-gray-400">Uncensored Code & Chat Interface</p> |
| | </div> |
| | </div> |
| | |
| | <div class="hidden md:flex items-center gap-6"> |
| | <a href="#dashboard" class="text-sm font-medium text-gray-300 hover:text-white transition">Dashboard</a> |
| | <a href="#install" class="text-sm font-medium text-gray-300 hover:text-white transition">Installation</a> |
| | <a href="#cloud" class="text-sm font-medium text-gray-300 hover:text-white transition">Cloud Options</a> |
| | <a href="https://huggingface.co/DavidAU/GLM-4.7-Flash-Uncensored-Heretic-NEO-CODE-Imatrix-MAX-GGUF" target="_blank" class="text-sm font-medium text-indigo-400 hover:text-indigo-300 flex items-center gap-2"> |
| | <i class="fa-brands fa-huggingface"></i> Model Hub |
| | </a> |
| | </div> |
| |
|
| | |
| | <a href="https://huggingface.co/spaces/akhaliq/anycoder" target="_blank" class="text-xs font-semibold text-gray-500 hover:text-indigo-400 transition flex items-center gap-1"> |
| | <i class="fa-solid fa-link"></i> Built with anycoder |
| | </a> |
| | </div> |
| | </div> |
| | </nav> |
| |
|
| | |
| | <main class="flex-grow container mx-auto px-4 py-8 space-y-12"> |
| |
|
| | |
| | <section id="dashboard" class="text-center space-y-6 py-10"> |
| | <div class="inline-block px-4 py-1 rounded-full bg-indigo-900/30 border border-indigo-500/30 text-indigo-300 text-sm font-medium mb-4"> |
| | Model: GLM-4.7-Flash-Uncensored-Heretic-NEO-CODE |
| | </div> |
| | <h2 class="text-4xl md:text-6xl font-extrabold tracking-tight leading-tight"> |
| | Run the <span class="text-gradient">Ultimate</span> Local AI |
| | </h2> |
| | <p class="text-xl text-gray-400 max-w-3xl mx-auto"> |
| | A comprehensive dashboard to install, configure, and deploy the GLM-4.7 Flash model. |
| | Optimized for GGUF quantization to run efficiently on consumer hardware or cloud instances. |
| | </p> |
| |
|
| | <div class="flex flex-wrap justify-center gap-4 pt-4"> |
| | <a href="#install" class="px-8 py-3 rounded-lg bg-indigo-600 hover:bg-indigo-700 text-white font-bold shadow-lg shadow-indigo-500/30 transition transform hover:scale-105 flex items-center gap-2"> |
| | <i class="fa-solid fa-download"></i> Start Installation |
| | </a> |
| | <a href="#cloud" class="px-8 py-3 rounded-lg bg-gray-800 hover:bg-gray-700 text-white font-bold border border-gray-600 transition transform hover:scale-105 flex items-center gap-2"> |
| | <i class="fa-solid fa-cloud"></i> Cloud Deployment |
| | </a> |
| | </div> |
| | </section> |
| |
|
| | |
| | <section class="grid md:grid-cols-3 gap-6"> |
| | <div class="glass-panel p-6 rounded-xl feature-card"> |
| | <div class="w-12 h-12 rounded-lg bg-blue-900/50 flex items-center justify-center mb-4 text-blue-400"> |
| | <i class="fa-solid fa-memory text-2xl"></i> |
| | </div> |
| | <h3 class="text-lg font-bold mb-2">RAM Requirements</h3> |
| | <p class="text-sm text-gray-400"> |
| | The <strong>Imatrix-MAX</strong> version is large. We recommend at least <span class="text-green-400 font-bold">32GB RAM</span> for smooth inference without swapping. |
| | </p> |
| | </div> |
| | <div class="glass-panel p-6 rounded-xl feature-card"> |
| | <div class="w-12 h-12 rounded-lg bg-purple-900/50 flex items-center justify-center mb-4 text-purple-400"> |
| | <i class="fa-solid fa-microchip text-2xl"></i> |
| | </div> |
| | <h3 class="text-lg font-bold mb-2">GPU Recommendation</h3> |
| | <p class="text-sm text-gray-400"> |
| | An NVIDIA GPU with <span class="text-green-400 font-bold">8GB+ VRAM</span> (RTX 3060 or better) allows for acceleration via CUDA. |
| | </p> |
| | </div> |
| | <div class="glass-panel p-6 rounded-xl feature-card"> |
| | <div class="w-12 h-12 rounded-lg bg-emerald-900/50 flex items-center justify-center mb-4 text-emerald-400"> |
| | <i class="fa-solid fa-hard-drive text-2xl"></i> |
| | </div> |
| | <h3 class="text-lg font-bold mb-2">Storage</h3> |
| | <p class="text-sm text-gray-400"> |
| | The GGUF model typically ranges between <span class="text-green-400 font-bold">20GB - 40GB</span> depending on the specific quantization (Q4_K_M vs Q8_0). |
| | </p> |
| | </div> |
| | </section> |
| |
|
| | |
| | <section id="install" class="glass-panel rounded-2xl overflow-hidden"> |
| | <div class="p-6 border-b border-gray-700 bg-gray-900/50 flex justify-between items-center"> |
| | <div> |
| | <h3 class="text-xl font-bold flex items-center gap-2"> |
| | <i class="fa-solid fa-code"></i> Automated Setup Script |
| | </h3> |
| | <p class="text-sm text-gray-400">Generates the command line instructions for KoboldCpp & Ollama</p> |
| | </div> |
| | <button onclick="copyToClipboard()" id="copyBtn" class="px-4 py-2 rounded bg-gray-700 hover:bg-gray-600 text-white text-sm transition flex items-center gap-2"> |
| | <i class="fa-regular fa-copy"></i> Copy to Clipboard |
| | </button> |
| | </div> |
| | |
| | <div class="p-8 grid lg:grid-cols-2 gap-8"> |
| | |
| | <div class="space-y-6"> |
| | <h4 class="text-lg font-semibold text-indigo-400 border-b border-gray-700 pb-2">Configuration</h4> |
| | |
| | <div> |
| | <label class="block text-sm font-medium text-gray-300 mb-2">Execution Backend</label> |
| | <select id="backendSelect" class="w-full bg-gray-800 border border-gray-700 rounded-lg p-3 text-white focus:ring-2 focus:ring-indigo-500 focus:outline-none"> |
| | <option value="kobold">KoboldCpp (Local Web UI)</option> |
| | <option value="ollama">Ollama (System Service)</option> |
| | <option value="cmd">Direct CLI (llama.cpp)</option> |
| | </select> |
| | </div> |
| |
|
| | <div> |
| | <label class="block text-sm font-medium text-gray-300 mb-2">GPU Acceleration</label> |
| | <div class="flex gap-4"> |
| | <label class="flex items-center gap-2 text-gray-300 cursor-pointer"> |
| | <input type="radio" name="gpu" value="cpu" checked class="text-indigo-600 focus:ring-indigo-500"> |
| | <span>CPU Only (Slower)</span> |
| | </label> |
| | <label class="flex items-center gap-2 text-gray-300 cursor-pointer"> |
| | <input type="radio" name="gpu" value="cuda" class="text-indigo-600 focus:ring-indigo-500"> |
| | <span>NVIDIA (CUDA)</span> |
| | </label> |
| | <label class="flex items-center gap-2 text-gray-300 cursor-pointer"> |
| | <input type="radio" name="gpu" value="metal" class="text-indigo-600 focus:ring-indigo-500"> |
| | <span>Apple Silicon (Metal)</span> |
| | </label> |
| | </div> |
| | </div> |
| |
|
| | <div> |
| | <label class="block text-sm font-medium text-gray-300 mb-2">Threads (CPU Cores)</label> |
| | <input type="range" id="threadSlider" min="1" max="32" value="8" class="w-full h-2 bg-gray-700 rounded-lg appearance-none cursor-pointer accent-indigo-500"> |
| | <div class="flex justify-between text-xs text-gray-500 mt-1"> |
| | <span>1 Thread</span> |
| | <span id="threadValue" class="text-indigo-400 font-bold">8 Threads</span> |
| | <span>32 Threads</span> |
| | </div> |
| | </div> |
| |
|
| | <div class="bg-indigo-900/20 p-4 rounded-lg border border-indigo-500/20"> |
| | <h5 class="text-sm font-bold text-indigo-300 mb-2"><i class="fa-solid fa-circle-info"></i> Note on Imatrix-MAX</h5> |
| | <p class="text-xs text-gray-400"> |
| | The "Imatrix-MAX" version implies an optimized quantization matrix. Ensure you download the specific <code>.gguf</code> file from the HuggingFace link provided. The script below assumes standard GGUF loading. |
| | </p> |
| | </div> |
| | </div> |
| |
|
| | |
| | <div class="relative"> |
| | <div class="terminal-window"> |
| | <div class="terminal-header"> |
| | <div class="dot red"></div> |
| | <div class="dot yellow"></div> |
| | <div class="dot green"></div> |
| | <span class="text-xs text-gray-400 ml-2">root@server:~/glm-deploy</span> |
| | </div> |
| | <div class="terminal-body" id="codeOutput"> |
| | <div class="mb-2"> |
| | <span class="path-text">~</span> |
| | <span class="text-white"> $ </span> |
| | <span class="cmd-text"># Initializing GLM-4.7 Flash Setup...</span> |
| | </div> |
| | <div id="scriptContent"> |
| | <div class="mb-2"> |
| | <span class="path-text">~</span> |
| | <span class="text-white"> $ </span> |
| | <span class="cmd-text">git clone https://github.com/ggerganov/llama.cpp</span> |
| | </div> |
| | <div class="mb-2"> |
| | <span class="path-text">~</span> |
| | <span class="text-white"> $ </span> |
| | <span class="cmd-text">cd llama.cpp && make</span> |
| | </div> |
| | <div class="mb-2"> |
| | <span class="path-text">~/llama.cpp</span> |
| | <span class="text-white"> $ </span> |
| | <span class="cmd-text">./llama-server</span> |
| | <span class="text-gray-500"> \</span> |
| | </div> |
| | <div class="mb-2 ml-4"> |
| | <span class="text-gray-500"> --model ./models/glm-4.7-flash-uncensored.Q4_K_M.gguf</span> |
| | </div> |
| | <div class="mb-2 ml-4"> |
| | <span class="text-gray-500"> --n_ctx 4096 --n_gpu_layers 35</span> |
| | </div> |
| | <div class="mb-2 ml-4"> |
| | <span class="text-gray-500"> --port 8080 --host 0.0.0.0</span> |
| | </div> |
| | <div class="mt-4"> |
| | <span class="success-text">[INFO] Server starting on http://localhost:8080</span> |
| | </div> |
| | </div> |
| | </div> |
| | </div> |
| | <div class="mt-4 text-xs text-gray-500 text-center"> |
| | * Adjust <code>--n_gpu_layers</code> based on your VRAM capacity. |
| | </div> |
| | </div> |
| | </div> |
| | </section> |
| |
|
| | |
| | <section id="cloud" class="space-y-8"> |
| | <div class="text-center"> |
| | <h2 class="text-3xl font-bold">Cloud Deployment Options</h2> |
| | <p class="text-gray-400 mt-2">Running this model in the cloud requires instances with high RAM and VRAM.</p> |
| | </div> |
| |
|
| | <div class="grid md:grid-cols-3 gap-6"> |
| | |
| | <div class="glass-panel p-6 rounded-xl border-t-4 border-pink-500 flex flex-col justify-between hover:border-pink-400 transition"> |
| | <div> |
| | <div class="w-10 h-10 rounded bg-pink-900/30 flex items-center justify-center text-pink-400 mb-4"> |
| | <i class="fa-solid fa-cloud-bolt"></i> |
| | </div> |
| | <h3 class="text-xl font-bold mb-2">RunPod</h3> |
| | <p class="text-sm text-gray-400 mb-4"> |
| | Best for short bursts. Use an A100 or H100 pod. Upload your GGUF file to Pod Storage and run the server. |
| | </p> |
| | <div class="bg-gray-900/50 p-3 rounded text-xs font-mono text-pink-300 mb-4"> |
| | nvidia-a100-80gb |
| | </div> |
| | <div class="flex items-center gap-2 text-xs text-gray-500"> |
| | <i class="fa-solid fa-check text-green-500"></i> Pay per second |
| | </div> |
| | </div> |
| | <a href="https://runpod.io" target="_blank" class="mt-4 w-full text-center py-2 rounded bg-pink-600 hover:bg-pink-700 text-white text-sm font-medium transition"> |
| | Launch Pod |
| | </a> |
| | </div> |
| |
|
| | |
| | <div class="glass-panel p-6 rounded-xl border-t-4 border-orange-500 flex flex-col justify-between hover:border-orange-400 transition"> |
| | <div> |
| | <div class="w-10 h-10 rounded bg-orange-900/30 flex items-center justify-center text-orange-400 mb-4"> |
| | <i class="fa-solid fa-server"></i> |
| | </div> |
| | <h3 class="text-xl font-bold mb-2">Vast.ai</h3> |
| | <p class="text-sm text-gray-400 mb-4"> |
| | Marketplaces for GPU instances. Look for "RTX 4090" or "A100 80GB" instances. Very cost-effective. |
| | </p> |
| | <div class="bg-gray-900/50 p-3 rounded text-xs font-mono text-orange-300 mb-4"> |
| | 96GB RAM + 24GB VRAM |
| | </div> |
| | <div class="flex items-center gap-2 text-xs text-gray-500"> |
| | <i class="fa-solid fa-check text-green-500"></i> Bid/On-demand options |
| | </div> |
| | </div> |
| | <a href="https://vast.ai" target="_blank" class="mt-4 w-full text-center py-2 rounded bg-orange-600 hover:bg-orange-700 text-white text-sm font-medium transition"> |
| | Browse Instances |
| | </a> |
| | </div> |
| |
|
| | |
| | <div class="glass-panel p-6 rounded-xl border-t-4 border-blue-500 flex flex-col justify-between hover:border-blue-400 transition"> |
| | <div> |
| | <div class="w-10 h-10 rounded bg-blue-900/30 flex items-center justify-center text-blue-400 mb-4"> |
| | <i class="fa-solid faRocket"></i> |
| | </div> |
| | <h3 class="text-xl font-bold mb-2">Lambda Labs</h3> |
| | <p class="text-sm text-gray-400 mb-4"> |
| | User-friendly interface. Good for A10s and H100s. Easy to set up SSH and transfer models. |
| | </p> |
| | <div class="bg-gray-900/50 p-3 rounded text-xs font-mono text-blue-300 mb-4"> |
| | H100 80GB or A100 40/80GB |
| | </div> |
| | <div class="flex items-center gap-2 text-xs text-gray-500"> |
| | <i class="fa-solid fa-check text-green-500"></i> Flat rate pricing |
| | </div> |
| | </div> |
| | <a href="https://lambdalabs.com" target="_blank" class="mt-4 w-full text-center py-2 rounded bg-blue-600 hover:bg-blue-700 text-white text-sm font-medium transition"> |
| | Get API Key |
| | </a> |
| | </div> |
| | </div> |
| |
|
| | <div class="glass-panel p-6 rounded-xl bg-gradient-to-r from-gray-900 to-gray-800 border border-gray-700"> |
| | <h3 class="text-lg font-bold text-white mb-4"><i class="fa-solid fa-list-check"></i> Quick Cloud Setup Checklist</h3> |
| | <ul class="space-y-3"> |
| | <li class="flex items-start gap-3"> |
| | <div class="mt-1 w-5 h-5 rounded-full bg-green-900 flex items-center justify-center text-green-400 text-xs"><i class="fa-solid fa-check"></i></div> |
| | <div> |
| | <span class="font-bold text-gray-200">Select Instance:</span> |
| | <span class="text-gray-400 text-sm"> Ensure instance has > 32GB System RAM AND > 16GB VRAM for the Imatrix-MAX version.</span> |
| | </div> |
| | </li> |
| | <li class="flex items-start gap-3"> |
| | <div class="mt-1 w-5 h-5 rounded-full bg-green-900 flex items-center justify-center text-green-400 text-xs"><i class="fa-solid fa-check"></i></div> |
| | <div> |
| | <span class="font-bold text-gray-200">Download Model:</span> |
| | <span class="text-gray-400 text-sm"> Use `wget` or `huggingface-cli` on the cloud instance to download the GGUF file directly to storage.</span> |
| | </div> |
| | </li> |
| | <li class="flex items-start gap-3"> |
| | <div class="mt-1 w-5 h-5 rounded-full bg-green-900 flex items-center justify-center text-green-400 text-xs"><i class="fa-solid fa-check"></i></div> |
| | <div> |
| | <span class="font-bold text-gray-200">Start Server:</span> |
| | <span class="text-gray-400 text-sm"> Run the `llama-server` command with `--host 0.0.0.0` to allow external web access.</span> |
| | </div> |
| | </li> |
| | </ul> |
| | </div> |
| | </section> |
| |
|
| | |
| | <footer class="border-t border-gray-800 pt-8 pb-4 text-center"> |
| | <div class="flex items-center justify-center gap-2 mb-4 opacity-70"> |
| | <i class="fa-solid fa-robot text-indigo-500"></i> |
| | <span class="font-semibold text-gray-400">GLM-4.7 Flash Deployer</span> |
| | </div> |
| | <p class="text-sm text-gray-600"> |
| | This tool is a UI helper for deploying open-source AI models. |
| | Please ensure you comply with the model's license agreement. |
| | </p> |
| | <div class="mt-4 text-xs text-gray-700"> |
| | © <span id="year"></span> AI Deployment Tools. All rights reserved. |
| | </div> |
| | </footer> |
| |
|
| | </main> |
| |
|
| | |
| | <script> |
| | // Update Year |
| | document.getElementById('year').textContent = new Date().getFullYear(); |
| | |
| | // Thread Slider Logic |
| | const slider = document.getElementById('threadSlider'); |
| | const threadValue = document.getElementById('threadValue'); |
| | |
| | slider.addEventListener('input', function() { |
| | threadValue.textContent = this.value + " Threads"; |
| | updateScript(); |
| | }); |
| | |
| | // Backend Selection Logic |
| | const backendSelect = document.getElementById('backendSelect'); |
| | |
| | backendSelect.addEventListener('change', updateScript); |
| | |
| | // GPU Selection Logic |
| | const gpuRadios = document.querySelectorAll('input[name="gpu"]'); |
| | gpuRadios.forEach(radio => { |
| | radio.addEventListener('change', updateScript); |
| | }); |
| | |
| | function updateScript() { |
| | const backend = backendSelect.value; |
| | const threads = slider.value; |
| | const gpu = document.querySelector('input[name="gpu"]:checked').value; |
| | const outputDiv = document.getElementById('scriptContent'); |
| | |
| | let scriptHTML = ''; |
| | |
| | if (backend === 'kobold') { |
| | scriptHTML = ` |
| | <div class="mb-2"> |
| | <span class="path-text">~</span> |
| | <span class="text-white"> $ </span> |
| | <span class="cmd-text">git clone https://github.com/LostRuins/koboldcpp.git</span> |
| | </div> |
| | <div class="mb-2"> |
| | <span class="path-text">~</span> |
| | <span class="text-white"> $ </span> |
| | <span class="cmd-text">cd koboldcpp</span> |
| | </div> |
| | <div class="mb-2"> |
| | <span class="path-text">~/koboldcpp</span> |
| | <span class="text-white"> $ </span> |
| | <span class="cmd-text">mkdir -p models</span> |
| | </div> |
| | <div class="mb-2"> |
| | <span class="path-text">~/koboldcpp/models</span> |
| | <span class="text-white"> $ </span> |
| | <span class="cmd-text"># Place your GLM-4.7-Flash-Q4_K_M.gguf here</span> |
| | </div> |
| | <div class="mb-2"> |
| | <span class="path-text">~/koboldcpp</span> |
| | <span class="text-white"> $ </span> |
| | <span class="cmd-text">python3 koboldcpp.py</span> |
| | <span class="text-gray-500"> \</span> |
| | </div> |
| | <div class="mb-2 ml-4"> |
| | <span class="text-gray-500"> --model models/glm-4.7-flash-uncensored.Q4_K_M.gguf</span> |
| | </div> |
| | <div class="mb-2 ml-4"> |
| | <span class="text-gray-500"> --contextsize ${threads * 512} --threads ${threads}</span> |
| | </div> |
| | <div class="mb-2 ml-4"> |
| | <span class="text-gray-500"> --port 5001 --host 0.0.0.0</span> |
| | </div> |
| | <div class="mt-4"> |
| | <span class="success-text">[INFO] KoboldAI UI ready at http://localhost:5001</span> |
| | </div> |
| | `; |
| | } |
| | else if (backend === 'ollama') { |
| | scriptHTML = ` |
| | <div class="mb-2"> |
| | <span class="path-text">~</span> |
| | <span class="text-white"> $ </span> |
| | <span class="cmd-text">curl -fsSL https://ollama.com/install.sh | sh</span> |
| | </div> |
| | <div class="mb-2"> |
| | <span class="path-text">~</span> |
| | <span class="text-white"> $ </span> |
| | <span class="cmd-text">ollama pull nomic-ai/gpt4all-falcon</span> |
| | <span class="text-gray-500">/* Fallback if direct GLM pull fails */</span> |
| | </div> |
| | <div class="mb-2"> |
| | <span class="path-text">~</span> |
| | <span class="text-white"> $ </span> |
| | <span class="cmd-text">ollama create glm-4-flash -f Modelfile</span> |
| | </div> |
| | <div class="mb-2 bg-gray-800 p-3 rounded border-l-4 border-yellow-500"> |
| | <span class="text-xs text-gray-400 block mb-1"># Create a Modelfile:</span> |
| | <span class="text-xs text-white">FROM glm-4.7-flash-uncensored.Q4_K_M.gguf</span> |
| | <span class="text-xs text-white">PARAMETER num_thread ${threads}</span> |
| | </div> |
| | <div class="mt-4"> |
| | <span class="success-text">[SUCCESS] Model loaded into Ollama. Run 'ollama run glm-4-flash'</span> |
| | </div> |
| | `; |
| | } |
| | else { |
| | // Direct CLI |
| | scriptHTML = ` |
| | <div class="mb-2"> |
| | <span class="path-text">~</span> |
| | <span class="text-white"> $ </span> |
| | <span class="cmd-text">git clone https://github.com/ggerganov/llama.cpp</span> |
| | </div> |
| | <div class="mb-2"> |
| | <span class="path-text">~</span> |
| | <span class="text-white"> $ </span> |
| | <span class="cmd-text">cd llama.cpp && make</span> |
| | </div> |
| | <div class="mb-2"> |
| | <span class="path-text">~/llama.cpp</span> |
| | <span class="text-white"> $ </span> |
| | <span class="cmd-text">mkdir -p models</span> |
| | </div> |
| | <div class="mb-2"> |
| | <span class="path-text">~/llama.cpp/models</span> |
| | <span class="text-white"> $ </span> |
| | <span class="cmd-text"># Download GLM-4.7-Flash-Q4_K_M.gguf here</span> |
| | </div> |
| | <div class="mb-2"> |
| | <span class="path-text">~/llama.cpp</span> |
| | <span class="text-white"> $ </span> |
| | <span class="cmd-text">./main</span> |
| | <span class="text-gray-500"> \</span> |
| | </div> |
| | <div class="mb-2 ml-4"> |
| | <span class="text-gray-500"> -m models/glm-4.7-flash-uncensored.Q4_K_M.gguf</span> |
| | </div> |
| | <div class="mb-2 ml-4"> |
| | <span class="text-gray-500"> -n 512 --repeat_penalty 1.1</span> |
| | </div> |
| | <div class="mb-2 ml-4"> |
| | <span class="text-gray-500"> -c ${threads * 512} -t ${threads} -b 1</span> |
| | </div> |
| | <div class="mt-4"> |
| | <span class="success-text">[INFO] Entering interactive mode. Type 'exit' to quit.</span> |
| | </div> |
| | `; |
| | } |
| | |
| | // GPU Specific Flags |
| | if (gpu === 'cuda') { |
| | // Insert GPU flag logic roughly by appending to relevant lines or adding a new block |
| | // For simplicity in this template, we just update the text in the script |
| | const tempDiv = document.createElement('div'); |
| | tempDiv.innerHTML = scriptHTML; |
| | // Simple replacement logic based on backend would be complex, so we stick to generic output |
| | // In a real app, we'd inject --n_gpu_layers based on backend |
| | if (backend === 'kobold') { |
| | scriptHTML = scriptHTML.replace('--port 5001', '--port 5001 --gpu_layer_count 35'); |
| | } else if (backend === 'cmd') { |
| | scriptHTML = scriptHTML.replace('-t ' + threads, `-t ${threads} -ngl 35`); |
| | } |
| | } |
| | |
| | outputDiv.innerHTML = scriptHTML; |
| | } |
| | |
| | function copyToClipboard() { |
| | const codeElement = document.getElementById('scriptContent'); |
| | const textToCopy = codeElement.innerText; |
| | |
| | navigator.clipboard.writeText(textToCopy).then(() => { |
| | const btn = document.getElementById('copyBtn'); |
| | const originalText = btn.innerHTML; |
| | btn.innerHTML = '<i class="fa-solid fa-check"></i> Copied!'; |
| | btn.classList.remove('bg-gray-700'); |
| | btn.classList.add('bg-green-600'); |
| | |
| | setTimeout(() => { |
| | btn.innerHTML = originalText; |
| | btn.classList.add('bg-gray-700'); |
| | btn.classList.remove('bg-green-600'); |
| | }, 3000); |
| | }).catch(err => { |
| | console.error('Failed to copy: ', err); |
| | }); |
| | } |
| | </script> |
| | </body> |
| | </html> |