T3Sam3D / app.py
spanofzero's picture
Update app.py
da70066 verified
import gradio as gr
from huggingface_hub import InferenceClient
from array import array
import os
import re
import time
# Securely retrieve the token from Space secrets
HF_TOKEN = os.getenv("HF_TOKEN")
# Initialize the inference client
client = InferenceClient("Qwen/Qwen2.5-7B-Instruct", token=HF_TOKEN)
class StateController:
__slots__ = ("_state", "_metric", "_batch", "_reg", "_rendered", "_rom60", "_symbols")
def __init__(self):
self._state = array("B", [0]) * 121
self._metric = 60
self._batch = 10
self._reg = {}
self._rendered = self._build_render()
# Pre-compute the 60x60 lookup ROM for O(1) verification
self._rom60 = tuple(
tuple((i * j) % 60 for j in range(60))
for i in range(60)
)
# 60-character alphanumeric array for cross-platform rendering
self._symbols = "ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz01234567"
def _build_render(self) -> str:
"""Construct the fixed 121-point visualization once during initialization."""
return "".join(
" [NODE_120] " if i == 120 else ("<" if i % 10 == 0 else ".")
for i in range(121)
)
def diagnostic(self) -> str:
"""Execute diagnostic sequence and return the formatted status report."""
for i in range(51):
self._state[i] = i % self._batch
self._reg.clear()
self._reg["STATUS"] = "RESOLVED"
return (
"Diagnostic sequence initiated.\n\n"
"Grid initialized: 5 active blocks.\n\n"
"Rendering 121-point array:\n"
f"{self._rendered}\n\n"
"Executing state resolution:\n"
"System resolved. State array reset to zero."
)
def generate_receipt(self, a: int, b: int, c: int) -> str:
"""Calculates the base-60 checksum for any 3-node allocation."""
rom_val = self._rom60[a % 60][b % 60]
checksum_index = (rom_val ^ (c % 60)) % 60
return f"0{self._symbols[checksum_index]}"
def validate_receipt(self, receipt: str, a: int, b: int, c: int) -> str:
"""Routes through NODE_120 to verify the checksum mathematically."""
expected_receipt = self.generate_receipt(a, b, c)
if receipt == expected_receipt:
self._state[120] = 1 # Activate NODE_120
return f"[NODE_120: ACTIVE] Checksum Validated. Receipt {receipt} matches allocation ({a}, {b}, {c})."
else:
self._state[120] = 0 # Deactivate NODE_120
return f"[NODE_120: INACTIVE] Verification Failed. Expected receipt {expected_receipt}, received {receipt}."
# Global singleton instance for resource reuse
controller = StateController()
SYSTEM_MSG = {
"role": "system",
"content": ("You are a logic-focused inference engine. "
"Utilize strict state-hold memory and parallel integer blocks. "
"Provide direct, technical, and accurate responses.")
}
def generate_response(message: str, history: list):
start_time = time.perf_counter()
msg_lower = message.lower().strip()
# Hardware diagnostic intercept
if msg_lower == "run grid diagnostic":
output = controller.diagnostic()
elapsed_time = time.perf_counter() - start_time
yield f"{output}\n\n---\n*Telemetry: Compute Time {elapsed_time:.4f}s | Source: Local Engine*"
return
# Deterministic Checksum Intercept (Bypasses AI completely)
verify_match = re.match(r"verify receipt\s+([a-zA-Z0-9]{2})\s+for\s+(\d+),\s*(\d+),\s*(\d+)", msg_lower, re.IGNORECASE)
if verify_match:
receipt = verify_match.group(1)
a = int(verify_match.group(2))
b = int(verify_match.group(3))
c = int(verify_match.group(4))
output = controller.validate_receipt(receipt, a, b, c)
elapsed_time = time.perf_counter() - start_time
yield f"{output}\n\n---\n*Telemetry: Compute Time {elapsed_time:.6f}s | Source: Local ROM Math*"
return
# Build the message list for inference
messages = [SYSTEM_MSG]
for turn in history:
messages.append({"role": "user", "content": turn[0]})
messages.append({"role": "assistant", "content": turn[1]})
messages.append({"role": "user", "content": message})
try:
# Stream text generation
stream = client.chat_completion(
messages,
max_tokens=1024,
stream=True,
temperature=0.15
)
partial_response = ""
for chunk in stream:
token = chunk.choices[0].delta.content or ""
partial_response += token
yield partial_response
# Post-processing telemetry after the stream completes
elapsed_time = time.perf_counter() - start_time
word_count = len(partial_response.split())
est_speed = word_count / elapsed_time if elapsed_time > 0 else 0
# Append telemetry footer to the final response
final_output = partial_response + f"\n\n---\n*Telemetry: Compute Time {elapsed_time:.2f}s | Est. Speed: {est_speed:.2f} words/sec | Source: Inference API*"
yield final_output
except Exception as exc:
yield f"System Error: {str(exc)}. Verify API token and permissions."
custom_css = """
body, .gradio-container { background-color: #0b0f19 !important; }
footer {display: none !important}
.message.user { background-color: #1e293b !important; border: 1px solid #3b82f6 !important; }
.message.bot { background-color: #0f172a !important; color: #60a5fa !important; }
"""
with gr.Blocks() as demo:
gr.Markdown("# Glyph.io Logic Interface")
gr.ChatInterface(
fn=generate_response,
description="Inference layer utilizing state-hold logic and deterministic ROM verification.",
examples=[
"Run grid diagnostic",
"Calculate the integer distribution for 120 units across 3 nodes.",
"Verify receipt 0e for 60, 30, 30"
],
cache_examples=False
)
if __name__ == "__main__":
demo.queue().launch(
theme=gr.themes.Soft(primary_hue="blue"),
css=custom_css,
ssr_mode=False
)