Spaces:
Runtime error
Runtime error
import gradio as gr | |
from llama_cpp import Llama | |
import os | |
# Path to the first shard of the model | |
model_path = "DeepSeek-R1-Zero-Q4_K_M/DeepSeek-R1-Zero-Q4_K_M-00001-of-00009.gguf" | |
# Debugging: Verify working directory and model path | |
print("Current working directory:", os.getcwd()) | |
print("Full model path:", os.path.join(os.getcwd(), model_path)) | |
# Initialize the model | |
try: | |
model = Llama(model_path=model_path, n_threads=8) | |
except ValueError as e: | |
print(f"Error initializing the model: {e}") | |
exit(1) | |
# Define the prediction function | |
def predict(prompt): | |
try: | |
# Generate output using the model | |
output = model(prompt) | |
# Extract and return the text from the response | |
return output["choices"][0]["text"] | |
except Exception as e: | |
return f"Error during inference: {e}" | |
# Create the Gradio interface | |
iface = gr.Interface( | |
fn=predict, | |
inputs="text", | |
outputs="text", | |
title="DeepSeek-R1-Zero", | |
description="A Gradio interface for the DeepSeek-R1-Zero model" | |
) | |
if __name__ == "__main__": | |
iface.launch() | |