Joel-Marin commited on
Commit
9045974
·
verified ·
1 Parent(s): 109af35

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +11 -18
app.py CHANGED
@@ -1,24 +1,17 @@
1
  import gradio as gr
2
- import torch
3
- from transformers import AutoModelForCausalLM, AutoTokenizer, TextGenerationPipeline
4
 
5
- MODEL_NAME = "deepseek-ai/deepseek-llm-7b-chat" # Change for LLaMA if needed
 
6
 
7
- # Load model & tokenizer
8
- tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)
9
- model = AutoModelForCausalLM.from_pretrained(
10
- MODEL_NAME,
11
- torch_dtype=torch.float16,
12
- device_map="auto" # Uses available GPU/CPU
13
- )
14
-
15
- # Create text generation pipeline
16
- pipe = TextGenerationPipeline(model=model, tokenizer=tokenizer)
17
-
18
- # Define Gradio interface
19
  def chat(prompt):
20
- output = pipe(prompt, max_length=512, do_sample=True, temperature=0.7)
21
- return output[0]["generated_text"]
 
 
 
 
22
 
23
- iface = gr.Interface(fn=chat, inputs="text", outputs="text", title="DeepSeek 7B Chatbot")
24
  iface.launch()
 
1
  import gradio as gr
2
+ import requests
3
+ import os
4
 
5
+ HF_API_KEY = os.getenv('TOKEN')
6
+ MODEL_NAME = "deepseek-ai/deepseek-llm-7b-chat"
7
 
 
 
 
 
 
 
 
 
 
 
 
 
8
  def chat(prompt):
9
+ response = requests.post(
10
+ f"https://api-inference.huggingface.co/models/{MODEL_NAME}",
11
+ headers={"Authorization": f"Bearer {HF_API_KEY}"},
12
+ json={"inputs": prompt}
13
+ )
14
+ return response.json()[0]["generated_text"]
15
 
16
+ iface = gr.Interface(fn=chat, inputs="text", outputs="text", title="DeepSeek API Chatbot")
17
  iface.launch()