dioarafl commited on
Commit
2ee9b1f
1 Parent(s): 780dfe3

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +11 -9
app.py CHANGED
@@ -1,23 +1,25 @@
1
  from transformers import AutoTokenizer, AutoModelForCausalLM
2
  import gradio as gr
3
 
4
- # Load model and tokenizer directly
5
- tokenizer = AutoTokenizer.from_pretrained("meta-llama/Meta-Llama-3-8B")
6
- model = AutoModelForCausalLM.from_pretrained("meta-llama/Meta-Llama-3-8B")
 
 
 
 
 
 
 
 
7
 
8
- # Define function to translate code
9
  def translate_code(input_code, prompt=""):
10
- # Combine input code and prompt
11
  input_text = f"{prompt}\n\n{input_code}"
12
- # Tokenize input text
13
  input_ids = tokenizer.encode(input_text, return_tensors="pt", max_length=1024, truncation=True)
14
- # Generate output sequence
15
  output = model.generate(input_ids, max_length=1024, num_return_sequences=1, temperature=0.7)
16
- # Decode output sequence
17
  translated_code = tokenizer.decode(output[0], skip_special_tokens=True)
18
  return translated_code
19
 
20
- # Launch Gradio interface
21
  gr.Interface(
22
  fn=translate_code,
23
  inputs=["text", "text"],
 
1
  from transformers import AutoTokenizer, AutoModelForCausalLM
2
  import gradio as gr
3
 
4
+ import os
5
+
6
+
7
+ huggingface_token = os.getenv("HUGGINGFACE_TOKEN")
8
+
9
+ if huggingface_token is None:
10
+ print("Token Hugging Face tidak ditemukan. Pastikan Anda telah menetapkan variabel lingkungan HUGGINGFACE_TOKEN.")
11
+ exit()
12
+
13
+ tokenizer = AutoTokenizer.from_pretrained("meta-llama/Meta-Llama-3-8B", token=huggingface_token)
14
+ model = AutoModelForCausalLM.from_pretrained("meta-llama/Meta-Llama-3-8B", token=huggingface_token)
15
 
 
16
  def translate_code(input_code, prompt=""):
 
17
  input_text = f"{prompt}\n\n{input_code}"
 
18
  input_ids = tokenizer.encode(input_text, return_tensors="pt", max_length=1024, truncation=True)
 
19
  output = model.generate(input_ids, max_length=1024, num_return_sequences=1, temperature=0.7)
 
20
  translated_code = tokenizer.decode(output[0], skip_special_tokens=True)
21
  return translated_code
22
 
 
23
  gr.Interface(
24
  fn=translate_code,
25
  inputs=["text", "text"],