AtAndDev commited on
Commit
56c0ca0
·
verified ·
1 Parent(s): 0d2b193

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +1 -2
app.py CHANGED
@@ -7,7 +7,6 @@ import spaces
7
  import gradio as gr
8
  from transformers import AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig, TextIteratorStreamer
9
 
10
- subprocess.run('pip install transformers==4.45.2 torch==2.4.1', shell=True)
11
  subprocess.run('pip install --upgrade flash-attn --no-build-isolation', env={'FLASH_ATTENTION_SKIP_CUDA_BUILD': "TRUE"}, shell=True)
12
 
13
  MODEL_ID = "AtAndDev/marco-qwq-7B"
@@ -110,7 +109,7 @@ quantization_config = BitsAndBytesConfig(
110
  load_in_4bit=True,
111
  bnb_4bit_compute_dtype=torch.bfloat16
112
  )
113
- tokenizer = AutoTokenizer.from_pretrained(MODEL_ID, trust_remote_code=True)
114
  model = AutoModelForCausalLM.from_pretrained(
115
  MODEL_ID,
116
  device_map="auto",
 
7
  import gradio as gr
8
  from transformers import AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig, TextIteratorStreamer
9
 
 
10
  subprocess.run('pip install --upgrade flash-attn --no-build-isolation', env={'FLASH_ATTENTION_SKIP_CUDA_BUILD': "TRUE"}, shell=True)
11
 
12
  MODEL_ID = "AtAndDev/marco-qwq-7B"
 
109
  load_in_4bit=True,
110
  bnb_4bit_compute_dtype=torch.bfloat16
111
  )
112
+ tokenizer = AutoTokenizer.from_pretrained("AIDC-AI/Marco-o1", trust_remote_code=True)
113
  model = AutoModelForCausalLM.from_pretrained(
114
  MODEL_ID,
115
  device_map="auto",