Text Generation
Transformers
Safetensors
English
Arabic
llama
text-generation-inference
4-bit precision
awq
MohamedRashad commited on
Commit
9db1926
1 Parent(s): c3fcc0e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -96,7 +96,7 @@ model_name_or_path = "MohamedRashad/AceGPT-7B-chat-AWQ"
96
  tokenizer = AutoTokenizer.from_pretrained(model_name_or_path, padding_side="right")
97
  model = AutoModelForCausalLM.from_pretrained(
98
  model_name_or_path,
99
- use_flash_attention_2=True,
100
  torch_dtype=torch.float16,
101
  low_cpu_mem_usage=True,
102
  device_map="auto"
 
96
  tokenizer = AutoTokenizer.from_pretrained(model_name_or_path, padding_side="right")
97
  model = AutoModelForCausalLM.from_pretrained(
98
  model_name_or_path,
99
+ use_flash_attention_2=True, # disable if you have problems with flash attention 2
100
  torch_dtype=torch.float16,
101
  low_cpu_mem_usage=True,
102
  device_map="auto"