Commit
·
22f789f
1
Parent(s):
da7b55e
Update README.md
Browse files
README.md
CHANGED
@@ -144,7 +144,6 @@ model = AutoModelForCausalLM.from_pretrained(
|
|
144 |
"azale-ai/DukunLM-7B-V1.0-Uncensored-sharded",
|
145 |
load_in_4bit=True,
|
146 |
torch_dtype=torch.float32,
|
147 |
-
trust_remote_code=True,
|
148 |
quantization_config=BitsAndBytesConfig(
|
149 |
load_in_4bit=True,
|
150 |
llm_int8_threshold=6.0,
|
@@ -204,7 +203,6 @@ model = AutoModelForCausalLM.from_pretrained(
|
|
204 |
"azale-ai/DukunLM-7B-V1.0-Uncensored-sharded",
|
205 |
load_in_4bit=True,
|
206 |
torch_dtype=torch.float32,
|
207 |
-
trust_remote_code=True,
|
208 |
quantization_config=BitsAndBytesConfig(
|
209 |
load_in_4bit=True,
|
210 |
llm_int8_threshold=6.0,
|
|
|
144 |
"azale-ai/DukunLM-7B-V1.0-Uncensored-sharded",
|
145 |
load_in_4bit=True,
|
146 |
torch_dtype=torch.float32,
|
|
|
147 |
quantization_config=BitsAndBytesConfig(
|
148 |
load_in_4bit=True,
|
149 |
llm_int8_threshold=6.0,
|
|
|
203 |
"azale-ai/DukunLM-7B-V1.0-Uncensored-sharded",
|
204 |
load_in_4bit=True,
|
205 |
torch_dtype=torch.float32,
|
|
|
206 |
quantization_config=BitsAndBytesConfig(
|
207 |
load_in_4bit=True,
|
208 |
llm_int8_threshold=6.0,
|