Spaces:
Runtime error
Runtime error
Antoine Chaffin
commited on
Commit
•
995dea4
1
Parent(s):
19b6ec4
Adding padding token
Browse files
app.py
CHANGED
@@ -32,6 +32,7 @@ If a question does not make any sense, or is not factually coherent, explain why
|
|
32 |
model = AutoModelForCausalLM.from_pretrained(args.model, use_auth_token=hf_token, torch_dtype=torch.float16,
|
33 |
device_map='auto').to(device)
|
34 |
tokenizer = AutoTokenizer.from_pretrained(args.model, use_auth_token=hf_token)
|
|
|
35 |
|
36 |
def embed(user, max_length, window_size, method, prompt):
|
37 |
uid = USERS.index(user)
|
@@ -40,7 +41,8 @@ def embed(user, max_length, window_size, method, prompt):
|
|
40 |
watermarked_texts = watermarker.embed(key=args.key, messages=[ uid ],
|
41 |
max_length=max_length, method=method, prompt=prompt)
|
42 |
print("watermarked_texts: ", watermarked_texts)
|
43 |
-
|
|
|
44 |
return watermarked_texts[0].replace(prompt, "")
|
45 |
|
46 |
def detect(attacked_text, window_size, method, prompt):
|
|
|
32 |
model = AutoModelForCausalLM.from_pretrained(args.model, use_auth_token=hf_token, torch_dtype=torch.float16,
|
33 |
device_map='auto').to(device)
|
34 |
tokenizer = AutoTokenizer.from_pretrained(args.model, use_auth_token=hf_token)
|
35 |
+
tokenizer.pad_token = tokenizer.eos_token
|
36 |
|
37 |
def embed(user, max_length, window_size, method, prompt):
|
38 |
uid = USERS.index(user)
|
|
|
41 |
watermarked_texts = watermarker.embed(key=args.key, messages=[ uid ],
|
42 |
max_length=max_length, method=method, prompt=prompt)
|
43 |
print("watermarked_texts: ", watermarked_texts)
|
44 |
+
print(watermarked_texts[0].replace(prompt, ""))
|
45 |
+
watermarked_texts[0].split("[/INST]")[-1]
|
46 |
return watermarked_texts[0].replace(prompt, "")
|
47 |
|
48 |
def detect(attacked_text, window_size, method, prompt):
|