fix: Adjust call to generate_replacements to use token IDs for prefix_tokens
Browse files
app.py
CHANGED
|
@@ -72,7 +72,7 @@ start_time = time.time()
|
|
| 72 |
for word in tqdm(low_prob_words, desc="Processing words"):
|
| 73 |
iteration_start_time = time.time()
|
| 74 |
prefix_index = word.first_token_index
|
| 75 |
-
prefix_tokens = [token for token, _ in result][:prefix_index + 1]
|
| 76 |
replacements = generate_replacements(model, tokenizer, prefix_tokens, device)
|
| 77 |
print(f"Original word: {word.text}, Log Probability: {word.logprob:.4f}")
|
| 78 |
print(f"Proposed replacements: {replacements}")
|
|
|
|
| 72 |
for word in tqdm(low_prob_words, desc="Processing words"):
|
| 73 |
iteration_start_time = time.time()
|
| 74 |
prefix_index = word.first_token_index
|
| 75 |
+
prefix_tokens = tokenizer.convert_tokens_to_ids([token for token, _ in result][:prefix_index + 1])
|
| 76 |
replacements = generate_replacements(model, tokenizer, prefix_tokens, device)
|
| 77 |
print(f"Original word: {word.text}, Log Probability: {word.logprob:.4f}")
|
| 78 |
print(f"Proposed replacements: {replacements}")
|