Update README.md
Browse files
README.md
CHANGED
@@ -101,7 +101,6 @@ I've tested several other news headline generators on Hugging Face and across th
|
|
101 |
## How to use?
|
102 |
|
103 |
```python
|
104 |
-
# Load model directly
|
105 |
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
|
106 |
|
107 |
tokenizer = AutoTokenizer.from_pretrained("Ateeqq/news-title-generator")
|
@@ -110,22 +109,14 @@ model = AutoModelForSeq2SeqLM.from_pretrained("Ateeqq/news-title-generator")
|
|
110 |
```python
|
111 |
def generate_title(input_text):
|
112 |
|
113 |
-
# Encode the input text
|
114 |
input_ids = tokenizer.encode(input_text, return_tensors="pt")
|
115 |
-
|
116 |
-
# Generate the output
|
117 |
output = model.generate(input_ids)
|
118 |
-
|
119 |
-
# Decode the generated tokens
|
120 |
decoded_text = tokenizer.decode(output[0], skip_special_tokens=True)
|
121 |
-
|
122 |
return decoded_text
|
123 |
|
124 |
-
# Example usage
|
125 |
input_text = "A group of scientists discovered a new planet."
|
126 |
generated_title = generate_title(input_text)
|
127 |
|
128 |
-
print(f"Input Text: {input_text}")
|
129 |
print(f"Generated Title: {generated_title}")
|
130 |
```
|
131 |
|
|
|
101 |
## How to use?
|
102 |
|
103 |
```python
|
|
|
104 |
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
|
105 |
|
106 |
tokenizer = AutoTokenizer.from_pretrained("Ateeqq/news-title-generator")
|
|
|
109 |
```python
|
110 |
def generate_title(input_text):
|
111 |
|
|
|
112 |
input_ids = tokenizer.encode(input_text, return_tensors="pt")
|
|
|
|
|
113 |
output = model.generate(input_ids)
|
|
|
|
|
114 |
decoded_text = tokenizer.decode(output[0], skip_special_tokens=True)
|
|
|
115 |
return decoded_text
|
116 |
|
|
|
117 |
input_text = "A group of scientists discovered a new planet."
|
118 |
generated_title = generate_title(input_text)
|
119 |
|
|
|
120 |
print(f"Generated Title: {generated_title}")
|
121 |
```
|
122 |
|