Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,40 @@
|
|
1 |
---
|
|
|
|
|
|
|
|
|
2 |
license: gpl
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
+
language: en
|
3 |
+
tags:
|
4 |
+
- exbert
|
5 |
+
|
6 |
license: gpl
|
7 |
---
|
8 |
+
|
9 |
+
|
10 |
+
# TCP 2023 for NTU students
|
11 |
+
|
12 |
+
Fine tuning pre-trained language models for text generation.
|
13 |
+
|
14 |
+
Pretrained model on Chinese language using a GPT2 for Large Language Head Model objective(GPT2LMHeadModel).
|
15 |
+
|
16 |
+
## Model description
|
17 |
+
|
18 |
+
TCP 2023 is a transformers model that has undergone fine-tuning using the GPT-2 architecture.
|
19 |
+
It was initially pretrained on an extensive corpus of Chinese data in a self-supervised manner.
|
20 |
+
This implies that the pretraining process involved using raw text data without any human annotations, allowing the model to make use of a wide range of publicly available data.
|
21 |
+
The model leveraged an automatic process to derive inputs and corresponding labels from these texts.
|
22 |
+
To be more specific, the pretraining aimed at predicting the subsequent word in sentences.
|
23 |
+
it was trained to guess the next word in sentences.
|
24 |
+
|
25 |
+
### How to use
|
26 |
+
|
27 |
+
You can use this model directly with a pipeline for text generation. Since the generation relies on some randomness, we
|
28 |
+
set a seed for reproducibility:
|
29 |
+
|
30 |
+
```python
|
31 |
+
>>> from transformers import GPT2LMHeadModel, AutoTokenizer, pipeline
|
32 |
+
|
33 |
+
>>> model_name = "DavidLanz/tcp2023"
|
34 |
+
|
35 |
+
>>> model = GPT2LMHeadModel.from_pretrained(model_name)
|
36 |
+
>>> tokenizer = AutoTokenizer.from_pretrained(model_name)
|
37 |
+
|
38 |
+
>>> text_generator = pipeline('text-generation', model=model, tokenizer=tokenizer)
|
39 |
+
>>> generated_text = text_generator(input_text, max_length=max_len, num_return_sequences=1)
|
40 |
+
>>> print(generated_text[0]['generated_text'])
|