DavidLanz commited on
Commit
323ab58
1 Parent(s): 83a77f4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +37 -0
README.md CHANGED
@@ -1,3 +1,40 @@
1
  ---
 
 
 
 
2
  license: gpl
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ language: en
3
+ tags:
4
+ - exbert
5
+
6
  license: gpl
7
  ---
8
+
9
+
10
+ # TCP 2023 for NTU students
11
+
12
+ Fine tuning pre-trained language models for text generation.
13
+
14
+ Pretrained model on Chinese language using a GPT2 for Large Language Head Model objective(GPT2LMHeadModel).
15
+
16
+ ## Model description
17
+
18
+ TCP 2023 is a transformers model that has undergone fine-tuning using the GPT-2 architecture.
19
+ It was initially pretrained on an extensive corpus of Chinese data in a self-supervised manner.
20
+ This implies that the pretraining process involved using raw text data without any human annotations, allowing the model to make use of a wide range of publicly available data.
21
+ The model leveraged an automatic process to derive inputs and corresponding labels from these texts.
22
+ To be more specific, the pretraining aimed at predicting the subsequent word in sentences.
23
+ it was trained to guess the next word in sentences.
24
+
25
+ ### How to use
26
+
27
+ You can use this model directly with a pipeline for text generation. Since the generation relies on some randomness, we
28
+ set a seed for reproducibility:
29
+
30
+ ```python
31
+ >>> from transformers import GPT2LMHeadModel, AutoTokenizer, pipeline
32
+
33
+ >>> model_name = "DavidLanz/tcp2023"
34
+
35
+ >>> model = GPT2LMHeadModel.from_pretrained(model_name)
36
+ >>> tokenizer = AutoTokenizer.from_pretrained(model_name)
37
+
38
+ >>> text_generator = pipeline('text-generation', model=model, tokenizer=tokenizer)
39
+ >>> generated_text = text_generator(input_text, max_length=max_len, num_return_sequences=1)
40
+ >>> print(generated_text[0]['generated_text'])