|
# KoGPT2-emotion-chatbot |
|
kogpt2 on hugging face Transformers for Psychological Counseling |
|
- [full project link](https://github.com/jiminAn/Capstone_2022) |
|
|
|
## how to use |
|
``` |
|
from transformers import GPT2LMHeadModel, PreTrainedTokenizerFast |
|
|
|
model = GPT2LMHeadModel.from_pretrained("withU/kogpt2-emotion-chatbot") |
|
tokenizer = PreTrainedTokenizerFast.from_pretrained("withU/kogpt2-emotion-chatbot") |
|
|
|
input_ids = tokenizer.encode("์๋
", add_special_tokens=False, return_tensors="pt") |
|
output_sequences = model.generate(input_ids=input_ids, do_sample=True, max_length=80, num_return_sequences=4) |
|
for generated_sequence in output_sequences: |
|
generated_sequence = generated_sequence.tolist() |
|
print("GENERATED SEQUENCE : {0}".format(tokenizer.decode(generated_sequence, clean_up_tokenization_spaces=True))) |
|
``` |
|
## dataset finetuned on |
|
- [wellness dataset](https://aihub.or.kr/opendata/keti-data/recognition-laguage/KETI-02-006) |
|
- [emotion corpus of conversations](https://aihub.or.kr/opendata/keti-data/recognition-laguage/KETI-02-010) |
|
- [chatbot data](https://jeongukjae.github.io/tfds-korean/datasets/korean_chatbot_qa_data.html) |
|
## references |
|
- [WelllnessConversation-LanguageModel](https://github.com/nawnoes/WellnessConversation-LanguageModel) |
|
- [KoGPT2: SKT-AI](https://github.com/SKT-AI/KoGPT2) |