|
--- |
|
license: bsd-3-clause |
|
datasets: |
|
- daily_dialog |
|
language: |
|
- en |
|
tags: |
|
- chat |
|
- dialog |
|
- natural language |
|
- codegen |
|
widget: |
|
- text: "Bot responds in friendly, longer messages.\n<Chat history>\n[User]: 'Hello!'\n[Bot]: 'Hi there! :D How are you?'\n[User]: 'Good, you?'\n[Bot]: '" |
|
example_title: "Greeting chat" |
|
co2_eq_emissions: |
|
emissions: 40 |
|
source: https://mlco2.github.io/impact/#compute |
|
training_type: fine-tuning |
|
geographical_location: Oregon, USA |
|
hardware_used: 1x NVIDIA P100, Kaggle |
|
library_name: transformers |
|
--- |
|
|
|
# DailyChat-350M |
|
A finetuned version of [Codegen-350M-nl](https://huggingface.co/Salesforce/Codegen-350M-nl) on the 'daily_dialog' dataset. |
|
The idea of this model is to create one that is capable of holding a decent conversation. |
|
|
|
# Training Procedure |
|
This was trained on Kaggle's servers using 1x NVIDIA P100. |
|
This model was trained for 1 epoch with learning rate 1e-2. |
|
|
|
# Biases & Limitations |
|
This likely contains the same biases and limitations as the original model that it is based on, and additionally heavy biases from the dataset. |
|
It *can* generate offensive input when prompted, so user discretion is advised. |
|
|
|
# Intended Use |
|
Dialog generation, chat agents. |