--- license: bsd-3-clause datasets: - daily_dialog language: - en tags: - chat - dialog - natural language - codegen widget: - text: "Bot responds in friendly, longer messages.\n\n[User]: 'Hello!'\n[Bot]: 'Hi there! :D How are you?'\n[User]: 'Good, you?'\n[Bot]: '" example_title: "Greeting chat" co2_eq_emissions: emissions: 40 source: https://mlco2.github.io/impact/#compute training_type: fine-tuning geographical_location: Oregon, USA hardware_used: 1x NVIDIA P100, Kaggle library_name: transformers --- # DailyChat-350M A finetuned version of [Codegen-350M-nl](https://huggingface.co/Salesforce/Codegen-350M-nl) on the 'daily_dialog' dataset. The idea of this model is to create one that is capable of holding a decent conversation. # Training Procedure This was trained on Kaggle's servers using 1x NVIDIA P100. This model was trained for 1 epoch with learning rate 1e-2. # Biases & Limitations This likely contains the same biases and limitations as the original model that it is based on, and additionally heavy biases from the dataset. It *can* generate offensive input when prompted, so user discretion is advised. # Intended Use Dialog generation, chat agents.