--- language: - en metrics: - accuracy co2_eq_emissions: emissions: "10" source: "mlco2.github.io" training_type: "fine-tuning" geographical_location: "West Java, Indonesia" hardware_used: "1 T4" license: cc-by-nc-sa-4.0 widget: - text: 'You: "Hey kekbot! Whats up?"\nKekbot: "' example_title: "Asking what's up" - text: 'You: "Hey kekbot! How r u?"\nKekbot: "' example_title: "Asking how he is" --- > THIS MODEL IS INTENDED FOR RESEARCH PURPOSES ONLY # Kekbot Mini Based on a `distilgpt2` model, fine-tuned to a select subset (65k<= messages) of Art Union's general-chat channel chat history. ### Limits and biases As this is trained on chat history, it is possible that discriminatory or even offensive materials to be outputted. Author holds his ground on the fact that ML models are mere statistical representation of the dataset used to train it, and that due to the nature of the dataset it is practically impossible to be certain of the degree of "cleanliness" that the data contained within holds. Author can confirm, however, that from heuristical testing that the model was not found to be offensive to the author himself, hopefully this opinion stays true for everyone in the audience.