--- datasets: - BI55/MedText metrics: - bleu - perplexity pipeline_tag: text-generation --- A further fine-tuned version of Locutusque/gpt2-large-conversational on MedText # Evaluation This model was evaluated using GPT-3.5, and it was asked medical questions. It achieved an average accuracy of 75.8%.