--- library_name: transformers datasets: - kallantis/Greek-Humorous-Dataset language: - el pipeline_tag: text-classification --- # # # The model is based on XLM-RoBERTa large ("xlm-roberta-large") fine-tuned for Humor Recognition in Greek language. # ## Model Details The model was pre-trained over 10 epochs on Greek Humorous Dataset # ## Pre-processing details The text needs to be pre-processed by removing all greek diacritics and punctuation and converting all letters to lowercase ## Load Pretrained Model ```python from transformers import AutoTokenizer, XLMRobertaForSequenceClassification tokenizer = AutoTokenizer.from_pretrained("kallantis/Humor-Recognition-Greek-XLM-R-large") model = XLMRobertaForSequenceClassification.from_pretrained("kallantis/Humor-Recognition-Greek-XLM-R-large", num_labels=2, ignore_mismatched_sizes=True) ```