--- tags: autotrain language: en widget: - text: "I quite enjoy using AutoTrain due to its simplicity." datasets: - hidude562/autotrain-data-SimpleDetect co2_eq_emissions: 0.21691606119445225 --- # Model Description This model detects if you are writing in a format that is more similar to Simple English Wikipedia or English Wikipedia. This can be extended to applications that aren't Wikipedia as well and to some extent, it can be used for other languages. Please also note there is a major bias to special characters (Mainly the hyphen mark, but it also applies to others) so I would recommend removing them from your input text. # Model Trained Using AutoTrain - Problem type: Multi-class Classification - Model ID: 837726721 - CO2 Emissions (in grams): 0.21691606119445225 ## Validation Metrics - Loss: 0.010096958838403225 - Accuracy: 0.996223414828066 - Macro F1: 0.996179398826373 - Micro F1: 0.996223414828066 - Weighted F1: 0.996223414828066 - Macro Precision: 0.996179398826373 - Micro Precision: 0.996223414828066 - Weighted Precision: 0.996223414828066 - Macro Recall: 0.996179398826373 - Micro Recall: 0.996223414828066 - Weighted Recall: 0.996223414828066 ## Usage You can use cURL to access this model: ``` $ curl -X POST -H "Authorization: Bearer YOUR_API_KEY" -H "Content-Type: application/json" -d '{"inputs": "I quite enjoy using AutoTrain due to its simplicity."}' https://api-inference.huggingface.co/models/hidude562/Wiki-Complexity ``` Or Python API: ``` from transformers import AutoModelForSequenceClassification, AutoTokenizer model = AutoModelForSequenceClassification.from_pretrained("hidude562/Wiki-Complexity", use_auth_token=True) tokenizer = AutoTokenizer.from_pretrained("hidude562/Wiki-Complexity", use_auth_token=True) inputs = tokenizer("I quite enjoy using AutoTrain due to its simplicity.", return_tensors="pt") outputs = model(**inputs) ```