JungleLee commited on
Commit
9c59104
1 Parent(s): b7cd0f6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -22,12 +22,12 @@ model_path = "JungleLee/bert-toxic-comment-classification"
22
  tokenizer = BertTokenizer.from_pretrained(model_path)
23
  model = BertForSequenceClassification.from_pretrained(model_path, num_labels=2)
24
 
25
- pipeline = TextClassificationPipeline(model=model, tokenizer=tokenizer)
26
  print(pipeline("You're a fucking nerd."))
27
  ```
28
 
29
  ## Training data
30
- The training data comes this [Kaggle competition](https://www.kaggle.com/c/jigsaw-unintended-bias-in-toxicity-classification/data). We use 90% of the `train.csv` data to train the model.
31
 
32
  ## Evaluation results
33
 
 
22
  tokenizer = BertTokenizer.from_pretrained(model_path)
23
  model = BertForSequenceClassification.from_pretrained(model_path, num_labels=2)
24
 
25
+ pipeline = TextClassificationPipeline(model=model, tokenizer=tokenizer)
26
  print(pipeline("You're a fucking nerd."))
27
  ```
28
 
29
  ## Training data
30
+ The training data comes from this [Kaggle competition](https://www.kaggle.com/c/jigsaw-unintended-bias-in-toxicity-classification/data). We use 90% of the `train.csv` data to train the model.
31
 
32
  ## Evaluation results
33