MisRoberta_lstm / config.json
UNCANNY69's picture
Upload 6 files
9eb7d2b verified
raw
history blame
405 Bytes
{
"architectures": [
"RoBERTaLSTMForSequenceClassification"
],
"dropout_rate": 0.1,
"embed_dim": 768,
"hidden_dim_lstm": 256,
"id2label": {
"0": "fake",
"1": "true"
},
"label2id": {
"fake": 0,
"true": 1
},
"model_type": "robertaLSTMForSequenceClassification",
"num_classes": 2,
"num_layers": 12,
"torch_dtype": "float32",
"transformers_version": "4.38.2"
}