Token Classification
Transformers
PyTorch
English
roberta
feature-extraction
Inference Endpoints
File size: 1,260 Bytes
6518761
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
edf16a0
6518761
 
 
 
44543be
6518761
 
 
 
ab8b6ed
 
 
602334e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6518761
 
 
 
cc5d64d
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
---
datasets:
- tner/bc5cdr
- tner/bionlp2004
- tner/btc
- tner/conll2003
- tner/fin
- tner/mit_movie_trivia
- tner/mit_restaurant
- tner/multinerd
- tner/ontonotes5
- tner/tweebank_ner
- tner/tweetner7
- tner/wikineural
- tner/wnut2017
language:
- en
metrics:
- accuracy
- f1
pipeline_tag: token-classification
---
# RoBERTa Span Detection 
This model is a fine-tuned model of [roberta-large](https://huggingface.co/roberta-large) after being trained on a **mixture of NER datasets**.

Basically, this model can detect NER spans (with <u>no differenciation on classes</u>). Labels use the IBO format and are: 
- 'B-TAG': beginning token of span
- 'I-TAG': inside token of span
- 'O': token not a span

# Usage
This model has been trained in an efficient way and thus cannot be load directly from HuggingFace's hub. To use that model, please follow instructions on this [repo](https://github.com/AntoineBlanot/efficient-llm).

# Data used for training
- tner/bc5cdr
- tner/bionlp2004
- tner/btc
- tner/conll2003
- tner/fin
- tner/mit_movie_trivia
- tner/mit_restaurant
- tner/multinerd
- tner/ontonotes5
- tner/tweebank_ner
- tner/tweetner7
- tner/wikineural
- tner/wnut2017

# Evaluation results

| Data | Accuracy |
|:---:|:---------:|
| validation | 0.972 |