File size: 1,230 Bytes
4d43185
 
 
 
 
86bc504
4d43185
86bc504
eb3bc9d
07e52cb
eb3bc9d
c19cf37
4d43185
 
 
 
 
 
 
 
 
 
 
6d9b33f
eb3bc9d
e9bec17
eb3bc9d
e9bec17
 
 
4d43185
e9bec17
6d9b33f
4d43185
c19cf37
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
---
# For reference on model card metadata, see the spec: https://github.com/huggingface/hub-docs/blob/main/modelcard.md?plain=1
# Doc / guide: https://huggingface.co/docs/hub/model-cards
{}
---

# Model Card for Model ID

This model is developed to tag Names, Organisations and addresses. I have used a data combined fro Conll, ontonotes5, and a custom address dataset that was self made. Cleaned
out the tags. Detects U.S addresses.
[\"O\", \"B-ORG\", \"I-ORG\", \"B-PER\", \"I-PER\",'B-addr','I-addr']

### Model Description

- **Developed by:** ctrlbuzz
- **Model type:** Bert
- **Language(s) (NLP):** Named Entity recognition
- **Finetuned from model [optional]:** bert-base-cased

## Uses

### Direct Use

```python
from transformers import AutoTokenizer, AutoModelForTokenClassification
from transformers import pipeline
tokenizer = AutoTokenizer.from_pretrained('bert-base-cased')
model = AutoModelForTokenClassification.from_pretrained("ctrlbuzz/bert-addresses")
nlp = pipeline("ner", model=model, tokenizer=tokenizer)
example = "While Maria was representing Johnson & Associates at a conference in Spain, she mailed me a letter from her new office at 123 Elm St., Apt. 4B, Springfield, IL.",

print(nlp(example))
```