dnzblgn commited on
Commit
0b6147a
1 Parent(s): 17ffbe8

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +44 -0
README.md ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Entity Recognition (NER) Model Card
2
+ Model Overview
3
+ Model Name: NER LSTM Model
4
+ Description: A LSTM-based model for Named Entity Recognition (NER) task. The model aims to classify words in text into their respective named entity categories such as Person, Organization, Location, etc.
5
+
6
+ Intended Use
7
+ Primary Use Case: Extracting named entities (e.g., names of people, organizations, locations) from text.
8
+
9
+ Usage Instructions:
10
+
11
+ Install the required libraries: Ensure that the required libraries, such as pandas, scikit-learn, keras, and tensorflow, are installed.
12
+ Load the model and tokenizer: Use the Hugging Face Transformers library to load the model and tokenizer from the provided files.
13
+ Tokenize input text: Preprocess input text and tokenize it using the loaded tokenizer.
14
+ Make predictions: Feed the tokenized input through the model to obtain predictions for named entity categories.
15
+ Post-process predictions: Use the LabelEncoder to map model predictions back to human-readable named entity categories.
16
+ Performance and Evaluation
17
+ Performance Metrics:
18
+
19
+ Test Loss: The loss value achieved on the test dataset.
20
+ Test Accuracy: The accuracy achieved on the test dataset.
21
+ Training Accuracy: The accuracy achieved on the training dataset.
22
+ Validation Accuracy: The accuracy achieved on the validation dataset.
23
+ Performance Summary:
24
+
25
+ The model achieved an accuracy of approximately [Test Accuracy] on the test dataset.
26
+ Training and validation accuracies are provided for reference.
27
+ Dataset
28
+ Dataset Name: NER dataset.csv
29
+ Description: The dataset contains labeled data for named entity recognition. It includes columns for 'Word' and 'POS' (Part-of-Speech) labels.
30
+
31
+ Model Details
32
+ Architecture:
33
+
34
+ Embedding Layer: Converts input tokens into dense vectors.
35
+ LSTM Layer: Processes the sequence of word embeddings.
36
+ Dense Layer: Produces a probability distribution over named entity categories.
37
+ Hyperparameters:
38
+
39
+ Embedding Dimension: 100
40
+ LSTM Units: 128
41
+ Batch Size: 64
42
+ Max Sequence Length: 100
43
+ Optimizer: Adam
44
+ Loss Function: Sparse Categorical Cross-Entropy