phanerozoic commited on
Commit
8848cd1
1 Parent(s): fa258f7

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +67 -3
README.md CHANGED
@@ -1,3 +1,67 @@
1
- ---
2
- license: cc-by-nc-4.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-nc-4.0
3
+ language:
4
+ - en
5
+ tags:
6
+ - bert
7
+ - question-classification
8
+ - trec
9
+ widget:
10
+ - text: |
11
+ Enter your question here to classify its type.
12
+ example_title: "Classify Question Type"
13
+ ---
14
+
15
+ # BERT-Question-Classifier
16
+
17
+ The BERT-Question-Classifier is a refined model based on the `bert-base-uncased` architecture. It has been fine-tuned specifically for classifying the types of questions (Description, Entity, Expression, Human, Location, Numeric) using the TREC question classification dataset.
18
+
19
+ - **Developed by**: phanerozoic
20
+ - **Model type**: BertForSequenceClassification
21
+ - **Source model**: `bert-base-uncased`
22
+ - **License**: cc-by-nc-4.0
23
+ - **Languages**: English
24
+
25
+ ## Model Details
26
+
27
+ The BERT-Question-Classifier utilizes a self-attention mechanism to assess the relevance of each word in the context of a question, optimized for categorizing question types.
28
+
29
+ ### Configuration
30
+ - **Attention probs dropout prob**: 0.1
31
+ - **Hidden act**: gelu
32
+ - **Hidden size**: 768
33
+ - **Number of attention heads**: 12
34
+ - **Number of hidden layers**: 12
35
+
36
+ ## Training and Evaluation Data
37
+
38
+ This model is trained on the TREC dataset, which contains a diverse set of question types each labeled under categories such as Description, Entity, Expression, Human, Location, and Numeric.
39
+
40
+ ## Training Procedure
41
+
42
+ The training process was systematically automated to evaluate various hyperparameters, ensuring the selection of optimal settings for the best model performance.
43
+
44
+ - **Initial exploratory training**: Various configurations of epochs, batch sizes, and learning rates were tested.
45
+ - **Focused refinement training**: Post initial testing, the model underwent intensive training with selected hyperparameters to ensure consistent performance and generalization.
46
+
47
+ ### Optimal Hyperparameters Identified
48
+ - **Epochs**: 5
49
+ - **Batch size**: 48
50
+ - **Learning rate**: 2e-5
51
+
52
+ ### Performance
53
+ Post-refinement, the model exhibits high efficacy in question type classification:
54
+ - **Accuracy**: 91%
55
+ - **F1 Score**: 92%
56
+
57
+ ## Usage
58
+
59
+ This model excels in classifying question types in English, ideal for systems needing to interpret and categorize user queries accurately.
60
+
61
+ ## Limitations
62
+
63
+ The BERT-Question-Classifier performs best on question data similar to that found in the TREC dataset. Performance may vary when applied to different domains or languages.
64
+
65
+ ## Acknowledgments
66
+
67
+ Special thanks to the developers of the BERT architecture and the contributions from the Hugging Face team, whose tools and libraries were crucial in the development of this classifier.