Model Card for Model ID
The GlueQwen model is fine-tuned on four distinct tasks from the GLUE benchmark: SST-2 (Sentiment Analysis), MRPC (Paraphrase Detection), CoLA (Linguistic Acceptability), and MNLI (Natural Language Inference). The base model used is Qwen/Qwen2.5-7B, which has 7 billion parameters. Qwen2.5-7B is designed to enhance various natural language understanding tasks through pre-training on diverse datasets, followed by fine-tuning for task-specific improvements.
The fine-tuning of GlueQwen involves optimizing the model for these GLUE tasks, aiming to measure catastrophic forgetting, model learning abilities, and overall training performance across different language tasks. These benchmarks provide insight into how well the model retains previous knowledge while learning new tasks sequentially.
Model Details
Model Description
Benchmark Table for GlueQwen Fine-Tuning Performance
Model | Parameter Size (B) | Pretrained Performance | Forgetting | Learning | Training Performance |
---|---|---|---|---|---|
Llama-3.2-1B | 1 | 0.50 | 0.24 | 0.33 | 0.54 |
Llama-3.2-3B | 3 | 0.56 | 0.225 | 0.36 | 0.61 |
Llama-3.1-8B | 8 | 0.56 | 0.59 | 0.84 | 0.67 |
Llama-3-8B | 8 | 0.53 | 0.39 | 0.98 | 0.70 |
Llama-2-7B | 7 | 0.67 | 0.23 | 0.12 | 0.63 |
GPT-J-6B | 6 | 0.50 | 0.39 | 0.45 | 0.54 |
Phi-2 | 2.7 | 0.59 | 0.10 | 0.15 | 0.61 |
Phi-3.5-mini | 3.82 | 0.69 | 0.02 | 0.30 | 0.76 |
Orca-2-7b | 7 | 0.76 | 0.185 | 0.33 | 0.81 |
Qwen2.5-0.5B | 0.5 | 0.52 | 0.23 | 0.56 | 0.61 |
Qwen2.5-7B | 7 | 0.56 | 0.51 | 1.12 | 0.77 |
Qwen2.5-14B | 14 | 0.71 | 0.935 | 0.66 | 0.80 |
GlueQwen | 7 | 0.59 | 0.42 | 0.97 | 0.73 |
Analysis
GlueQwen, fine-tuned on multiple tasks from the GLUE dataset, demonstrates a pre-trained performance of 0.59. Its forgetting rate is moderate at 0.42, reflecting some loss of previously learned information. However, the model exhibits a strong learning capability with a learning score of 0.97. The overall training performance stands at 0.73, positioning GlueQwen as a balanced model that manages forgetting while achieving significant improvements in task-specific learning.
- Developed by: [More Information Needed]
- Funded by [optional]: [More Information Needed]
- Shared by [optional]: [More Information Needed]
- Model type: [More Information Needed]
- Language(s) (NLP): [More Information Needed]
- License: [More Information Needed]
- Finetuned from model [optional]: [More Information Needed]
Model Sources [optional]
- Repository: [More Information Needed]
- Paper [optional]: [More Information Needed]
- Demo [optional]: [More Information Needed]
Uses
Direct Use
[More Information Needed]
Downstream Use [optional]
[More Information Needed]
Out-of-Scope Use
[More Information Needed]
Bias, Risks, and Limitations
[More Information Needed]
Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
Training Details
Training Data
[More Information Needed]
Training Procedure
Preprocessing [optional]
[More Information Needed]
Training Hyperparameters
- Training regime: [More Information Needed]
Speeds, Sizes, Times [optional]
[More Information Needed]
Evaluation
Testing Data, Factors & Metrics
Testing Data
[More Information Needed]
Factors
[More Information Needed]
Metrics
[More Information Needed]
Results
[More Information Needed]
Summary
Model Examination [optional]
[More Information Needed]
Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type: [More Information Needed]
- Hours used: [More Information Needed]
- Cloud Provider: [More Information Needed]
- Compute Region: [More Information Needed]
- Carbon Emitted: [More Information Needed]
Technical Specifications [optional]
Model Architecture and Objective
[More Information Needed]
Compute Infrastructure
[More Information Needed]
Hardware
[More Information Needed]
Software
[More Information Needed]
Citation [optional]
BibTeX:
[More Information Needed]
APA:
[More Information Needed]
Glossary [optional]
[More Information Needed]
More Information [optional]
[More Information Needed]
Model Card Authors [optional]
[More Information Needed]
Model Card Contact
[More Information Needed]
- Downloads last month
- 12
Model tree for naimul011/GlueQwen
Base model
Qwen/Qwen2.5-7B