File size: 1,841 Bytes
3cd0e33
 
 
 
 
 
 
 
 
 
 
 
d031d17
674284e
3cd0e33
ac4e8f4
3cd0e33
5efbb2c
3cd0e33
ac4e8f4
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3cd0e33
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
---
base_model: Xclbr7/Arcanum-12b
language:
- en
license: apache-2.0
tags:
- text-generation-inference
- transformers
- unsloth
- mistral
- trl
---
<img src="https://cdn-uploads.huggingface.co/production/uploads/66dcee3321f901b049f48002/Fpdr8qCx9Xx4RHWgptCGD.png" width="800"/>


# Aether-12b

Aether-12b is a fine-tuned large language model based on Arcanum-12b, further trained on the CleverBoi-Data-20k dataset. 

## Model Details πŸ“Š
- Developed by: AIXON Lab
- Model type: Causal Language Model
- Language(s): English (primarily), may support other languages
- License: apache-2.0
- Repository: https://huggingface.co/aixonlab/Aether-12b

## Model Architecture πŸ—οΈ
- Base model: Arcanum-12b
- Parameter count: ~12 billion
- Architecture specifics: Transformer-based language model

## Open LLM Leaderboard Evaluation Results
Coming Soon !

## Training & Fine-tuning πŸ”„
Aether-12b was fine-tuned on the following dataset:
- Dataset: theprint/CleverBoi-Data-20k
- Fine-tuning method: TRL SFTTrainer with AdamW optimizer, cosine decay LR scheduler, bfloat16 precision.

The CleverBoi-Data-20k dataset improved the model in the following ways:
1. Enhanced reasoning and problem-solving capabilities
2. Broader knowledge across various topics
3. Improved performance on specific tasks like writing, analysis, and problem-solving
4. Better contextual understanding and response generation

## Intended Use 🎯
As an assistant or specific role bot.

## Ethical Considerations πŸ€”
As a fine-tuned model based on Arcanum-12b, this model may inherit biases and limitations from its parent model and the fine-tuning dataset. Users should be aware of potential biases in generated content and use the model responsibly.


## Acknowledgments πŸ™
We acknowledge the contributions of:
- theprint for the amazing CleverBoi-Data-20k dataset