passthepizza
commited on
Commit
•
83a2196
1
Parent(s):
efb3463
Update README.md
Browse files
README.md
CHANGED
@@ -1,85 +1,24 @@
|
|
1 |
-
|
2 |
-
license: mit
|
3 |
-
language:
|
4 |
-
- en
|
5 |
-
base_model:
|
6 |
-
- meta-llama/Llama-3.1-70B-Instruct
|
7 |
-
---
|
8 |
-
|
9 |
-
# Cakrawala-70B
|
10 |
-
|
11 |
-
## Model Description
|
12 |
-
|
13 |
-
Cakrawala-70B is a fine-tuned variant of the Llama-3.1-70B-Instruct model, specifically optimized for generating rich roleplaying conversations and character interactions. The model uses QLoRA (Quantized Low-Rank Adaptation) fine-tuning techniques to efficiently adapt the large language model for this specialized use case.
|
14 |
-
|
15 |
-
## Intended Use
|
16 |
|
17 |
-
|
18 |
-
Cakrawala-70B is designed specifically for generating high-quality roleplaying conversations with the following key characteristics:
|
19 |
-
- Rich, descriptive character interactions
|
20 |
-
- Consistent character voice and emotional development
|
21 |
-
- Show-don't-tell emotional states
|
22 |
-
- Clear separation between character perspectives
|
23 |
-
- Structured turn-taking in conversations
|
24 |
-
- Detailed physical descriptions and environmental awareness
|
25 |
|
26 |
-
|
27 |
-
- Game developers creating interactive narratives
|
28 |
-
- Writers seeking AI assistance in character development
|
29 |
-
- RPG platforms and applications
|
30 |
-
- Interactive fiction developers
|
31 |
-
- Educational platforms teaching creative writing or character development
|
32 |
|
33 |
-
|
34 |
|
35 |
-
|
36 |
-
- Total examples: 5,867 conversation pairs
|
37 |
-
- Format: JSON Lines (.jsonl)
|
38 |
-
- Structure: Conversations field containing alternating messages between participants
|
39 |
-
- Validation split: 5% of total data
|
40 |
|
41 |
-
###
|
42 |
-
|
43 |
-
|
44 |
-
|
45 |
-
3. Rich descriptive elements including:
|
46 |
-
- Physical actions
|
47 |
-
- Facial expressions
|
48 |
-
- Tone indicators
|
49 |
-
- Environmental details
|
50 |
-
- Character reactions
|
51 |
|
52 |
-
###
|
53 |
-
-
|
54 |
-
-
|
55 |
-
-
|
56 |
-
- Sequence length is set to 2048 tokens
|
57 |
-
- Sample packing is enabled for efficient training
|
58 |
|
59 |
-
## Training
|
60 |
-
|
61 |
-
### Base Model
|
62 |
-
- Architecture: meta-llama/Llama-3.1-70B-Instruct
|
63 |
-
- Model Type: LlamaForCausalLM
|
64 |
-
- Tokenizer: AutoTokenizer
|
65 |
-
|
66 |
-
### Fine-tuning Approach
|
67 |
-
- Method: QLoRA (Quantized Low-Rank Adaptation)
|
68 |
-
- Quantization: 4-bit precision
|
69 |
-
- Sequence Length: 2048 tokens
|
70 |
-
- Training Duration: 3 epochs
|
71 |
-
|
72 |
-
### LoRA Configuration
|
73 |
-
- Rank (r): 32
|
74 |
-
- Alpha: 64
|
75 |
-
- Dropout: 0.1
|
76 |
-
- Target Modules:
|
77 |
-
- Query Projection (q_proj)
|
78 |
-
- Key Projection (k_proj)
|
79 |
-
- Value Projection (v_proj)
|
80 |
-
- Output Projection (o_proj)
|
81 |
-
|
82 |
-
### Training Parameters
|
83 |
- Gradient Accumulation Steps: 16
|
84 |
- Micro Batch Size: 4
|
85 |
- Learning Rate: 0.0003
|
@@ -87,19 +26,14 @@ Each training example consists of:
|
|
87 |
- Scheduler: Cosine
|
88 |
- Mixed Precision: BF16 & FP16 with TF32 support
|
89 |
|
90 |
-
##
|
91 |
-
|
92 |
-
## Limitations
|
93 |
|
94 |
-
|
95 |
-
- Training data size (5,867 examples) may limit variety in some scenarios
|
96 |
-
- Specialized for roleplaying conversations, may not generalize well to other tasks
|
97 |
|
98 |
-
|
|
|
99 |
|
100 |
-
|
101 |
-
- Pad Token: <|end_of_text|>
|
102 |
|
103 |
-
|
104 |
-
- Supports 8 x H100 NVL configuration
|
105 |
-
- Utilizes 128 vCPU and 1509 GB RAM
|
|
|
1 |
+
# 🎭 Cakrawala-70B
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
2 |
|
3 |
+
> *"Where Worlds Converge and Adventures Begin!"*
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
4 |
|
5 |
+
## 🌟 What's Special About This Model?
|
|
|
|
|
|
|
|
|
|
|
6 |
|
7 |
+
Cakrawala-70B is a fine-tuned variant of the Llama-3.1-70B-Instruct model, specifically optimised for generating rich roleplaying conversations and character interactions. The model has been trained to excel at producing detailed, contextually appropriate character dialogues with rich descriptions of physical actions, expressions, and emotional states while maintaining consistent character voices and perspectives throughout extended interactions.
|
8 |
|
9 |
+
## 🧪 The Secret Sauce
|
|
|
|
|
|
|
|
|
10 |
|
11 |
+
### Training Diet:
|
12 |
+
- Fed with 5,867 conversation pairs
|
13 |
+
- Each conversation is a minimum 12-13 turns long
|
14 |
+
- Focused heavily details like facial expressions, environmental descriptions, and character reactions that are focused a lot on **keeping the model in character.**
|
|
|
|
|
|
|
|
|
|
|
|
|
15 |
|
16 |
+
### Tech Wizardry:
|
17 |
+
- Trained on the mighty Llama-3.1-70B-Instruct
|
18 |
+
- Fine-tuned using QLoRA
|
19 |
+
- Trained over 3 epochs
|
|
|
|
|
20 |
|
21 |
+
## Training Parameters
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
22 |
- Gradient Accumulation Steps: 16
|
23 |
- Micro Batch Size: 4
|
24 |
- Learning Rate: 0.0003
|
|
|
26 |
- Scheduler: Cosine
|
27 |
- Mixed Precision: BF16 & FP16 with TF32 support
|
28 |
|
29 |
+
## 🔧 Under the Hood
|
30 |
+
- Trained on 8 x H100 NVL GPUs
|
|
|
31 |
|
32 |
+
## 🎬 License & Credits
|
|
|
|
|
33 |
|
34 |
+
- Licensed under MIT
|
35 |
+
- Based on meta-llama/Llama-3.1-70B-Instruct
|
36 |
|
37 |
+
---
|
|
|
38 |
|
39 |
+
*Built with ❤️ for roleplayers, by roleplayers*
|
|
|
|