nicholasKluge commited on
Commit
5a158e8
1 Parent(s): 91f1762

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -0
README.md CHANGED
@@ -46,6 +46,7 @@ co2_eq_emissions:
46
  # TeenyTinyLlama-460m-Chat-awq
47
 
48
  **Note: This model is a quantized version of [TeenyTinyLlama-460m](https://huggingface.co/nicholasKluge/TeenyTinyLlama-460m). Quantization was performed using [AutoAWQ](https://github.com/casper-hansen/AutoAWQ), allowing this version to be 80% lighter, 20% faster, and with almost no performance loss. A GPU is required to run the AWQ-quantized models.**
 
49
  TeenyTinyLlama is a pair of small foundational models trained in Brazilian Portuguese.
50
 
51
  This repository contains a version of [TeenyTinyLlama-460m](https://huggingface.co/nicholasKluge/TeenyTinyLlama-460m) (`TeenyTinyLlama-460m-Chat`) fine-tuned on the [Instruct-Aira Dataset version 2.0](https://huggingface.co/datasets/nicholasKluge/instruct-aira-dataset-v2).
@@ -56,6 +57,7 @@ This repository contains a version of [TeenyTinyLlama-460m](https://huggingface.
56
  - **Batch size:** 4
57
  - **Optimizer:** `torch.optim.AdamW` (warmup_steps = 1e3, learning_rate = 1e-5, epsilon = 1e-8)
58
  - **GPU:** 1 NVIDIA A100-SXM4-40GB
 
59
  - **Quantization Configuration:**
60
  - `bits`: 4
61
  - `group_size`: 128
 
46
  # TeenyTinyLlama-460m-Chat-awq
47
 
48
  **Note: This model is a quantized version of [TeenyTinyLlama-460m](https://huggingface.co/nicholasKluge/TeenyTinyLlama-460m). Quantization was performed using [AutoAWQ](https://github.com/casper-hansen/AutoAWQ), allowing this version to be 80% lighter, 20% faster, and with almost no performance loss. A GPU is required to run the AWQ-quantized models.**
49
+
50
  TeenyTinyLlama is a pair of small foundational models trained in Brazilian Portuguese.
51
 
52
  This repository contains a version of [TeenyTinyLlama-460m](https://huggingface.co/nicholasKluge/TeenyTinyLlama-460m) (`TeenyTinyLlama-460m-Chat`) fine-tuned on the [Instruct-Aira Dataset version 2.0](https://huggingface.co/datasets/nicholasKluge/instruct-aira-dataset-v2).
 
57
  - **Batch size:** 4
58
  - **Optimizer:** `torch.optim.AdamW` (warmup_steps = 1e3, learning_rate = 1e-5, epsilon = 1e-8)
59
  - **GPU:** 1 NVIDIA A100-SXM4-40GB
60
+ - **Carbon emissions** stats are logged in this [file](emissions.csv).
61
  - **Quantization Configuration:**
62
  - `bits`: 4
63
  - `group_size`: 128