File size: 2,031 Bytes
311d1da
f16bf60
afb21a1
 
 
 
 
 
 
311d1da
afb21a1
 
 
 
906b214
2fcd04e
906b214
b63ff14
807de2e
afb21a1
906b214
afb21a1
906b214
afb21a1
906b214
afb21a1
906b214
afb21a1
906b214
afb21a1
906b214
afb21a1
906b214
afb21a1
906b214
afb21a1
906b214
afb21a1
906b214
afb21a1
906b214
afb21a1
906b214
afb21a1
906b214
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
---
base_model: ./core2
tags:
- generated_from_trainer
metrics:
- accuracy
model-index:
- name: core2
  results: []
---

<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->

# Model Card: LLama 2 - Version 7b (Embedding + Output + 1 Hidden Layer)

## Overview

- **Link to Training Progress:** [WandB Training Progress](https://wandb.ai/inteligenciaartificialcursos/huggingface/runs/qpsun6iq?workspace=user-inteligenciaartificialcursos)

- **Model Name:** LLama 2 - Version 7b

- **Total Parameters:** 446 million

## Training Data

The model has been trained with the following sequence of datasets:

1. **GPT-2 Data (Done):** The initial training phase involves GPT-2 data and is currently in the finalization stage.

2. **Wikipedia QA in Markdown (In Progress):** The model's training will continue with Wikipedia question-answering data in Markdown format.

3. **QA with Rhetoric (Future Stages):** The model will further be fine-tuned with question-answering data generated from various LLama models, incorporating rhetorical elements.

## Model Description

The LLama 2 - Version 7b model is a powerful language model with a total of 446 million parameters. It utilizes embeddings, an output layer, and one hidden layer to perform a wide range of natural language processing tasks. The training is conducted in multiple stages, each focused on different datasets and objectives.

## Disclaimer

This model card provides an overview of the LLama 2 - Version 7b model, its training data, and intended use cases. Keep in mind that the model's performance may vary depending on the specific task or dataset. Users are encouraged to evaluate the model's suitability for their applications and exercise caution when using it in real-world scenarios.

For any further inquiries or issues related to this model, please contact the model developers through the provided training progress link.

---