0xOracle's picture
header img
311bc36
|
raw
history blame
No virus
2.2 kB
---
tags:
- autotrain
- text-generation
- mistral
- fine-tune
- text-generation-inference
- chat
- Trained with Auto-train
- pytorch
widget:
- text: 'I love AutoTrain because '
license: apache-2.0
language:
- en
library_name: transformers
pipeline_tag: text-generation
---
# Model Trained Using AutoTrain
![LLM_IMAGE](WhatsApp Image 2023-11-20 at 19.06.49.jpeg)
The mistral-7b-fraud2-finetuned Large Language Model (LLM) is a fine-tuned version of the [Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) generative text model using a variety of synthetically generated Fraudulent transcripts datasets.
For full details of this model please read [release blog post](https://mistral.ai/news/announcing-mistral-7b/)
## Instruction format
In order to leverage instruction fine-tuning, your prompt should be surrounded by `[INST]` and `[\INST]` tokens. The very first instruction should begin with a begin of sentence id. The next instructions should not. The assistant generation will be ended by the end-of-sentence token id.
E.g.
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
device = "cuda" # the device to load the model onto
model = AutoModelForCausalLM.from_pretrained("mistralai/Mistral-7B-Instruct-v0.1")
tokenizer = AutoTokenizer.from_pretrained("mistralai/Mistral-7B-Instruct-v0.1")
text = "<s>[INST] Below is a conversation transcript [/INST]"
"Your credit card has been stolen, and you need to contact us to resolve the issue. We will help you protect your information and prevent further fraud.</s> "
"[INST] Analyze the conversation and determine if it's fraudulent or legitimate. [/INST]"
encodeds = tokenizer(text, return_tensors="pt", add_special_tokens=False)
model_inputs = encodeds.to(device)
model.to(device)
generated_ids = model.generate(**model_inputs, max_new_tokens=1000, do_sample=True)
decoded = tokenizer.batch_decode(generated_ids)
print(decoded[0])
```
## Model Architecture
This instruction model is based on Mistral-7B-v0.1, a transformer model with the following architecture choices:
- Grouped-Query Attention
- Sliding-Window Attention
- Byte-fallback BPE tokenizer
## The Team
- BILIC TEAM OF AI ENGINEERS