Vigalpaca-French-7B-ties
Vigalpaca-French-7B-ties is a merge of the following models:
jpacifico/French-Alpaca-7B-Instruct-beta
bofenghuang/vigostral-7b-chat
base model : jpacifico/French-Alpaca-7B-Instruct-beta
## Usage
```python
!pip install -qU transformers accelerate
from transformers import AutoTokenizer
import transformers
import torch
model = "jpacifico/Vigalpaca-French-7B-ties"
messages = [{"role": "user", "content": "What is a large language model?"}]
tokenizer = AutoTokenizer.from_pretrained(model)
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
pipeline = transformers.pipeline(
"text-generation",
model=model,
torch_dtype=torch.float16,
device_map="auto",
)
outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
print(outputs[0]["generated_text"])
Limitations
The Vigalpaca model is a quick demonstration that a base 7B model can be easily merged/fine-tuned to specialize in a particular language. It does not have any moderation mechanisms.
- Developed by: Jonathan Pacifico. Vigostral model by Bofeng Huang (special thanks), 2024
- Model type: LLM
- Language(s) (NLP): French
- License: Apache-2.0
- Downloads last month
- 13
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.
Model tree for jpacifico/Vigalpaca-French-7B-ties
Merge model
this model