|
--- |
|
library_name: peft |
|
base_model: meta-llama/Llama-2-7b-chat-hf |
|
license: apache-2.0 |
|
language: |
|
- en |
|
--- |
|
## Training: |
|
|
|
Check out our github: https://github.com/AI4Finance-Foundation/FinGPT/tree/master/fingpt/FinGPT_Forecaster |
|
|
|
## Inference |
|
``` python |
|
from datasets import load_dataset |
|
from transformers import AutoTokenizer, AutoModelForCausalLM |
|
from peft import PeftModel |
|
|
|
|
|
base_model = AutoModelForCausalLM.from_pretrained( |
|
'meta-llama/Llama-2-7b-chat-hf', |
|
trust_remote_code=True, |
|
device_map="auto", |
|
torch_dtype=torch.float16, # optional if you have enough VRAM |
|
) |
|
tokenizer = AutoTokenizer.from_pretrained('meta-llama/Llama-2-7b-chat-hf') |
|
|
|
model = PeftModel.from_pretrained(base_model, 'FinGPT/fingpt-forecaster_dow30_llama2-7b_lora') |
|
model = model.eval() |
|
``` |
|
|
|
- PEFT 0.5.0 |