Edit model card

LogiPhi-2.7B-ModelStock

image/jpeg

LogiPhi-2.7B-ModelStock is a merge of the following models using LazyMergekit: Special thanks to Charles Goddard for the quick implementation!

πŸ† Evaluation

Coming Soon

🧩 Configuration

models:
  - model: microsoft/phi-2
  - model: MaziyarPanahi/phi-2-logical-sft
  - model: rhysjones/phi-2-orange-v2
merge_method: model_stock
base_model: microsoft/phi-2
dtype: float16

πŸ’» Usage

!pip install -qU transformers accelerate

from transformers import AutoTokenizer
import transformers
import torch

model = "johnsnowlabs/LogiPhi-2.7B-ModelStock"
messages = [{"role": "user", "content": "What is a large language model?"}]

tokenizer = AutoTokenizer.from_pretrained(model)
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
pipeline = transformers.pipeline(
    "text-generation",
    model=model,
    torch_dtype=torch.float16,
    device_map="auto",
)

outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
print(outputs[0]["generated_text"])
Downloads last month
2
Safetensors
Model size
2.65B params
Tensor type
F32
Β·
Inference API
Model is too large to load in Inference API (serverless). To try the model, launch it on Inference Endpoints (dedicated) instead.