Edit model card

Galen

Galen is fine-tuned from Mistral-7B-Instruct-v0.2, using medical quesion answering dataset

Galen's view about future of medicine and AI:

alt text

Get Started

Install "accelerate" to use CUDA GPU

pip install accelerate
from transformers import AutoTokenizer, pipeline
tokenizer = AutoTokenizer.from_pretrained('ahmed-ai/galen')
model_pipeline = pipeline(task="text-generation", model='ahmed-ai/galen', tokenizer=tokenizer, max_length=256, temperature=0.5, top_p=0.6)
result = model_pipeline('What is squamous carcinoma')
#print the generated text
print(result[0]['generated_text'][len(prompt):])
Downloads last month
1
Safetensors
Model size
3.86B params
Tensor type
F32
·
BF16
·
U8
·
Inference API
Input a message to start chatting with ahmed-ai/galen.
Inference API (serverless) has been turned off for this model.

Dataset used to train ahmed-ai/galen