Transformers
Safetensors
English
medical
text-generation-inference
llm
biomistral
Inference Endpoints

You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

Model Card for BioMistral Multi-Turn Doctor Conversation Model

Model Details

Model Description

This model is a fine-tuned version of the BioMistral model, specifically tailored for multi-turn doctor-patient conversations. It leverages the powerful language generation capabilities of BioMistral to provide accurate and context-aware responses in medical dialogue scenarios.

  • Developed by: Siyahul Haque T P
  • Model type: Text-generation (LLM)
  • Language(s) (NLP): English (en)
  • License: Apache-2.0
  • Finetuned from model: BioMistral

Uses

Direct Use

This model can be directly used for generating responses in multi-turn medical conversations, making it useful for applications like virtual health assistants and medical chatbots.

Downstream Use

This model can be further fine-tuned or integrated into larger healthcare applications, such as patient management systems or automated symptom checkers.

Out-of-Scope Use

The model is not suitable for use in emergency medical situations, providing final diagnoses, or replacing professional medical advice.

Bias, Risks, and Limitations

The model may reflect biases present in the training data, including underrepresentation of certain medical conditions or demographic groups. The model should not be used as a sole source of medical information and must be supervised by qualified healthcare professionals.

Recommendations

Users should be aware of the potential biases and limitations of the model. It is recommended to use the model as a supplementary tool rather than a primary source of medical advice.

How to Get Started with the Model

from transformers import AutoModelForCausalLM, AutoTokenizer

# Load the tokenizer and model from the Hugging Face Hub
tokenizer = AutoTokenizer.from_pretrained("BioMistral/BioMistral-7B")
model = AutoModelForCausalLM.from_pretrained("siyah1/BioMistral-7b-Chat-Doctor")

# Example input: patient describing a symptom
input_text = "Hello, doctor, I have a headache."

# Tokenize the input text
inputs = tokenizer(input_text, return_tensors="pt")

# Generate a response from the model
outputs = model.generate(**inputs, max_length=100, num_return_sequences=1)

# Decode the generated response
response = tokenizer.decode(outputs[0], skip_special_tokens=True)

# Print the model's response
print("Doctor:", response)
Downloads last month
8
Safetensors
Model size
304M params
Tensor type
F32
·
FP16
·
Inference API
Unable to determine this model’s pipeline type. Check the docs .

Datasets used to train siyah1/BioMistral-7b-Chat-Doctor