--- license: mit datasets: - keivalya/MedQuad-MedicalQnADataset language: - en library_name: diffusers tags: - medical --- # Model Card for GaiaMiniMed This is a medical fine tuned model from the [Falcon-7b-Instruction](https://huggingface.co/tiiuae/falcon-7b-instruct) Base using 500 steps & 6 epochs with [MedAware](https://huggingface.co/datasets/keivalya/MedQuad-MedicalQnADataset) Dataset from [keivalya](https://huggingface.co/datasets/keivalya) ## Model Details ### Model Description - **Developed by:** [Tonic](https://www.huggingface.co/tonic) - **Shared by :** [Tonic](https://www.huggingface.co/tonic) - **Model type:** Medical Fine-Tuned Conversational Falcon 7b (Instruct) - **Language(s) (NLP):** English - **License:** MIT - **Finetuned from model:**[tiiuae/falcon-7b-instruct](https://huggingface.co/tiiuae/falcon-7b-instruct) - ### Model Sources [optional] - **Repository:** [Github](https://github.com/Josephrp/AI-challenge-hackathon/blob/master/falcon_7b_instruct_GaiaMiniMed_dataset.ipynb) - **Demo [optional]:** {{ demo | default("[More Information Needed]", true)}} ## Uses Use this model like you would use Falcon Instruct Models ### Direct Use This model is intended for educational purposes only , always consult a doctor for the best advice. This model should perform better at medical QnA tasks in a conversational manner. It is our hope that it will help improve patient outcomes and public health. ### Downstream Use Use this model next to others and have group conversations to produce diagnoses , public health advisory , and personal hygene improvements. ### Out-of-Scope Use This model is not meant as a decision support system in the wild, only for educational use. ## Bias, Risks, and Limitations {{ bias_risks_limitations | default("[More Information Needed]", true)}} ## How to Get Started with the Model Use the code below to get started with the model. {{ get_started_code | default("[More Information Needed]", true)}} ## Training Details ### Results ![image/png](https://cdn-uploads.huggingface.co/production/uploads/62a3bb1cd0d8c2c2169f0b88/F8GfMSJcAaH7pXvpUK_r3.png) ```json TrainOutput(global_step=6150, training_loss=1.0597990553941183, {'epoch': 6.0}) ``` ### Training Data ```json DatasetDict({ train: Dataset({ features: ['qtype', 'Question', 'Answer'], num_rows: 16407 }) }) ``` ### Training Procedure #### Preprocessing [optional] ``` trainable params: 4718592 || all params: 3613463424 || trainables%: 0.13058363808693696 ``` #### Training Hyperparameters - **Training regime:** {{ training_regime | default("[More Information Needed]", true)}} #### Speeds, Sizes, Times [optional] ```json metrics={'train_runtime': 30766.4612, 'train_samples_per_second': 3.2, 'train_steps_per_second': 0.2, 'total_flos': 1.1252790565109983e+18, 'train_loss': 1.0597990553941183,", true)}} ``` ## Environmental Impact Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** {{ hardware | default("[More Information Needed]", true)}} - **Hours used:** {{ hours_used | default("[More Information Needed]", true)}} - **Cloud Provider:** {{ cloud_provider | default("[More Information Needed]", true)}} - **Compute Region:** {{ cloud_region | default("[More Information Needed]", true)}} - **Carbon Emitted:** {{ co2_emitted | default("[More Information Needed]", true)}} ## Technical Specifications ### Model Architecture and Objective ```json PeftModelForCausalLM( (base_model): LoraModel( (model): FalconForCausalLM( (transformer): FalconModel( (word_embeddings): Embedding(65024, 4544) (h): ModuleList( (0-31): 32 x FalconDecoderLayer( (self_attention): FalconAttention( (maybe_rotary): FalconRotaryEmbedding() (query_key_value): Linear4bit( in_features=4544, out_features=4672, bias=False (lora_dropout): ModuleDict( (default): Dropout(p=0.05, inplace=False) ) (lora_A): ModuleDict( (default): Linear(in_features=4544, out_features=16, bias=False) ) (lora_B): ModuleDict( (default): Linear(in_features=16, out_features=4672, bias=False) ) (lora_embedding_A): ParameterDict() (lora_embedding_B): ParameterDict() ) (dense): Linear4bit(in_features=4544, out_features=4544, bias=False) (attention_dropout): Dropout(p=0.0, inplace=False) ) (mlp): FalconMLP( (dense_h_to_4h): Linear4bit(in_features=4544, out_features=18176, bias=False) (act): GELU(approximate='none') (dense_4h_to_h): Linear4bit(in_features=18176, out_features=4544, bias=False) ) (input_layernorm): LayerNorm((4544,), eps=1e-05, elementwise_affine=True) ) ) (ln_f): LayerNorm((4544,), eps=1e-05, elementwise_affine=True) ) (lm_head): Linear(in_features=4544, out_features=65024, bias=False) ) ) ) ``` ### Compute Infrastructure Google Collaboratory #### Hardware A100 ## Model Card Authors [Tonic](https://huggingface.co/tonic) ## Model Card Contact "[Tonic](https://huggingface.co/tonic)