aashish1904 commited on
Commit
3a658c2
·
verified ·
1 Parent(s): 619cde5

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +84 -0
README.md ADDED
@@ -0,0 +1,84 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+
4
+ license: mit
5
+ language:
6
+ - en
7
+ metrics:
8
+ - f1
9
+ tags:
10
+ - medical
11
+
12
+ ---
13
+
14
+ [![QuantFactory Banner](https://lh7-rt.googleusercontent.com/docsz/AD_4nXeiuCm7c8lEwEJuRey9kiVZsRn2W-b4pWlu3-X534V3YmVuVc2ZL-NXg2RkzSOOS2JXGHutDuyyNAUtdJI65jGTo8jT9Y99tMi4H4MqL44Uc5QKG77B0d6-JfIkZHFaUA71-RtjyYZWVIhqsNZcx8-OMaA?key=xt3VSDoCbmTY7o-cwwOFwQ)](https://hf.co/QuantFactory)
15
+
16
+
17
+ # QuantFactory/MentaLLaMA-chat-7B-GGUF
18
+ This is quantized version of [klyang/MentaLLaMA-chat-7B](https://huggingface.co/klyang/MentaLLaMA-chat-7B) created using llama.cpp
19
+
20
+ # Original Model Card
21
+
22
+
23
+ # Introduction
24
+
25
+ MentaLLaMA-chat-7B is part of the [MentaLLaMA](https://github.com/SteveKGYang/MentalLLaMA) project, the first open-source large language model (LLM) series for
26
+ interpretable mental health analysis with instruction-following capability. This model is finetuned based on the Meta LLaMA2-chat-7B foundation model and the full IMHI instruction tuning data.
27
+ The model is expected to make complex mental health analysis for various mental health conditions and give reliable explanations for each of its predictions.
28
+ It is fine-tuned on the IMHI dataset with 75K high-quality natural language instructions to boost its performance in downstream tasks.
29
+ We perform a comprehensive evaluation on the IMHI benchmark with 20K test samples. The result shows that MentalLLaMA approaches state-of-the-art discriminative
30
+ methods in correctness and generates high-quality explanations.
31
+
32
+ # Ethical Consideration
33
+
34
+ Although experiments on MentaLLaMA show promising performance on interpretable mental health analysis, we stress that
35
+ all predicted results and generated explanations should only used
36
+ for non-clinical research, and the help-seeker should get assistance
37
+ from professional psychiatrists or clinical practitioners. In addition,
38
+ recent studies have indicated LLMs may introduce some potential
39
+ bias, such as gender gaps. Meanwhile, some incorrect prediction results, inappropriate explanations, and over-generalization
40
+ also illustrate the potential risks of current LLMs. Therefore, there
41
+ are still many challenges in applying the model to real-scenario
42
+ mental health monitoring systems.
43
+
44
+ ## Other Models in MentaLLaMA
45
+
46
+ In addition to MentaLLaMA-chat-7B, the MentaLLaMA project includes another model: MentaLLaMA-chat-13B, MentalBART, MentalT5.
47
+
48
+ - **MentaLLaMA-chat-13B**: This model is finetuned based on the Meta LLaMA2-chat-13B foundation model and the full IMHI instruction tuning data. The training data covers 10 mental health analysis tasks.
49
+
50
+ - **MentalBART**: This model is finetuned based on the BART-large foundation model and the full IMHI-completion data. The training data covers 10 mental health analysis tasks. This model doesn't have instruction-following ability but is more lightweight and performs well in interpretable mental health analysis in a completion-based manner.
51
+
52
+ - **MentalT5**: This model is finetuned based on the T5-large foundation model and the full IMHI-completion data. The training data covers 10 mental health analysis tasks. This model doesn't have instruction-following ability but is more lightweight and performs well in interpretable mental health analysis in a completion-based manner.
53
+
54
+ ## Usage
55
+
56
+ You can use the MentaLLaMA-chat-7B model in your Python project with the Hugging Face Transformers library. Here is a simple example of how to load the model:
57
+
58
+ ```python
59
+ from transformers import LlamaTokenizer, LlamaForCausalLM
60
+ tokenizer = LlamaTokenizer.from_pretrained('klyang/MentaLLaMA-chat-7B')
61
+ model = LlamaForCausalLM.from_pretrained('klyang/MentaLLaMA-chat-7B', device_map='auto')
62
+ ```
63
+
64
+ In this example, LlamaTokenizer is used to load the tokenizer, and LlamaForCausalLM is used to load the model. The `device_map='auto'` argument is used to automatically
65
+ use the GPU if it's available.
66
+
67
+ ## License
68
+
69
+ MentaLLaMA-chat-7B is licensed under MIT. For more details, please see the MIT file.
70
+
71
+ ## Citation
72
+
73
+ If you use MentaLLaMA-chat-7B in your work, please cite the our paper:
74
+
75
+ ```bibtex
76
+ @misc{yang2023mentalllama,
77
+ title={MentalLLaMA: Interpretable Mental Health Analysis on Social Media with Large Language Models},
78
+ author={Kailai Yang and Tianlin Zhang and Ziyan Kuang and Qianqian Xie and Sophia Ananiadou},
79
+ year={2023},
80
+ eprint={2309.13567},
81
+ archivePrefix={arXiv},
82
+ primaryClass={cs.CL}
83
+ }
84
+ ```