GGUF
fakezeta commited on
Commit
5e68fc7
1 Parent(s): 698935c

Upload 2 files

Browse files
Files changed (3) hide show
  1. .gitattributes +1 -0
  2. README.md +76 -0
  3. neural-chat-7b-v3-1_Q5_K_M.gguf +3 -0
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ neural-chat-7b-v3-1_Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,3 +1,79 @@
1
  ---
2
  license: apache-2.0
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
  ---
4
+
5
+ neural-chat-7b-v3-1 - GGUF
6
+
7
+ Model creator: [Intel](https://huggingface.co/Intel)
8
+ Original model: [neural-chat-7b-v3-1](https://huggingface.co/Intel/neural-chat-7b-v3-1)
9
+
10
+ Description
11
+
12
+ This repo contains GGUF format model files for Intel's neural-chat-7b-v3-1.
13
+
14
+ These files were quantised with Q5_K_M.
15
+
16
+ ## Original Readme from Intel
17
+
18
+ ## Finetuning on [habana](https://habana.ai/) HPU
19
+
20
+ This model is a fine-tuned model based on [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on the open source dataset [Open-Orca/SlimOrca](https://huggingface.co/datasets/Open-Orca/SlimOrca). Then we align it with DPO algorithm. For more details, you can refer our blog: [NeuralChat: Simplifying Supervised Instruction Fine-Tuning and Reinforcement Aligning](https://medium.com/intel-analytics-software/neuralchat-simplifying-supervised-instruction-fine-tuning-and-reinforcement-aligning-for-chatbots-d034bca44f69).
21
+
22
+ ## Model date
23
+ Neural-chat-7b-v3 was trained between September and October, 2023.
24
+
25
+ ## Evaluation
26
+
27
+ We submit our model to [open_llm_leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard), and the model performance has been **improved significantly** as we see from the average metric of 7 tasks from the leaderboard.
28
+
29
+ | Model | Average ⬆️| ARC (25-s) ⬆️ | HellaSwag (10-s) ⬆️ | MMLU (5-s) ⬆️| TruthfulQA (MC) (0-s) ⬆️ | Winogrande (5-s) | GSM8K (5-s) | DROP (3-s) |
30
+ | --- | --- | --- | --- | --- | --- | --- | --- | --- |
31
+ |[mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) | 50.32 | 59.58 | 83.31 | 64.16 | 42.15 | 78.37 | 18.12 | 6.14 |
32
+ | [Intel/neural-chat-7b-v3](https://huggingface.co/Intel/neural-chat-7b-v3) | **57.31** | 67.15 | 83.29 | 62.26 | 58.77 | 78.06 | 1.21 | 50.43 |
33
+ | [Intel/neural-chat-7b-v3-1](https://huggingface.co/Intel/neural-chat-7b-v3-1) | **59.06** | 66.21 | 83.64 | 62.37 | 59.65 | 78.14 | 19.56 | 43.84 |
34
+
35
+ ## Training procedure
36
+
37
+ ### Training hyperparameters
38
+
39
+ The following hyperparameters were used during training:
40
+ - learning_rate: 1e-04
41
+ - train_batch_size: 1
42
+ - eval_batch_size: 2
43
+ - seed: 42
44
+ - distributed_type: multi-HPU
45
+ - num_devices: 8
46
+ - gradient_accumulation_steps: 8
47
+ - total_train_batch_size: 64
48
+ - total_eval_batch_size: 8
49
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
50
+ - lr_scheduler_type: linear
51
+ - lr_scheduler_warmup_ratio: 0.02
52
+ - num_epochs: 2.0
53
+
54
+ ## Inference with transformers
55
+
56
+ ```shell
57
+ import transformers
58
+ model = transformers.AutoModelForCausalLM.from_pretrained(
59
+ 'Intel/neural-chat-7b-v3'
60
+ )
61
+ ```
62
+
63
+ ## Ethical Considerations and Limitations
64
+ neural-chat-7b-v3 can produce factually incorrect output, and should not be relied on to produce factually accurate information. neural-chat-7b-v3 was trained on [Open-Orca/SlimOrca](https://huggingface.co/datasets/Open-Orca/SlimOrca) based on [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1). Because of the limitations of the pretrained model and the finetuning datasets, it is possible that this model could generate lewd, biased or otherwise offensive outputs.
65
+
66
+ Therefore, before deploying any applications of neural-chat-7b-v3, developers should perform safety testing.
67
+
68
+ ## Disclaimer
69
+
70
+ The license on this model does not constitute legal advice. We are not responsible for the actions of third parties who use this model. Please cosult an attorney before using this model for commercial purposes.
71
+
72
+ ## Organizations developing the model
73
+
74
+ The NeuralChat team with members from Intel/SATG/AIA/AIPT. Core team members: Kaokao Lv, Liang Lv, Chang Wang, Wenxin Zhang, Xuhui Ren, and Haihao Shen.
75
+
76
+ ## Useful links
77
+ * Intel Neural Compressor [link](https://github.com/intel/neural-compressor)
78
+ * Intel Extension for Transformers [link](https://github.com/intel/intel-extension-for-transformers)
79
+ * Intel Extension for PyTorch [link](https://github.com/intel/intel-extension-for-pytorch)
neural-chat-7b-v3-1_Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71ccc0ba42e2ac86f593d7905f9be865e3870496ade370537ae1db0c13637bc1
3
+ size 5131409024