TheBloke commited on
Commit
7a677dc
1 Parent(s): 49afe19

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +48 -9
README.md CHANGED
@@ -5,7 +5,17 @@ license: apache-2.0
5
  model_creator: Intel
6
  model_name: Neural Chat 7B v3-1
7
  model_type: mistral
8
- prompt_template: '{prompt}
 
 
 
 
 
 
 
 
 
 
9
 
10
  '
11
  quantized_by: TheBloke
@@ -53,11 +63,17 @@ These files were quantised using hardware kindly provided by [Massed Compute](ht
53
  <!-- repositories-available end -->
54
 
55
  <!-- prompt-template start -->
56
- ## Prompt template: Unknown
57
 
58
  ```
 
 
 
 
59
  {prompt}
60
 
 
 
61
  ```
62
 
63
  <!-- prompt-template end -->
@@ -226,7 +242,13 @@ from huggingface_hub import InferenceClient
226
  endpoint_url = "https://your-endpoint-url-here"
227
 
228
  prompt = "Tell me about AI"
229
- prompt_template=f'''{prompt}
 
 
 
 
 
 
230
  '''
231
 
232
  client = InferenceClient(endpoint_url)
@@ -282,7 +304,13 @@ model = AutoModelForCausalLM.from_pretrained(model_name_or_path,
282
  tokenizer = AutoTokenizer.from_pretrained(model_name_or_path, use_fast=True)
283
 
284
  prompt = "Tell me about AI"
285
- prompt_template=f'''{prompt}
 
 
 
 
 
 
286
  '''
287
 
288
  print("\n\n*** Generate:")
@@ -357,9 +385,9 @@ And thank you again to a16z for their generous grant.
357
  # Original model card: Intel's Neural Chat 7B v3-1
358
 
359
 
360
- ## Finetuning on [habana](https://habana.ai/) HPU
361
 
362
- This model is a fine-tuned model based on [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on the open source dataset [Open-Orca/SlimOrca](https://huggingface.co/datasets/Open-Orca/SlimOrca). Then we align it with DPO algorithm. For more details, you can refer our blog: [NeuralChat: Simplifying Supervised Instruction Fine-Tuning and Reinforcement Aligning](https://medium.com/intel-analytics-software/neuralchat-simplifying-supervised-instruction-fine-tuning-and-reinforcement-aligning-for-chatbots-d034bca44f69) and [The Practice of Supervised Fine-tuning and Direct Preference Optimization on Habana Gaudi2](https://medium.com/@NeuralCompressor/the-practice-of-supervised-finetuning-and-direct-preference-optimization-on-habana-gaudi2-a1197d8a3cd3).
363
 
364
  ## Model date
365
  Neural-chat-7b-v3-1 was trained between September and October, 2023.
@@ -389,10 +417,22 @@ The following hyperparameters were used during training:
389
  - total_train_batch_size: 64
390
  - total_eval_batch_size: 8
391
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
392
- - lr_scheduler_type: linear
393
- - lr_scheduler_warmup_ratio: 0.02
394
  - num_epochs: 2.0
395
 
 
 
 
 
 
 
 
 
 
 
 
 
396
  ## Inference with transformers
397
 
398
  ```shell
@@ -418,4 +458,3 @@ The NeuralChat team with members from Intel/SATG/AIA/AIPT. Core team members: Ka
418
  ## Useful links
419
  * Intel Neural Compressor [link](https://github.com/intel/neural-compressor)
420
  * Intel Extension for Transformers [link](https://github.com/intel/intel-extension-for-transformers)
421
- * Intel Extension for PyTorch [link](https://github.com/intel/intel-extension-for-pytorch)
 
5
  model_creator: Intel
6
  model_name: Neural Chat 7B v3-1
7
  model_type: mistral
8
+ prompt_template: '### System:
9
+
10
+ {system_message}
11
+
12
+
13
+ ### User:
14
+
15
+ {prompt}
16
+
17
+
18
+ ### Assistant:
19
 
20
  '
21
  quantized_by: TheBloke
 
63
  <!-- repositories-available end -->
64
 
65
  <!-- prompt-template start -->
66
+ ## Prompt template: Orca-Hashes
67
 
68
  ```
69
+ ### System:
70
+ {system_message}
71
+
72
+ ### User:
73
  {prompt}
74
 
75
+ ### Assistant:
76
+
77
  ```
78
 
79
  <!-- prompt-template end -->
 
242
  endpoint_url = "https://your-endpoint-url-here"
243
 
244
  prompt = "Tell me about AI"
245
+ prompt_template=f'''### System:
246
+ {system_message}
247
+
248
+ ### User:
249
+ {prompt}
250
+
251
+ ### Assistant:
252
  '''
253
 
254
  client = InferenceClient(endpoint_url)
 
304
  tokenizer = AutoTokenizer.from_pretrained(model_name_or_path, use_fast=True)
305
 
306
  prompt = "Tell me about AI"
307
+ prompt_template=f'''### System:
308
+ {system_message}
309
+
310
+ ### User:
311
+ {prompt}
312
+
313
+ ### Assistant:
314
  '''
315
 
316
  print("\n\n*** Generate:")
 
385
  # Original model card: Intel's Neural Chat 7B v3-1
386
 
387
 
388
+ ## Fine-tuning on [Habana](https://habana.ai/) Gaudi2
389
 
390
+ This model is a fine-tuned model based on [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on the open source dataset [Open-Orca/SlimOrca](https://huggingface.co/datasets/Open-Orca/SlimOrca). Then we align it with DPO algorithm. For more details, you can refer our blog: [The Practice of Supervised Fine-tuning and Direct Preference Optimization on Habana Gaudi2](https://medium.com/@NeuralCompressor/the-practice-of-supervised-finetuning-and-direct-preference-optimization-on-habana-gaudi2-a1197d8a3cd3).
391
 
392
  ## Model date
393
  Neural-chat-7b-v3-1 was trained between September and October, 2023.
 
417
  - total_train_batch_size: 64
418
  - total_eval_batch_size: 8
419
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
420
+ - lr_scheduler_type: cosine
421
+ - lr_scheduler_warmup_ratio: 0.03
422
  - num_epochs: 2.0
423
 
424
+ ## Prompt Template
425
+
426
+ ```
427
+ ### System:
428
+ {system}
429
+ ### User:
430
+ {usr}
431
+ ### Assistant:
432
+
433
+ ```
434
+
435
+
436
  ## Inference with transformers
437
 
438
  ```shell
 
458
  ## Useful links
459
  * Intel Neural Compressor [link](https://github.com/intel/neural-compressor)
460
  * Intel Extension for Transformers [link](https://github.com/intel/intel-extension-for-transformers)