sarath-shekkizhar commited on
Commit
268e444
1 Parent(s): 75d20ad

adding model card

Browse files
Files changed (1) hide show
  1. README.md +8 -3
README.md CHANGED
@@ -1,8 +1,13 @@
1
  ---
2
- license: {apache-2.0}
3
- base_model: {openchat/openchat_3.5}
 
 
 
 
 
 
4
  ---
5
-
6
  # TenyxChat: Language Model Alignment using Tenyx Fine-tuning
7
 
8
  Introducing TenyxChat, a series of ChatGPT-like models trained to function as useful assistants through preference tuning, using Tenyx's recently released advanced fine-tuning technology ([VentureBeat article](https://venturebeat.com/ai/tenyx-aims-to-fix-llms-catastrophic-forgetting-problem/)). Our first chat model in the series, TenyxChat-7B-v1, is trained using the [Direct Preference Optimization (DPO)](https://arxiv.org/abs/2305.18290) framework on the open-source AI feedback dataset [UltraFeedback](https://huggingface.co/datasets/HuggingFaceH4/ultrafeedback_binarized).
 
1
  ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ library_name: transformers
6
+ tags:
7
+ - tenyx-fine-tuning
8
+ - dpo
9
+ - tenyxchat
10
  ---
 
11
  # TenyxChat: Language Model Alignment using Tenyx Fine-tuning
12
 
13
  Introducing TenyxChat, a series of ChatGPT-like models trained to function as useful assistants through preference tuning, using Tenyx's recently released advanced fine-tuning technology ([VentureBeat article](https://venturebeat.com/ai/tenyx-aims-to-fix-llms-catastrophic-forgetting-problem/)). Our first chat model in the series, TenyxChat-7B-v1, is trained using the [Direct Preference Optimization (DPO)](https://arxiv.org/abs/2305.18290) framework on the open-source AI feedback dataset [UltraFeedback](https://huggingface.co/datasets/HuggingFaceH4/ultrafeedback_binarized).