huu-ontocord commited on
Commit
bdabe7c
1 Parent(s): db9a90b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +22 -3
README.md CHANGED
@@ -1,7 +1,26 @@
1
  ---
2
- library_name: transformers
3
- tags: []
 
4
  ---
5
- This is an experiment in modifying the ``</s>`` and ``<s>`` tags of the original Felix-8b DPO model to fix the issue of being too verbose. Use the original tokenizer for Felix-8b. We will push the tokenizer shortly.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6
 
7
 
 
1
  ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
  ---
6
+ April 17, 2024
7
+
8
+ # Felix-8B-v2: A model built with lawfulness alignment
9
+
10
+ Felix-8B-v2 is an experimental language model developed by Ontocord.ai, specializing in addressing lawfulness concerns under the Biden-Harris Executive Order on AI and the principles of the EU AI Act. This model has achieved one of the highest scores on the TruthfulQA benchmark compared to models of its size, showcasing its exceptional performance in providing accurate and reliable responses.
11
+ Felix-8B-v2 is **experimental and a research work product** and a DPO reinforcement learning version of [ontocord/sft-4e-exp2](https://huggingface.co/ontocord/sft-4e-exp2) which in turn is a fine-tuned version of [TencentARC/Mistral_Pro_8B_v0.1](https://huggingface.co/TencentARC/Mistral_Pro_8B_v0.1).
12
+
13
+
14
+ This model is exactly the same as [Felix-8B](https://huggingface.co/ontocord/Felix-8B) except we modified the ``</s>`` and ``<s>`` tags of the original Felix-8b DPO model to fix the issue of being too verbose.
15
+
16
+
17
+ **Please give feedback in the Community section. If you find any issues please let us know in the Community section so we can improve the model.**
18
+
19
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/630430583926de1f7ec62c6b/zKoEI-B21KnmeRWvKzrx1.png)
20
+
21
+ ## Model Description
22
+
23
+ Felix-8B is an 8 billion parameter language model trained using Ontocord.ai's proprietary auto-purpleteaming technique. The model has been fine-tuned and optimized using synthetic data, with the goal of improving its robustness and ability to handle a wide range of tasks while maintaining a strong focus on safety and truthfulness.
24
+
25
 
26