JordiBayarri commited on
Commit
4062100
1 Parent(s): 5693173

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -5
README.md CHANGED
@@ -82,11 +82,9 @@ The Beta model has been developed to excel in several different medical tasks. F
82
 
83
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6620f941eba5274b5c12f83d/2NW3im0aH2u6RKp969sjx.png)
84
 
85
- We also compared the performance of the model in the general domain, using the OpenLLM Leaderboard benchmark:
86
 
87
- TO BE UPDATED
88
-
89
- ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6620f941eba5274b5c12f83d/Ym6v3LsMdfwetXbg6twQP.png)
90
 
91
  ## Uses
92
 
@@ -276,7 +274,7 @@ The model is aligned using the Direct Preference Optimization (DPO) technique th
276
  2. Red-Teaming Alignment: This step further fine-tunes the model to resist a variety of potential attacks, enhancing its robustness and security. Dataset will be shared soon. In this stage, we set the learning rate to 1e-7.
277
 
278
  <!---
279
- ^^^ LINKS TO DPO DATA ^^^
280
  -->
281
 
282
 
 
82
 
83
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6620f941eba5274b5c12f83d/2NW3im0aH2u6RKp969sjx.png)
84
 
85
+ We also compared the performance of the model in the general domain, using the OpenLLM Leaderboard benchmark. Aloe-Beta gets competitive results with the current SOTA general models in the most used general benchmarks and outperforms the medical models:
86
 
87
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6620f941eba5274b5c12f83d/zbT3BzclSegfh6p2hC-3a.png)
 
 
88
 
89
  ## Uses
90
 
 
274
  2. Red-Teaming Alignment: This step further fine-tunes the model to resist a variety of potential attacks, enhancing its robustness and security. Dataset will be shared soon. In this stage, we set the learning rate to 1e-7.
275
 
276
  <!---
277
+ ^^^ LINKS TO DPO DATA (DPO added, missing the RT^^^
278
  -->
279
 
280