Update README.md
Browse files
README.md
CHANGED
@@ -15,7 +15,7 @@ base_model:
|
|
15 |
</p> -->
|
16 |
|
17 |
|
18 |
-
# Aria-Chat
|
19 |
|
20 |
<!--
|
21 |
- Aria is the **first open multimodal native MoE** model, capable of seamlessly handling various input modalities within a MoE architecture.
|
@@ -25,7 +25,7 @@ base_model:
|
|
25 |
|
26 |
## Key features
|
27 |
|
28 |
-
- **Especially Optimized For Multimodal Chat**: Unlike previous models, Aria-Chat
|
29 |
- **Improved Stability**: We have especially improved its stability for long outputs, reducing probabilities for previously-reported bad cases like incomplete responses on Markdown tables, or endless responses on listwise outputs.
|
30 |
- **Better Multi-Lingual Abilities**: We have optimized its ability on non-English scenarios (Chinese, Spanish, French, Japanese, *etc*), including both multi-lingual OCR and multi-lingual dialogs.
|
31 |
|
@@ -49,7 +49,7 @@ This checkpoint is not designed for benchmarks, but for real-world open-ended ap
|
|
49 |
| Model | Score |
|
50 |
|---------------------------|---------|
|
51 |
| gpt-4o | 89.15 |
|
52 |
-
| **Aria-Chat
|
53 |
| gpt-4-vision-preview | 79.78 |
|
54 |
| Aria | 74.1 |
|
55 |
| Reka-Flash | 64.65 |
|
@@ -81,7 +81,7 @@ import torch
|
|
81 |
from PIL import Image
|
82 |
from transformers import AutoModelForCausalLM, AutoProcessor
|
83 |
|
84 |
-
model_id_or_path = "rhymes-ai/Aria-Chat
|
85 |
|
86 |
model = AutoModelForCausalLM.from_pretrained(model_id_or_path, device_map="auto", torch_dtype=torch.bfloat16, trust_remote_code=True)
|
87 |
|
|
|
15 |
</p> -->
|
16 |
|
17 |
|
18 |
+
# Aria-Chat Model Card
|
19 |
|
20 |
<!--
|
21 |
- Aria is the **first open multimodal native MoE** model, capable of seamlessly handling various input modalities within a MoE architecture.
|
|
|
25 |
|
26 |
## Key features
|
27 |
|
28 |
+
- **Especially Optimized For Multimodal Chat**: Unlike previous models, Aria-Chat is especially optimized for open-ended dialogs. We hope this version can provide seamless open-source multimodal chat experience.
|
29 |
- **Improved Stability**: We have especially improved its stability for long outputs, reducing probabilities for previously-reported bad cases like incomplete responses on Markdown tables, or endless responses on listwise outputs.
|
30 |
- **Better Multi-Lingual Abilities**: We have optimized its ability on non-English scenarios (Chinese, Spanish, French, Japanese, *etc*), including both multi-lingual OCR and multi-lingual dialogs.
|
31 |
|
|
|
49 |
| Model | Score |
|
50 |
|---------------------------|---------|
|
51 |
| gpt-4o | 89.15 |
|
52 |
+
| **Aria-Chat** |**81.3** |
|
53 |
| gpt-4-vision-preview | 79.78 |
|
54 |
| Aria | 74.1 |
|
55 |
| Reka-Flash | 64.65 |
|
|
|
81 |
from PIL import Image
|
82 |
from transformers import AutoModelForCausalLM, AutoProcessor
|
83 |
|
84 |
+
model_id_or_path = "rhymes-ai/Aria-Chat"
|
85 |
|
86 |
model = AutoModelForCausalLM.from_pretrained(model_id_or_path, device_map="auto", torch_dtype=torch.bfloat16, trust_remote_code=True)
|
87 |
|