Spaces:
Running
Running
Update README.md
Browse filesAdded links to the other orgs (although the comparisons doesn't work yet, since there's only one comparison, I think?)
README.md
CHANGED
@@ -1,10 +1,20 @@
|
|
1 |
---
|
2 |
title: README
|
3 |
-
emoji:
|
4 |
-
colorFrom:
|
5 |
-
colorTo:
|
6 |
sdk: static
|
7 |
pinned: false
|
|
|
|
|
|
|
8 |
---
|
9 |
|
10 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
title: README
|
3 |
+
emoji: π€
|
4 |
+
colorFrom: green
|
5 |
+
colorTo: purple
|
6 |
sdk: static
|
7 |
pinned: false
|
8 |
+
tags:
|
9 |
+
- evaluate
|
10 |
+
- measurement
|
11 |
---
|
12 |
|
13 |
+
π€ Evaluate provides access to a wide range of evaluation tools. It covers a range of modalities such as text, computer vision, audio, etc. as well as tools to evaluate models or datasets.
|
14 |
+
|
15 |
+
It has three types of evaluations:
|
16 |
+
- **Metric**: measures the performance of a model on a given dataset, usually by comparing the model's predictions to some ground truth labels -- these are covered in this space.
|
17 |
+
- **Comparison**: used useful to compare the performance of two or more models on a single test dataset., e.g. by comparing their predictions to ground truth labels and computing their agreement -- covered in the [Evaluate Comparison](https://huggingface.co/spaces/evaluate-comparison) Spaces.
|
18 |
+
- **Measurement**: for gaining more insights on datasets and model predictions based on their properties and characteristics -- covered in the [Evaluate Measurement](https://huggingface.co/evaluate-measurement) Spaces.
|
19 |
+
|
20 |
+
All three types of evaluation supported by the π€ Evaluate library are meant to be mutually complementary, and help our community carry out more mindful and responsible evaluation!
|