ignacioct's picture
erasing interactive and update intervals
d543e01
"""
Dashboard to visualize the progress of the SomosNLP project.
by Argilla.
This dashboard shows the progress of the SomosNLP project, including the number of annotated and pending records, the top annotators, and the remaining records to be annotated.
The data is fetched from the source datasets and updated every 5 minutes.
Due to Gradio's limitation on what can be passed as input to their graph methods, the data is fetched outside of the graph methods and stored in global variables. Therefore,
a function for each graph-dataset tuple is needed. Moreover, to also avoid circular imports, all the functions must be
in the same Python file. This behavior is not ideal, and could be improved knowing how to pass input parameter to graph functions in Gradio.
"""
import datetime
import os
from typing import Dict, List, Tuple
from uuid import UUID
import altair as alt
from apscheduler.schedulers.background import BackgroundScheduler
import argilla as rg
from argilla.feedback import FeedbackDataset
from argilla.client.feedback.dataset.remote.dataset import RemoteFeedbackDataset
import gradio as gr
import pandas as pd
def get_source_datasets() -> Tuple[
FeedbackDataset | RemoteFeedbackDataset,
FeedbackDataset | RemoteFeedbackDataset,
FeedbackDataset | RemoteFeedbackDataset,
]:
"""
This function returns the source datasets to be showed in the visualization. The datasets names
and the workspace name is obtained from the environment variables.
Returns:
A tuple with the three source datasets
"""
return (
rg.FeedbackDataset.from_argilla(
os.getenv("SOURCE_DATASET_1"), workspace=os.getenv("SOURCE_WORKSPACE")
),
rg.FeedbackDataset.from_argilla(
os.getenv("SOURCE_DATASET_2"), workspace=os.getenv("SOURCE_WORKSPACE")
),
rg.FeedbackDataset.from_argilla(
os.getenv("SOURCE_DATASET_3"), workspace=os.getenv("SOURCE_WORKSPACE")
),
)
def get_user_annotations_dictionary(
datasets: List[FeedbackDataset | RemoteFeedbackDataset],
) -> Dict[str, int]:
"""
This function returns a dictionary with the username as the key and the number of annotations as the value.
All annotationsfrom all datasets are introduced in the same dictionary.
Args:
datasets: A list with the datasets to be used to obtain the annotations and the annotators.
Returns:
A dictionary with the username as the key and the number of annotations as the value.
"""
output = {}
for dataset in datasets:
for record in dataset:
for response in record.responses:
if str(response.user_id) not in output.keys():
output[str(response.user_id)] = 1
else:
output[str(response.user_id)] += 1
# Changing the name of the keys, from the id to the username
for key in list(output.keys()):
output[rg.User.from_id(UUID(key)).username] = output.pop(key)
return output
def donut_chart_1() -> alt.Chart:
"""
This function returns a donut chart with the number of annotated and pending records, for the first dataset
Returns:
An altair chart with the donut chart.
"""
annotated_records = len(dataset1.filter_by(response_status=["submitted"]))
pending_records = len(dataset1) - annotated_records
source = pd.DataFrame(
{
"values": [annotated_records, pending_records],
"category": ["Annotated", "Pending"], # Add a new column for categories
}
)
base = alt.Chart(source).encode(
theta=alt.Theta("values:Q", stack=True),
radius=alt.Radius(
"values", scale=alt.Scale(type="sqrt", zero=True, rangeMin=20)
),
color=alt.Color("category:N", legend=alt.Legend(title="Category")),
)
c1 = base.mark_arc(innerRadius=20, stroke="#fff")
c2 = base.mark_text(radiusOffset=10).encode(text="values:Q")
chart = c1 + c2
return chart
def donut_chart_2() -> alt.Chart:
"""
This function returns a donut chart with the number of annotated and pending records, for the second dataset.
Returns:
An altair chart with the donut chart.
"""
annotated_records = len(dataset2.filter_by(response_status=["submitted"]))
pending_records = len(dataset2) - annotated_records
source = pd.DataFrame(
{
"values": [annotated_records, pending_records],
"category": ["Annotated", "Pending"], # Add a new column for categories
}
)
base = alt.Chart(source).encode(
theta=alt.Theta("values:Q", stack=True),
radius=alt.Radius(
"values", scale=alt.Scale(type="sqrt", zero=True, rangeMin=20)
),
color=alt.Color("category:N", legend=alt.Legend(title="Category")),
)
c1 = base.mark_arc(innerRadius=20, stroke="#fff")
c2 = base.mark_text(radiusOffset=10).encode(text="values:Q")
chart = c1 + c2
return chart
def donut_chart_3() -> alt.Chart:
"""
This function returns a donut chart with the number of annotated and pending records, for the third dataset.
Returns:
An altair chart with the donut chart.
"""
annotated_records = len(dataset3.filter_by(response_status=["submitted"]))
pending_records = len(dataset3) - annotated_records
source = pd.DataFrame(
{
"values": [annotated_records, pending_records],
"category": ["Annotated", "Pending"], # Add a new column for categories
}
)
base = alt.Chart(source).encode(
theta=alt.Theta("values:Q", stack=True),
radius=alt.Radius(
"values", scale=alt.Scale(type="sqrt", zero=True, rangeMin=20)
),
color=alt.Color("category:N", legend=alt.Legend(title="Category")),
)
c1 = base.mark_arc(innerRadius=20, stroke="#fff")
c2 = base.mark_text(radiusOffset=10).encode(text="values:Q")
chart = c1 + c2
return chart
def kpi_chart_submitted_1() -> alt.Chart:
"""
This function returns a KPI chart with the total amount of records that have been annotated, for the first dataset.
Returns:
An altair chart with the KPI chart.
"""
total = len(dataset1.filter_by(response_status=["submitted"]))
# Assuming you have a DataFrame with user data, create a sample DataFrame
data = pd.DataFrame({"Category": ["Total completed"], "Value": [total]})
# Create Altair chart
chart = (
alt.Chart(data)
.mark_text(fontSize=100, align="center", baseline="middle", color="steelblue")
.encode(text="Value:N")
.properties(title="Completados", width=250, height=200)
)
return chart
def kpi_chart_submitted_2() -> alt.Chart:
"""
This function returns a KPI chart with the total amount of records that have been annotated, for the second dataset.
Returns:
An altair chart with the KPI chart.
"""
total = len(dataset2.filter_by(response_status=["submitted"]))
# Assuming you have a DataFrame with user data, create a sample DataFrame
data = pd.DataFrame({"Category": ["Total completed"], "Value": [total]})
# Create Altair chart
chart = (
alt.Chart(data)
.mark_text(fontSize=100, align="center", baseline="middle", color="steelblue")
.encode(text="Value:N")
.properties(title="Completados", width=250, height=200)
)
return chart
def kpi_chart_submitted_3() -> alt.Chart:
"""
This function returns a KPI chart with the total amount of records that have been annotated, for the third dataset.
Returns:
An altair chart with the KPI chart.
"""
total = len(dataset3.filter_by(response_status=["submitted"]))
# Assuming you have a DataFrame with user data, create a sample DataFrame
data = pd.DataFrame({"Category": ["Total completed"], "Value": [total]})
# Create Altair chart
chart = (
alt.Chart(data)
.mark_text(fontSize=100, align="center", baseline="middle", color="steelblue")
.encode(text="Value:N")
.properties(title="Completados", width=250, height=200)
)
return chart
def kpi_chart_remaining_1() -> alt.Chart:
"""
This function returns a KPI chart with the remaining amount of records to be annotated, for the first dataset.
Returns:
An altair chart with the KPI chart.
"""
annotated_records = len(dataset1.filter_by(response_status=["submitted"]))
pending_records = len(dataset1) - annotated_records
# Assuming you have a DataFrame with user data, create a sample DataFrame
data = pd.DataFrame({"Category": ["Total remaining"], "Value": [pending_records]})
# Create Altair chart
chart = (
alt.Chart(data)
.mark_text(fontSize=100, align="center", baseline="middle", color="steelblue")
.encode(text="Value:N")
.properties(title="Restantes", width=250, height=200)
)
return chart
def kpi_chart_remaining_2() -> alt.Chart:
"""
This function returns a KPI chart with the remaining amount of records to be annotated, for the second dataset.
Returns:
An altair chart with the KPI chart.
"""
annotated_records = len(dataset2.filter_by(response_status=["submitted"]))
pending_records = len(dataset2) - annotated_records
# Assuming you have a DataFrame with user data, create a sample DataFrame
data = pd.DataFrame({"Category": ["Total remaining"], "Value": [pending_records]})
# Create Altair chart
chart = (
alt.Chart(data)
.mark_text(fontSize=100, align="center", baseline="middle", color="steelblue")
.encode(text="Value:N")
.properties(title="Restantes", width=250, height=200)
)
return chart
def kpi_chart_remaining_3() -> alt.Chart:
"""
This function returns a KPI chart with the remaining amount of records to be annotated, for the third dataset.
Returns:
An altair chart with the KPI chart.
"""
annotated_records = len(dataset3.filter_by(response_status=["submitted"]))
pending_records = len(dataset3) - annotated_records
# Assuming you have a DataFrame with user data, create a sample DataFrame
data = pd.DataFrame({"Category": ["Total remaining"], "Value": [pending_records]})
# Create Altair chart
chart = (
alt.Chart(data)
.mark_text(fontSize=100, align="center", baseline="middle", color="steelblue")
.encode(text="Value:N")
.properties(title="Restantes", width=250, height=200)
)
return chart
def render_hub_user_link(hub_id: str) -> str:
"""
This function formats the username with a link to the user's profile in the Hugging Face Hub.
Args:
hub_id: The user's id in the Hugging Face Hub.
Returns:
A string with the username formatted as a link to the user's profile in the Hugging Face Hub.
"""
link = f"https://huggingface.co/{hub_id}"
return f'<a target="_blank" href="{link}" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">{hub_id}</a>'
def kpi_chart_annotators() -> alt.Chart:
"""
This function returns a KPI chart with the total amount of annotators.
Returns:
An altair chart with the KPI chart.
"""
# Obtain the total amount of annotators
total_annotators = len(user_ids_annotations)
# Assuming you have a DataFrame with user data, create a sample DataFrame
data = pd.DataFrame(
{"Category": ["Total Contributors"], "Value": [total_annotators]}
)
# Create Altair chart
chart = (
alt.Chart(data)
.mark_text(fontSize=100, align="center", baseline="middle", color="steelblue")
.encode(text="Value:N")
.properties(title="Contribuidores Totales", width=250, height=200)
)
return chart
def obtain_top_users(user_ids_annotations: Dict[str, int]) -> pd.DataFrame:
"""
This function returns the top 50 users with the most annotations. The usernames are formatted as links to the user's profile in the Hugging Face Hub.
Args:
user_ids_annotations: A dictionary with the user ids as the key and the number of annotations as the value.
Returns:
A pandas dataframe with the top 5 users with the most annotations.
"""
dataframe = pd.DataFrame(
user_ids_annotations.items(), columns=["Name", "Submitted Responses"]
)
dataframe["Name"] = dataframe["Name"].apply(render_hub_user_link)
dataframe = dataframe.sort_values(by="Submitted Responses", ascending=False)
# Renaming the df columns to Spanish
dataframe.columns = ["Nombre", "Respuestas Enviadas"]
return dataframe.head(50)
def get_top() -> pd.DataFrame:
"""
This function returns the top users with the most annotations. The usernames are formatted as links to the user's profile in the Hugging Face Hub.
Returns:
A pandas dataframe with the top users with the most annotations.
"""
return obtain_top_users(user_ids_annotations)
def fetch_data() -> None:
"""
This function fetches the data from the source datasets and updates the global variables.
"""
print(f"Starting to fetch data: {datetime.datetime.now()}")
# Load the dataset as global variable to be able to use it in all Gradio graph methods,
# as they usually do not allow arguments.
global dataset1, dataset2, dataset3, user_ids_annotations
dataset1, dataset2, dataset3 = get_source_datasets()
user_ids_annotations = get_user_annotations_dictionary(
[dataset1, dataset2, dataset3]
)
# Print the current date and time
print(f"Data fetched: {datetime.datetime.now()}")
def main() -> None:
# Set the update interval
update_interval = 300 # seconds
update_interval_charts = 30 # seconds
# Connect to the space with rg.init()
rg.init(
api_url=os.getenv("ARGILLA_API_URL"),
api_key=os.getenv("ARGILLA_API_KEY"),
extra_headers={"Authorization": f"Bearer {os.getenv('HF_TOKEN')}"},
)
# Initial data fetching
fetch_data()
# To avoid the orange border for the Gradio elements that are in constant loading
css = """
.generating {
border: none;
}
"""
with gr.Blocks(css=css, title="LLM Benchmark en Español Dashboard") as demo:
# JSS code to force light theme
demo.load(
None,
None,
js="""
() => {
const params = new URLSearchParams(window.location.search);
if (!params.has('__theme')) {
params.set('__theme', 'light');
window.location.search = params.toString();
}
}""",
)
gr.Markdown(
"""
# 🗣️ SomosNLP LLM Benchmark en Español Dashboard
Esta dashboard, desarrollada en Gradio, muestra el progreso de anotación del Hackathon de LLM en Español, de SomosNLP. En este espacio, puedes valorar traducciones realizadas automática del inglés al español. Tus contribuciones y las del resto de anotadores aparecerán en esta visualización. Si quieres contribuir, puedes anotar siguiendo este [enlace](https://huggingface.co/spaces/somosnlp/benchmark-annotation-argilla).
"""
)
gr.Markdown(
f"""
## 🚀 Progreso del dataset {os.getenv("SOURCE_DATASET_1")}
El dataset ARC-C, creado por AllenAI, contiene un total de 2585 filas con preguntas de nivel escolar, de múltiples opciones, con el fin de evaluar Large Language Models (LLMs) en tareas de question-answering.
"""
)
with gr.Row():
plot = gr.Plot(label="Plot")
demo.load(
kpi_chart_submitted_1,
inputs=[],
outputs=[plot],
)
plot = gr.Plot(label="Plot")
demo.load(
kpi_chart_remaining_1,
inputs=[],
outputs=[plot],
)
# donut_chart_plotted_1 = gr.Plot(label="Plot")
# demo.load(
# donut_chart_1,
# inputs=[],
# outputs=[donut_chart_plotted_1],
# )
gr.Markdown(
f"""
## 🚀 Progreso del dataset {os.getenv("SOURCE_DATASET_2")}
HellaSwag es un conjunto de datos creado por Allen Institute for Artificial Intelligence (AI2) el cual consta de un contexto y de una serie de generaciones como continuación a dicho contexto, que suelen ser tareas de sentido común y fáciles de resolver por humanos.
"""
)
with gr.Row():
plot = gr.Plot(label="Plot")
demo.load(
kpi_chart_submitted_2,
inputs=[],
outputs=[plot],
)
plot = gr.Plot(label="Plot")
demo.load(
kpi_chart_remaining_2,
inputs=[],
outputs=[plot],
)
# donut_chart_plotted_2 = gr.Plot(label="Plot")
# demo.load(
# donut_chart_2,
# inputs=[],
# outputs=[donut_chart_plotted_2],
# )
gr.Markdown(
f"""
## 🚀 Progreso del dataset {os.getenv("SOURCE_DATASET_3")}
"Measuring Massive Multitask Language Understanding", abreviado como MMLU, es un conjunto de datos creado por Center for AI Safety que contiene preguntas con múltiples opciones de una gran diversidad de temas, llegando a cubrir hasta 57 temas o tareas distintas. De este modo, este conjunto de datos es utilizado para evaluar Large Language Models (LLMs), de modo que para que uno de estos modelos sea capaz de obtener un mejor resultado, ha de tener un conocimiento extenso y capacidad de resolución de problemas.
"""
)
with gr.Row():
plot = gr.Plot(label="Plot")
demo.load(
kpi_chart_submitted_3,
inputs=[],
outputs=[plot],
)
plot = gr.Plot(label="Plot")
demo.load(
kpi_chart_remaining_3,
inputs=[],
outputs=[plot],
)
# donut_chart_plotted_3 = gr.Plot(label="Plot")
# demo.load(
# donut_chart_3,
# inputs=[],
# outputs=[donut_chart_plotted_3],
# )
gr.Markdown(
"""
## 👾 Hall de la Fama
Aquí puedes ver el número de contribuidores y los contribuidores con más contribuciones:
"""
)
with gr.Row():
plot2 = gr.Plot(label="Plot")
demo.load(
kpi_chart_annotators,
inputs=[],
outputs=[plot2],
)
top_df_plot = gr.Dataframe(
headers=["Name", "Submitted Responses"],
datatype=[
"markdown",
"number",
],
row_count=50,
col_count=(2, "fixed"),
interactive=False,
)
demo.load(get_top, None, [top_df_plot])
# Launch the Gradio interface
demo.launch()
if __name__ == "__main__":
main()