id
stringlengths 6
6
| query_type
stringclasses 14
values | question
dict | paraphrased_question
sequence | query
dict | template_id
stringclasses 8
values | query_shape
stringclasses 7
values | query_class
stringclasses 5
values | auto_generated
bool 2
classes | number_of_patterns
int32 1
14
|
---|---|---|---|---|---|---|---|---|---|
AQ1279 | non-factoid | {
"string": "What is the top benchmark result (metric and value) over the dataset Atari 2600 River Raid?"
} | [] | {
"sparql": "SELECT DISTINCT ?metric ?metric_lbl (MAX(?value) AS ?score)\nWHERE {\n {\n SELECT ?metric ?metric_lbl ?value\n WHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"Atari 2600 River Raid\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?eval orkgp:HAS_VALUE ?value.\n OPTIONAL {?eval orkgp:HAS_METRIC ?metric.\n ?metric rdfs:label ?metric_lbl.}\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n OPTIONAL {?cont orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.}\n }\n ORDER BY DESC(?value)\n }\n}\nGROUP BY ?metric ?metric_lbl"
} | T04 | Tree | WHICH-WHAT | true | 13 |
AQ0099 | Factoid | {
"string": "What models are being evaluated on the HMDB51 (finetuned) dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?model ?model_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"HMDB51 (finetuned)\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?paper orkgp:HAS_BENCHMARK ?benchmark.\n OPTIONAL {?paper orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.}\n}"
} | T01 | Tree | WHICH-WHAT | true | 6 |
AQ1523 | Factoid | {
"string": "Which model has achieved the highest Accuracy score on the LAMBADA benchmark dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?model ?model_lbl\nWHERE {\n ?metric a orkgc:Metric;\n rdfs:label ?metric_lbl.\n FILTER (str(?metric_lbl) = \"Accuracy\")\n {\n SELECT ?model ?model_lbl\n WHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"LAMBADA\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?eval orkgp:HAS_VALUE ?value;\n orkgp:HAS_METRIC ?metric.\n ?cont orkgp:HAS_BENCHMARK ?benchmark;\n orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.\n }\n ORDER BY DESC(?value)\n LIMIT 1\n }\n}"
} | T05 | Tree | WHICH-WHAT | true | 12 |
AQ0986 | Factoid | {
"string": "What evaluation metrics are commonly used when benchmarking models on the Recipe dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?metric ?metric_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"Recipe\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n OPTIONAL {?eval orkgp:HAS_METRIC ?metric.\n ?metric rdfs:label ?metric_lbl.}\n}"
} | T03 | Tree | WHICH-WHAT | true | 6 |
AQ2369 | Factoid | {
"string": "Where can I find code references in papers that have used the LeViT-256 model for benchmarking purposes?"
} | [] | {
"sparql": "SELECT DISTINCT ?code\nWHERE {\n ?model a orkgc:Model;\n rdfs:label ?model_lbl.\n FILTER (str(?model_lbl) = \"LeViT-256\")\n ?benchmark orkgp:HAS_DATASET ?dataset.\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n ?cont orkgp:HAS_MODEL ?model;\n orkgp:HAS_SOURCE_CODE ?code.\n}"
} | T07 | Tree | WHICH-WHAT | true | 4 |
AQ0092 | Factoid | {
"string": "Could you provide a list of models that have been tested on the ImageNet 64x64 benchmark dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?model ?model_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"ImageNet 64x64\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?paper orkgp:HAS_BENCHMARK ?benchmark.\n OPTIONAL {?paper orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.}\n}"
} | T01 | Tree | WHICH-WHAT | true | 6 |
AQ1292 | non-factoid | {
"string": "What is the highest benchmark result achieved on the Atari 2600 Time Pilot dataset, including the metric and its value?"
} | [] | {
"sparql": "SELECT DISTINCT ?metric ?metric_lbl (MAX(?value) AS ?score)\nWHERE {\n {\n SELECT ?metric ?metric_lbl ?value\n WHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"Atari 2600 Time Pilot\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?eval orkgp:HAS_VALUE ?value.\n OPTIONAL {?eval orkgp:HAS_METRIC ?metric.\n ?metric rdfs:label ?metric_lbl.}\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n OPTIONAL {?cont orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.}\n }\n ORDER BY DESC(?value)\n }\n}\nGROUP BY ?metric ?metric_lbl"
} | T04 | Tree | WHICH-WHAT | true | 13 |
AQ1652 | Factoid | {
"string": "What is the name of the top performing model in terms of Accuracy (%) score when benchmarked on the FGVC Aircraft dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?model ?model_lbl\nWHERE {\n ?metric a orkgc:Metric;\n rdfs:label ?metric_lbl.\n FILTER (str(?metric_lbl) = \"Accuracy (%)\")\n {\n SELECT ?model ?model_lbl\n WHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"FGVC Aircraft\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?eval orkgp:HAS_VALUE ?value;\n orkgp:HAS_METRIC ?metric.\n ?cont orkgp:HAS_BENCHMARK ?benchmark;\n orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.\n }\n ORDER BY DESC(?value)\n LIMIT 1\n }\n}"
} | T05 | Tree | WHICH-WHAT | true | 12 |
AQ2239 | Factoid | {
"string": "Provide a list of papers that have utilized the BiLSTM-CRF+ELMo model and include the links to their code?"
} | [] | {
"sparql": "SELECT DISTINCT ?code\nWHERE {\n ?model a orkgc:Model;\n rdfs:label ?model_lbl.\n FILTER (str(?model_lbl) = \"BiLSTM-CRF+ELMo\")\n ?benchmark orkgp:HAS_DATASET ?dataset.\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n ?cont orkgp:HAS_MODEL ?model;\n orkgp:HAS_SOURCE_CODE ?code.\n}"
} | T07 | Tree | WHICH-WHAT | true | 4 |
AQ0502 | Factoid | {
"string": "What are the titles and IDs of research papers that include a benchmark for the MedSTS dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?paper ?paper_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"MedSTS\")\n ?benchmark orkgp:HAS_DATASET ?dataset.\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n ?paper orkgp:P31 ?cont;\n rdfs:label ?paper_lbl.\n}"
} | T02 | Tree | WHICH-WHAT | true | 5 |
AQ2373 | Factoid | {
"string": "Where can I find code references in papers that have used the CAIT-M36-448 model for benchmarking purposes?"
} | [] | {
"sparql": "SELECT DISTINCT ?code\nWHERE {\n ?model a orkgc:Model;\n rdfs:label ?model_lbl.\n FILTER (str(?model_lbl) = \"CAIT-M36-448\")\n ?benchmark orkgp:HAS_DATASET ?dataset.\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n ?cont orkgp:HAS_MODEL ?model;\n orkgp:HAS_SOURCE_CODE ?code.\n}"
} | T07 | Tree | WHICH-WHAT | true | 4 |
AQ0308 | Factoid | {
"string": "What models are being evaluated on the Paper Field dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?model ?model_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"Paper Field\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?paper orkgp:HAS_BENCHMARK ?benchmark.\n OPTIONAL {?paper orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.}\n}"
} | T01 | Tree | WHICH-WHAT | true | 6 |
AQ0194 | Factoid | {
"string": "Can you list the models that have been evaluated on the Reacher, easy (DMControl100k) dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?model ?model_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"Reacher, easy (DMControl100k)\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?paper orkgp:HAS_BENCHMARK ?benchmark.\n OPTIONAL {?paper orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.}\n}"
} | T01 | Tree | WHICH-WHAT | true | 6 |
AQ1205 | non-factoid | {
"string": "Can you provide the highest benchmark result, including the metric and score, for the UrbanSound8k dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?metric ?metric_lbl (MAX(?value) AS ?score)\nWHERE {\n {\n SELECT ?metric ?metric_lbl ?value\n WHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"UrbanSound8k\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?eval orkgp:HAS_VALUE ?value.\n OPTIONAL {?eval orkgp:HAS_METRIC ?metric.\n ?metric rdfs:label ?metric_lbl.}\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n OPTIONAL {?cont orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.}\n }\n ORDER BY DESC(?value)\n }\n}\nGROUP BY ?metric ?metric_lbl"
} | T04 | Tree | WHICH-WHAT | true | 13 |
AQ0723 | Factoid | {
"string": "What are the metrics of evaluation over the smallNLP-KG dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?metric ?metric_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"smallNLP-KG\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n OPTIONAL {?eval orkgp:HAS_METRIC ?metric.\n ?metric rdfs:label ?metric_lbl.}\n}"
} | T03 | Tree | WHICH-WHAT | true | 6 |
AQ1400 | Factoid | {
"string": "What is the best performing model benchmarking the ChemProt dataset in terms of F1 metric?"
} | [] | {
"sparql": "SELECT DISTINCT ?model ?model_lbl\nWHERE {\n ?metric a orkgc:Metric;\n rdfs:label ?metric_lbl.\n FILTER (str(?metric_lbl) = \"F1\")\n {\n SELECT ?model ?model_lbl\n WHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"ChemProt\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?eval orkgp:HAS_VALUE ?value;\n orkgp:HAS_METRIC ?metric.\n ?cont orkgp:HAS_BENCHMARK ?benchmark;\n orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.\n }\n ORDER BY DESC(?value)\n LIMIT 1\n }\n}"
} | T05 | Tree | WHICH-WHAT | true | 12 |
AQ0826 | Factoid | {
"string": "What are the metrics of evaluation over the Words in Context dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?metric ?metric_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"Words in Context\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n OPTIONAL {?eval orkgp:HAS_METRIC ?metric.\n ?metric rdfs:label ?metric_lbl.}\n}"
} | T03 | Tree | WHICH-WHAT | true | 6 |
AQ1750 | Factoid | {
"string": "Which model has achieved the highest Score score on the Atari 2600 Skiing benchmark dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?model ?model_lbl\nWHERE {\n ?metric a orkgc:Metric;\n rdfs:label ?metric_lbl.\n FILTER (str(?metric_lbl) = \"Score\")\n {\n SELECT ?model ?model_lbl\n WHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"Atari 2600 Skiing\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?eval orkgp:HAS_VALUE ?value;\n orkgp:HAS_METRIC ?metric.\n ?cont orkgp:HAS_BENCHMARK ?benchmark;\n orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.\n }\n ORDER BY DESC(?value)\n LIMIT 1\n }\n}"
} | T05 | Tree | WHICH-WHAT | true | 12 |
AQ1159 | non-factoid | {
"string": "What is the top benchmark score and its metric on the Rotowire (Content Selection) dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?metric ?metric_lbl (MAX(?value) AS ?score)\nWHERE {\n {\n SELECT ?metric ?metric_lbl ?value\n WHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"Rotowire (Content Selection)\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?eval orkgp:HAS_VALUE ?value.\n OPTIONAL {?eval orkgp:HAS_METRIC ?metric.\n ?metric rdfs:label ?metric_lbl.}\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n OPTIONAL {?cont orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.}\n }\n ORDER BY DESC(?value)\n }\n}\nGROUP BY ?metric ?metric_lbl"
} | T04 | Tree | WHICH-WHAT | true | 13 |
AQ1595 | Factoid | {
"string": "Which model has achieved the highest ROUGE-1 score on the Pubmed benchmark dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?model ?model_lbl\nWHERE {\n ?metric a orkgc:Metric;\n rdfs:label ?metric_lbl.\n FILTER (str(?metric_lbl) = \"ROUGE-1\")\n {\n SELECT ?model ?model_lbl\n WHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"Pubmed\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?eval orkgp:HAS_VALUE ?value;\n orkgp:HAS_METRIC ?metric.\n ?cont orkgp:HAS_BENCHMARK ?benchmark;\n orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.\n }\n ORDER BY DESC(?value)\n LIMIT 1\n }\n}"
} | T05 | Tree | WHICH-WHAT | true | 12 |
AQ0974 | Factoid | {
"string": "Can you list the metrics used to evaluate models on the Atari 2600 Star Gunner dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?metric ?metric_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"Atari 2600 Star Gunner\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n OPTIONAL {?eval orkgp:HAS_METRIC ?metric.\n ?metric rdfs:label ?metric_lbl.}\n}"
} | T03 | Tree | WHICH-WHAT | true | 6 |
AQ1253 | non-factoid | {
"string": "Can you provide the highest benchmark result, including the metric and score, for the MLDoc Zero-Shot English-to-Chinese dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?metric ?metric_lbl (MAX(?value) AS ?score)\nWHERE {\n {\n SELECT ?metric ?metric_lbl ?value\n WHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"MLDoc Zero-Shot English-to-Chinese\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?eval orkgp:HAS_VALUE ?value.\n OPTIONAL {?eval orkgp:HAS_METRIC ?metric.\n ?metric rdfs:label ?metric_lbl.}\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n OPTIONAL {?cont orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.}\n }\n ORDER BY DESC(?value)\n }\n}\nGROUP BY ?metric ?metric_lbl"
} | T04 | Tree | WHICH-WHAT | true | 13 |
AQ0018 | Factoid | {
"string": "Can you list the models that have been evaluated on the SciERC dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?model ?model_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"SciERC\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?paper orkgp:HAS_BENCHMARK ?benchmark.\n OPTIONAL {?paper orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.}\n}"
} | T01 | Tree | WHICH-WHAT | true | 6 |
AQ0035 | Factoid | {
"string": "What are the models that have been benchmarked on the DRI Corpus dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?model ?model_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"DRI Corpus\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?paper orkgp:HAS_BENCHMARK ?benchmark.\n OPTIONAL {?paper orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.}\n}"
} | T01 | Tree | WHICH-WHAT | true | 6 |
AQ2258 | Factoid | {
"string": "Can you provide links to code used in papers that benchmark the Biinclusion (Euro500kReuters) model?"
} | [] | {
"sparql": "SELECT DISTINCT ?code\nWHERE {\n ?model a orkgc:Model;\n rdfs:label ?model_lbl.\n FILTER (str(?model_lbl) = \"Biinclusion (Euro500kReuters)\")\n ?benchmark orkgp:HAS_DATASET ?dataset.\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n ?cont orkgp:HAS_MODEL ?model;\n orkgp:HAS_SOURCE_CODE ?code.\n}"
} | T07 | Tree | WHICH-WHAT | true | 4 |
AQ0985 | Factoid | {
"string": "What are the metrics of evaluation over the Classic dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?metric ?metric_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"Classic\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n OPTIONAL {?eval orkgp:HAS_METRIC ?metric.\n ?metric rdfs:label ?metric_lbl.}\n}"
} | T03 | Tree | WHICH-WHAT | true | 6 |
AQ0120 | Factoid | {
"string": "What are the models that have been benchmarked on the TriviaQA dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?model ?model_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"TriviaQA\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?paper orkgp:HAS_BENCHMARK ?benchmark.\n OPTIONAL {?paper orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.}\n}"
} | T01 | Tree | WHICH-WHAT | true | 6 |
AQ2135 | Factoid | {
"string": "List the code links in papers that use the Transformer-XL + RMS dynamic eval model in any benchmark?"
} | [] | {
"sparql": "SELECT DISTINCT ?code\nWHERE {\n ?model a orkgc:Model;\n rdfs:label ?model_lbl.\n FILTER (str(?model_lbl) = \"Transformer-XL + RMS dynamic eval\")\n ?benchmark orkgp:HAS_DATASET ?dataset.\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n ?cont orkgp:HAS_MODEL ?model;\n orkgp:HAS_SOURCE_CODE ?code.\n}"
} | T07 | Tree | WHICH-WHAT | true | 4 |
AQ1908 | Factoid | {
"string": "Provide a list of papers that have utilized the LNN-EL/ens model and include the links to their code?"
} | [] | {
"sparql": "SELECT DISTINCT ?code\nWHERE {\n ?model a orkgc:Model;\n rdfs:label ?model_lbl.\n FILTER (str(?model_lbl) = \"LNN-EL/ens\")\n ?benchmark orkgp:HAS_DATASET ?dataset.\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n ?cont orkgp:HAS_MODEL ?model;\n orkgp:HAS_SOURCE_CODE ?code.\n}"
} | T07 | Tree | WHICH-WHAT | true | 4 |
AQ1609 | Factoid | {
"string": "Indicate the model that performed best in terms of Accuracy (%) metric on the STL-10 benchmark dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?model ?model_lbl\nWHERE {\n ?metric a orkgc:Metric;\n rdfs:label ?metric_lbl.\n FILTER (str(?metric_lbl) = \"Accuracy (%)\")\n {\n SELECT ?model ?model_lbl\n WHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"STL-10\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?eval orkgp:HAS_VALUE ?value;\n orkgp:HAS_METRIC ?metric.\n ?cont orkgp:HAS_BENCHMARK ?benchmark;\n orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.\n }\n ORDER BY DESC(?value)\n LIMIT 1\n }\n}"
} | T05 | Tree | WHICH-WHAT | true | 12 |
AQ2113 | Factoid | {
"string": "Can you provide links to code used in papers that benchmark the SRU++ Large model?"
} | [] | {
"sparql": "SELECT DISTINCT ?code\nWHERE {\n ?model a orkgc:Model;\n rdfs:label ?model_lbl.\n FILTER (str(?model_lbl) = \"SRU++ Large\")\n ?benchmark orkgp:HAS_DATASET ?dataset.\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n ?cont orkgp:HAS_MODEL ?model;\n orkgp:HAS_SOURCE_CODE ?code.\n}"
} | T07 | Tree | WHICH-WHAT | true | 4 |
AQ0276 | Factoid | {
"string": "What models are being evaluated on the Atari 2600 Private Eye dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?model ?model_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"Atari 2600 Private Eye\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?paper orkgp:HAS_BENCHMARK ?benchmark.\n OPTIONAL {?paper orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.}\n}"
} | T01 | Tree | WHICH-WHAT | true | 6 |
AQ2306 | Factoid | {
"string": "Provide a list of papers that have utilized the DQN hs model and include the links to their code?"
} | [] | {
"sparql": "SELECT DISTINCT ?code\nWHERE {\n ?model a orkgc:Model;\n rdfs:label ?model_lbl.\n FILTER (str(?model_lbl) = \"DQN hs\")\n ?benchmark orkgp:HAS_DATASET ?dataset.\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n ?cont orkgp:HAS_MODEL ?model;\n orkgp:HAS_SOURCE_CODE ?code.\n}"
} | T07 | Tree | WHICH-WHAT | true | 4 |
AQ0776 | Factoid | {
"string": "What evaluation metrics are commonly used when benchmarking models on the Stanford Dogs dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?metric ?metric_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"Stanford Dogs\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n OPTIONAL {?eval orkgp:HAS_METRIC ?metric.\n ?metric rdfs:label ?metric_lbl.}\n}"
} | T03 | Tree | WHICH-WHAT | true | 6 |
AQ1048 | non-factoid | {
"string": "Can you provide the highest benchmark result, including the metric and score, for the SemEval-2021 Task 11 dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?metric ?metric_lbl (MAX(?value) AS ?score)\nWHERE {\n {\n SELECT ?metric ?metric_lbl ?value\n WHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"SemEval-2021 Task 11\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?eval orkgp:HAS_VALUE ?value.\n OPTIONAL {?eval orkgp:HAS_METRIC ?metric.\n ?metric rdfs:label ?metric_lbl.}\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n OPTIONAL {?cont orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.}\n }\n ORDER BY DESC(?value)\n }\n}\nGROUP BY ?metric ?metric_lbl"
} | T04 | Tree | WHICH-WHAT | true | 13 |
AQ0483 | Factoid | {
"string": "What are the titles and IDs of research papers that include a benchmark for the Words in Context dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?paper ?paper_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"Words in Context\")\n ?benchmark orkgp:HAS_DATASET ?dataset.\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n ?paper orkgp:P31 ?cont;\n rdfs:label ?paper_lbl.\n}"
} | T02 | Tree | WHICH-WHAT | true | 5 |
AQ0011 | Factoid | {
"string": "What models are being evaluated on the ORKG-TDM dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?model ?model_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"ORKG-TDM\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?paper orkgp:HAS_BENCHMARK ?benchmark.\n OPTIONAL {?paper orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.}\n}"
} | T01 | Tree | WHICH-WHAT | true | 6 |
AQ0075 | Factoid | {
"string": "What models are being evaluated on the WMT2014 English-French dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?model ?model_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"WMT2014 English-French\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?paper orkgp:HAS_BENCHMARK ?benchmark.\n OPTIONAL {?paper orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.}\n}"
} | T01 | Tree | WHICH-WHAT | true | 6 |
AQ2175 | Factoid | {
"string": "List the code links in papers that use the Transformer-XL Large model in any benchmark?"
} | [] | {
"sparql": "SELECT DISTINCT ?code\nWHERE {\n ?model a orkgc:Model;\n rdfs:label ?model_lbl.\n FILTER (str(?model_lbl) = \"Transformer-XL Large\")\n ?benchmark orkgp:HAS_DATASET ?dataset.\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n ?cont orkgp:HAS_MODEL ?model;\n orkgp:HAS_SOURCE_CODE ?code.\n}"
} | T07 | Tree | WHICH-WHAT | true | 4 |
AQ1933 | Factoid | {
"string": "Provide a list of papers that have utilized the Stable and Transferable Mixture-of-Experts model and include the links to their code?"
} | [] | {
"sparql": "SELECT DISTINCT ?code\nWHERE {\n ?model a orkgc:Model;\n rdfs:label ?model_lbl.\n FILTER (str(?model_lbl) = \"Stable and Transferable Mixture-of-Experts\")\n ?benchmark orkgp:HAS_DATASET ?dataset.\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n ?cont orkgp:HAS_MODEL ?model;\n orkgp:HAS_SOURCE_CODE ?code.\n}"
} | T07 | Tree | WHICH-WHAT | true | 4 |
AQ0519 | Factoid | {
"string": "List the title and ID of research papers that contain a benchmark over the UrbanSound8k dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?paper ?paper_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"UrbanSound8k\")\n ?benchmark orkgp:HAS_DATASET ?dataset.\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n ?paper orkgp:P31 ?cont;\n rdfs:label ?paper_lbl.\n}"
} | T02 | Tree | WHICH-WHAT | true | 5 |
AQ2420 | Factoid | {
"string": "Provide a list of papers that have utilized the CAIT-XXS-24 model and include the links to their code?"
} | [] | {
"sparql": "SELECT DISTINCT ?code\nWHERE {\n ?model a orkgc:Model;\n rdfs:label ?model_lbl.\n FILTER (str(?model_lbl) = \"CAIT-XXS-24\")\n ?benchmark orkgp:HAS_DATASET ?dataset.\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n ?cont orkgp:HAS_MODEL ?model;\n orkgp:HAS_SOURCE_CODE ?code.\n}"
} | T07 | Tree | WHICH-WHAT | true | 4 |
AQ1324 | non-factoid | {
"string": "What is the top benchmark result (metric and value) over the dataset TREC-6?"
} | [] | {
"sparql": "SELECT DISTINCT ?metric ?metric_lbl (MAX(?value) AS ?score)\nWHERE {\n {\n SELECT ?metric ?metric_lbl ?value\n WHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"TREC-6\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?eval orkgp:HAS_VALUE ?value.\n OPTIONAL {?eval orkgp:HAS_METRIC ?metric.\n ?metric rdfs:label ?metric_lbl.}\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n OPTIONAL {?cont orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.}\n }\n ORDER BY DESC(?value)\n }\n}\nGROUP BY ?metric ?metric_lbl"
} | T04 | Tree | WHICH-WHAT | true | 13 |
AQ1091 | non-factoid | {
"string": "What is the highest benchmark result achieved on the SemEval-2010 Task 8 dataset, including the metric and its value?"
} | [] | {
"sparql": "SELECT DISTINCT ?metric ?metric_lbl (MAX(?value) AS ?score)\nWHERE {\n {\n SELECT ?metric ?metric_lbl ?value\n WHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"SemEval-2010 Task 8\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?eval orkgp:HAS_VALUE ?value.\n OPTIONAL {?eval orkgp:HAS_METRIC ?metric.\n ?metric rdfs:label ?metric_lbl.}\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n OPTIONAL {?cont orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.}\n }\n ORDER BY DESC(?value)\n }\n}\nGROUP BY ?metric ?metric_lbl"
} | T04 | Tree | WHICH-WHAT | true | 13 |
AQ2399 | Factoid | {
"string": "Provide a list of papers that have utilized the EfficientNet-L2-475 (SAM) model and include the links to their code?"
} | [] | {
"sparql": "SELECT DISTINCT ?code\nWHERE {\n ?model a orkgc:Model;\n rdfs:label ?model_lbl.\n FILTER (str(?model_lbl) = \"EfficientNet-L2-475 (SAM)\")\n ?benchmark orkgp:HAS_DATASET ?dataset.\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n ?cont orkgp:HAS_MODEL ?model;\n orkgp:HAS_SOURCE_CODE ?code.\n}"
} | T07 | Tree | WHICH-WHAT | true | 4 |
AQ1983 | Factoid | {
"string": "Can you provide links to code used in papers that benchmark the Transformer model?"
} | [] | {
"sparql": "SELECT DISTINCT ?code\nWHERE {\n ?model a orkgc:Model;\n rdfs:label ?model_lbl.\n FILTER (str(?model_lbl) = \"Transformer\")\n ?benchmark orkgp:HAS_DATASET ?dataset.\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n ?cont orkgp:HAS_MODEL ?model;\n orkgp:HAS_SOURCE_CODE ?code.\n}"
} | T07 | Tree | WHICH-WHAT | true | 4 |
AQ0998 | Factoid | {
"string": "Can you list the metrics used to evaluate models on the ScienceCite dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?metric ?metric_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"ScienceCite\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n OPTIONAL {?eval orkgp:HAS_METRIC ?metric.\n ?metric rdfs:label ?metric_lbl.}\n}"
} | T03 | Tree | WHICH-WHAT | true | 6 |
AQ0690 | Factoid | {
"string": "What are the metrics of evaluation over the AI-KG dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?metric ?metric_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"AI-KG\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n OPTIONAL {?eval orkgp:HAS_METRIC ?metric.\n ?metric rdfs:label ?metric_lbl.}\n}"
} | T03 | Tree | WHICH-WHAT | true | 6 |
AQ2283 | Factoid | {
"string": "List the code links in papers that use the Baseline : Random model in any benchmark?"
} | [] | {
"sparql": "SELECT DISTINCT ?code\nWHERE {\n ?model a orkgc:Model;\n rdfs:label ?model_lbl.\n FILTER (str(?model_lbl) = \"Baseline : Random\")\n ?benchmark orkgp:HAS_DATASET ?dataset.\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n ?cont orkgp:HAS_MODEL ?model;\n orkgp:HAS_SOURCE_CODE ?code.\n}"
} | T07 | Tree | WHICH-WHAT | true | 4 |
AQ2438 | Factoid | {
"string": "Provide a list of papers that have utilized the VGG11B(3x) + LocalLearning model and include the links to their code?"
} | [] | {
"sparql": "SELECT DISTINCT ?code\nWHERE {\n ?model a orkgc:Model;\n rdfs:label ?model_lbl.\n FILTER (str(?model_lbl) = \"VGG11B(3x) + LocalLearning\")\n ?benchmark orkgp:HAS_DATASET ?dataset.\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n ?cont orkgp:HAS_MODEL ?model;\n orkgp:HAS_SOURCE_CODE ?code.\n}"
} | T07 | Tree | WHICH-WHAT | true | 4 |
AQ2302 | Factoid | {
"string": "Can you provide links to code used in papers that benchmark the Recurrent Rational DQN Average model?"
} | [] | {
"sparql": "SELECT DISTINCT ?code\nWHERE {\n ?model a orkgc:Model;\n rdfs:label ?model_lbl.\n FILTER (str(?model_lbl) = \"Recurrent Rational DQN Average\")\n ?benchmark orkgp:HAS_DATASET ?dataset.\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n ?cont orkgp:HAS_MODEL ?model;\n orkgp:HAS_SOURCE_CODE ?code.\n}"
} | T07 | Tree | WHICH-WHAT | true | 4 |
AQ0225 | Factoid | {
"string": "What models are being evaluated on the MLDoc Zero-Shot English-to-French dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?model ?model_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"MLDoc Zero-Shot English-to-French\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?paper orkgp:HAS_BENCHMARK ?benchmark.\n OPTIONAL {?paper orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.}\n}"
} | T01 | Tree | WHICH-WHAT | true | 6 |
AQ1818 | Factoid | {
"string": "What is the name of the top performing model in terms of Macro F1 score when benchmarked on the PWC Leaderboards (restricted) dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?model ?model_lbl\nWHERE {\n ?metric a orkgc:Metric;\n rdfs:label ?metric_lbl.\n FILTER (str(?metric_lbl) = \"Macro F1\")\n {\n SELECT ?model ?model_lbl\n WHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"PWC Leaderboards (restricted)\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?eval orkgp:HAS_VALUE ?value;\n orkgp:HAS_METRIC ?metric.\n ?cont orkgp:HAS_BENCHMARK ?benchmark;\n orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.\n }\n ORDER BY DESC(?value)\n LIMIT 1\n }\n}"
} | T05 | Tree | WHICH-WHAT | true | 12 |
AQ0403 | Factoid | {
"string": "What are the titles and IDs of research papers that include a benchmark for the GAD dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?paper ?paper_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"GAD\")\n ?benchmark orkgp:HAS_DATASET ?dataset.\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n ?paper orkgp:P31 ?cont;\n rdfs:label ?paper_lbl.\n}"
} | T02 | Tree | WHICH-WHAT | true | 5 |
AQ2179 | Factoid | {
"string": "Can you provide links to code used in papers that benchmark the PAR Transformer Base model?"
} | [] | {
"sparql": "SELECT DISTINCT ?code\nWHERE {\n ?model a orkgc:Model;\n rdfs:label ?model_lbl.\n FILTER (str(?model_lbl) = \"PAR Transformer Base\")\n ?benchmark orkgp:HAS_DATASET ?dataset.\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n ?cont orkgp:HAS_MODEL ?model;\n orkgp:HAS_SOURCE_CODE ?code.\n}"
} | T07 | Tree | WHICH-WHAT | true | 4 |
AQ0237 | Factoid | {
"string": "Can you list the models that have been evaluated on the Atari 2600 Crazy Climber dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?model ?model_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"Atari 2600 Crazy Climber\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?paper orkgp:HAS_BENCHMARK ?benchmark.\n OPTIONAL {?paper orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.}\n}"
} | T01 | Tree | WHICH-WHAT | true | 6 |
AQ0691 | Factoid | {
"string": "Can you list the metrics used to evaluate models on the SciREX dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?metric ?metric_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"SciREX\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n OPTIONAL {?eval orkgp:HAS_METRIC ?metric.\n ?metric rdfs:label ?metric_lbl.}\n}"
} | T03 | Tree | WHICH-WHAT | true | 6 |
AQ0413 | Factoid | {
"string": "Give me a list of research papers along with their titles and IDs, that have performed benchmarks on the AG News dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?paper ?paper_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"AG News\")\n ?benchmark orkgp:HAS_DATASET ?dataset.\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n ?paper orkgp:P31 ?cont;\n rdfs:label ?paper_lbl.\n}"
} | T02 | Tree | WHICH-WHAT | true | 5 |
AQ0423 | Factoid | {
"string": "List the title and ID of research papers that contain a benchmark over the 20NEWS dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?paper ?paper_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"20NEWS\")\n ?benchmark orkgp:HAS_DATASET ?dataset.\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n ?paper orkgp:P31 ?cont;\n rdfs:label ?paper_lbl.\n}"
} | T02 | Tree | WHICH-WHAT | true | 5 |
AQ1862 | Factoid | {
"string": "What are the most commonly used benchmark datasets for the Text Generation research field?"
} | [] | {
"sparql": "SELECT DISTINCT ?dataset ?dataset_lbl\nWHERE {\n ?problem a orkgc:Problem;\n rdfs:label ?problem_lbl. \n FILTER (str(?problem_lbl) = \"Text Generation\")\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n ?benchmark orkgp:HAS_DATASET ?dataset.\n ?cont orkgp:HAS_BENCHMARK ?benchmark;\n orkgp:P32 ?problem.\n}"
} | T06 | Tree | WHICH-WHAT | true | 5 |
AQ0307 | Factoid | {
"string": "Can you list the models that have been evaluated on the Reuters En-De dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?model ?model_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"Reuters En-De\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?paper orkgp:HAS_BENCHMARK ?benchmark.\n OPTIONAL {?paper orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.}\n}"
} | T01 | Tree | WHICH-WHAT | true | 6 |
AQ1521 | Factoid | {
"string": "What is the best performing model benchmarking the enwik8 dataset in terms of Number of params metric?"
} | [] | {
"sparql": "SELECT DISTINCT ?model ?model_lbl\nWHERE {\n ?metric a orkgc:Metric;\n rdfs:label ?metric_lbl.\n FILTER (str(?metric_lbl) = \"Number of params\")\n {\n SELECT ?model ?model_lbl\n WHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"enwik8\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?eval orkgp:HAS_VALUE ?value;\n orkgp:HAS_METRIC ?metric.\n ?cont orkgp:HAS_BENCHMARK ?benchmark;\n orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.\n }\n ORDER BY DESC(?value)\n LIMIT 1\n }\n}"
} | T05 | Tree | WHICH-WHAT | true | 12 |
AQ0218 | Factoid | {
"string": "What are the models that have been benchmarked on the Reuters RCV1/RCV2 German-to-English dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?model ?model_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"Reuters RCV1/RCV2 German-to-English\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?paper orkgp:HAS_BENCHMARK ?benchmark.\n OPTIONAL {?paper orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.}\n}"
} | T01 | Tree | WHICH-WHAT | true | 6 |
AQ1489 | Factoid | {
"string": "What is the best performing model benchmarking the One Billion Word dataset in terms of PPL metric?"
} | [] | {
"sparql": "SELECT DISTINCT ?model ?model_lbl\nWHERE {\n ?metric a orkgc:Metric;\n rdfs:label ?metric_lbl.\n FILTER (str(?metric_lbl) = \"PPL\")\n {\n SELECT ?model ?model_lbl\n WHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"One Billion Word\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?eval orkgp:HAS_VALUE ?value;\n orkgp:HAS_METRIC ?metric.\n ?cont orkgp:HAS_BENCHMARK ?benchmark;\n orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.\n }\n ORDER BY DESC(?value)\n LIMIT 1\n }\n}"
} | T05 | Tree | WHICH-WHAT | true | 12 |
AQ0140 | Factoid | {
"string": "What are the models that have been benchmarked on the Words in Context dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?model ?model_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"Words in Context\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?paper orkgp:HAS_BENCHMARK ?benchmark.\n OPTIONAL {?paper orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.}\n}"
} | T01 | Tree | WHICH-WHAT | true | 6 |
AQ0642 | Factoid | {
"string": "Give me a list of research papers along with their titles and IDs, that have performed benchmarks on the Classic dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?paper ?paper_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"Classic\")\n ?benchmark orkgp:HAS_DATASET ?dataset.\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n ?paper orkgp:P31 ?cont;\n rdfs:label ?paper_lbl.\n}"
} | T02 | Tree | WHICH-WHAT | true | 5 |
AQ1896 | Factoid | {
"string": "Name the datasets that have been used for benchmarking in the Graph Embedding research problem?"
} | [] | {
"sparql": "SELECT DISTINCT ?dataset ?dataset_lbl\nWHERE {\n ?problem a orkgc:Problem;\n rdfs:label ?problem_lbl. \n FILTER (str(?problem_lbl) = \"Graph Embedding\")\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n ?benchmark orkgp:HAS_DATASET ?dataset.\n ?cont orkgp:HAS_BENCHMARK ?benchmark;\n orkgp:P32 ?problem.\n}"
} | T06 | Tree | WHICH-WHAT | true | 5 |
AQ1238 | non-factoid | {
"string": "What is the top benchmark score and its metric on the ClueWeb09-B dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?metric ?metric_lbl (MAX(?value) AS ?score)\nWHERE {\n {\n SELECT ?metric ?metric_lbl ?value\n WHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"ClueWeb09-B\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?eval orkgp:HAS_VALUE ?value.\n OPTIONAL {?eval orkgp:HAS_METRIC ?metric.\n ?metric rdfs:label ?metric_lbl.}\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n OPTIONAL {?cont orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.}\n }\n ORDER BY DESC(?value)\n }\n}\nGROUP BY ?metric ?metric_lbl"
} | T04 | Tree | WHICH-WHAT | true | 13 |
AQ0838 | Factoid | {
"string": "List the metrics that are used to evaluate models on the Penn Treebank (Word Level) benchmark dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?metric ?metric_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"Penn Treebank (Word Level)\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n OPTIONAL {?eval orkgp:HAS_METRIC ?metric.\n ?metric rdfs:label ?metric_lbl.}\n}"
} | T03 | Tree | WHICH-WHAT | true | 6 |
AQ1515 | Factoid | {
"string": "Which model has achieved the highest F1 score on the MedNLI benchmark dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?model ?model_lbl\nWHERE {\n ?metric a orkgc:Metric;\n rdfs:label ?metric_lbl.\n FILTER (str(?metric_lbl) = \"F1\")\n {\n SELECT ?model ?model_lbl\n WHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"MedNLI\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?eval orkgp:HAS_VALUE ?value;\n orkgp:HAS_METRIC ?metric.\n ?cont orkgp:HAS_BENCHMARK ?benchmark;\n orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.\n }\n ORDER BY DESC(?value)\n LIMIT 1\n }\n}"
} | T05 | Tree | WHICH-WHAT | true | 12 |
AQ2062 | Factoid | {
"string": "List the code links in papers that use the BERT-joint model in any benchmark?"
} | [] | {
"sparql": "SELECT DISTINCT ?code\nWHERE {\n ?model a orkgc:Model;\n rdfs:label ?model_lbl.\n FILTER (str(?model_lbl) = \"BERT-joint\")\n ?benchmark orkgp:HAS_DATASET ?dataset.\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n ?cont orkgp:HAS_MODEL ?model;\n orkgp:HAS_SOURCE_CODE ?code.\n}"
} | T07 | Tree | WHICH-WHAT | true | 4 |
AQ1952 | Factoid | {
"string": "Can you provide links to code used in papers that benchmark the BERT classifier model?"
} | [] | {
"sparql": "SELECT DISTINCT ?code\nWHERE {\n ?model a orkgc:Model;\n rdfs:label ?model_lbl.\n FILTER (str(?model_lbl) = \"BERT classifier\")\n ?benchmark orkgp:HAS_DATASET ?dataset.\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n ?cont orkgp:HAS_MODEL ?model;\n orkgp:HAS_SOURCE_CODE ?code.\n}"
} | T07 | Tree | WHICH-WHAT | true | 4 |
AQ0363 | Factoid | {
"string": "Give me a list of research papers along with their titles and IDs, that have performed benchmarks on the ner_dataset_recognition dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?paper ?paper_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"ner_dataset_recognition\")\n ?benchmark orkgp:HAS_DATASET ?dataset.\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n ?paper orkgp:P31 ?cont;\n rdfs:label ?paper_lbl.\n}"
} | T02 | Tree | WHICH-WHAT | true | 5 |
AQ0766 | Factoid | {
"string": "Can you list the metrics used to evaluate models on the 20NEWS dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?metric ?metric_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"20NEWS\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n OPTIONAL {?eval orkgp:HAS_METRIC ?metric.\n ?metric rdfs:label ?metric_lbl.}\n}"
} | T03 | Tree | WHICH-WHAT | true | 6 |
AQ0704 | Factoid | {
"string": "Can you list the metrics used to evaluate models on the SciERC dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?metric ?metric_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"SciERC\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n OPTIONAL {?eval orkgp:HAS_METRIC ?metric.\n ?metric rdfs:label ?metric_lbl.}\n}"
} | T03 | Tree | WHICH-WHAT | true | 6 |
AQ1653 | Factoid | {
"string": "What is the name of the top performing model in terms of FLOPS score when benchmarked on the FGVC Aircraft dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?model ?model_lbl\nWHERE {\n ?metric a orkgc:Metric;\n rdfs:label ?metric_lbl.\n FILTER (str(?metric_lbl) = \"FLOPS\")\n {\n SELECT ?model ?model_lbl\n WHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"FGVC Aircraft\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?eval orkgp:HAS_VALUE ?value;\n orkgp:HAS_METRIC ?metric.\n ?cont orkgp:HAS_BENCHMARK ?benchmark;\n orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.\n }\n ORDER BY DESC(?value)\n LIMIT 1\n }\n}"
} | T05 | Tree | WHICH-WHAT | true | 12 |
AQ1800 | Factoid | {
"string": "Which model has achieved the highest Accuracy score on the Reuters De-En benchmark dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?model ?model_lbl\nWHERE {\n ?metric a orkgc:Metric;\n rdfs:label ?metric_lbl.\n FILTER (str(?metric_lbl) = \"Accuracy\")\n {\n SELECT ?model ?model_lbl\n WHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"Reuters De-En\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?eval orkgp:HAS_VALUE ?value;\n orkgp:HAS_METRIC ?metric.\n ?cont orkgp:HAS_BENCHMARK ?benchmark;\n orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.\n }\n ORDER BY DESC(?value)\n LIMIT 1\n }\n}"
} | T05 | Tree | WHICH-WHAT | true | 12 |
AQ2178 | Factoid | {
"string": "Provide a list of papers that have utilized the Feedback Transformer (4 layers) model and include the links to their code?"
} | [] | {
"sparql": "SELECT DISTINCT ?code\nWHERE {\n ?model a orkgc:Model;\n rdfs:label ?model_lbl.\n FILTER (str(?model_lbl) = \"Feedback Transformer (4 layers)\")\n ?benchmark orkgp:HAS_DATASET ?dataset.\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n ?cont orkgp:HAS_MODEL ?model;\n orkgp:HAS_SOURCE_CODE ?code.\n}"
} | T07 | Tree | WHICH-WHAT | true | 4 |
AQ1618 | Factoid | {
"string": "What is the name of the top performing model in terms of Accuracy score when benchmarked on the ImageNet dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?model ?model_lbl\nWHERE {\n ?metric a orkgc:Metric;\n rdfs:label ?metric_lbl.\n FILTER (str(?metric_lbl) = \"Accuracy\")\n {\n SELECT ?model ?model_lbl\n WHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"ImageNet\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?eval orkgp:HAS_VALUE ?value;\n orkgp:HAS_METRIC ?metric.\n ?cont orkgp:HAS_BENCHMARK ?benchmark;\n orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.\n }\n ORDER BY DESC(?value)\n LIMIT 1\n }\n}"
} | T05 | Tree | WHICH-WHAT | true | 12 |
AQ2240 | Factoid | {
"string": "List the code links in papers that use the NCBI_BERT(base) (P) model in any benchmark?"
} | [] | {
"sparql": "SELECT DISTINCT ?code\nWHERE {\n ?model a orkgc:Model;\n rdfs:label ?model_lbl.\n FILTER (str(?model_lbl) = \"NCBI_BERT(base) (P)\")\n ?benchmark orkgp:HAS_DATASET ?dataset.\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n ?cont orkgp:HAS_MODEL ?model;\n orkgp:HAS_SOURCE_CODE ?code.\n}"
} | T07 | Tree | WHICH-WHAT | true | 4 |
AQ1849 | Factoid | {
"string": "Provide a list of benchmarked datasets related to the stochastic classification research area?"
} | [] | {
"sparql": "SELECT DISTINCT ?dataset ?dataset_lbl\nWHERE {\n ?problem a orkgc:Problem;\n rdfs:label ?problem_lbl. \n FILTER (str(?problem_lbl) = \"stochastic classification\")\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n ?benchmark orkgp:HAS_DATASET ?dataset.\n ?cont orkgp:HAS_BENCHMARK ?benchmark;\n orkgp:P32 ?problem.\n}"
} | T06 | Tree | WHICH-WHAT | true | 5 |
AQ0300 | Factoid | {
"string": "What models are being evaluated on the Recipe dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?model ?model_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"Recipe\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?paper orkgp:HAS_BENCHMARK ?benchmark.\n OPTIONAL {?paper orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.}\n}"
} | T01 | Tree | WHICH-WHAT | true | 6 |
AQ2335 | Factoid | {
"string": "Where can I find code references in papers that have used the Intrinsic Reward Agent model for benchmarking purposes?"
} | [] | {
"sparql": "SELECT DISTINCT ?code\nWHERE {\n ?model a orkgc:Model;\n rdfs:label ?model_lbl.\n FILTER (str(?model_lbl) = \"Intrinsic Reward Agent\")\n ?benchmark orkgp:HAS_DATASET ?dataset.\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n ?cont orkgp:HAS_MODEL ?model;\n orkgp:HAS_SOURCE_CODE ?code.\n}"
} | T07 | Tree | WHICH-WHAT | true | 4 |
AQ1392 | Factoid | {
"string": "Indicate the model that performed best in terms of RE Micro F1 metric on the ACE 2004 benchmark dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?model ?model_lbl\nWHERE {\n ?metric a orkgc:Metric;\n rdfs:label ?metric_lbl.\n FILTER (str(?metric_lbl) = \"RE Micro F1\")\n {\n SELECT ?model ?model_lbl\n WHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"ACE 2004\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?eval orkgp:HAS_VALUE ?value;\n orkgp:HAS_METRIC ?metric.\n ?cont orkgp:HAS_BENCHMARK ?benchmark;\n orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.\n }\n ORDER BY DESC(?value)\n LIMIT 1\n }\n}"
} | T05 | Tree | WHICH-WHAT | true | 12 |
AQ1193 | non-factoid | {
"string": "What is the top benchmark result (metric and value) over the dataset TempEval-3?"
} | [] | {
"sparql": "SELECT DISTINCT ?metric ?metric_lbl (MAX(?value) AS ?score)\nWHERE {\n {\n SELECT ?metric ?metric_lbl ?value\n WHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"TempEval-3\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?eval orkgp:HAS_VALUE ?value.\n OPTIONAL {?eval orkgp:HAS_METRIC ?metric.\n ?metric rdfs:label ?metric_lbl.}\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n OPTIONAL {?cont orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.}\n }\n ORDER BY DESC(?value)\n }\n}\nGROUP BY ?metric ?metric_lbl"
} | T04 | Tree | WHICH-WHAT | true | 13 |
AQ0223 | Factoid | {
"string": "What are the models that have been benchmarked on the MLDoc Zero-Shot English-to-Russian dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?model ?model_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"MLDoc Zero-Shot English-to-Russian\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?paper orkgp:HAS_BENCHMARK ?benchmark.\n OPTIONAL {?paper orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.}\n}"
} | T01 | Tree | WHICH-WHAT | true | 6 |
AQ0569 | Factoid | {
"string": "What are the titles and IDs of research papers that include a benchmark for the MLDoc Zero-Shot English-to-Italian dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?paper ?paper_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"MLDoc Zero-Shot English-to-Italian\")\n ?benchmark orkgp:HAS_DATASET ?dataset.\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n ?paper orkgp:P31 ?cont;\n rdfs:label ?paper_lbl.\n}"
} | T02 | Tree | WHICH-WHAT | true | 5 |
AQ1366 | non-factoid | {
"string": "What is the top benchmark score and its metric on the MUTAG dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?metric ?metric_lbl (MAX(?value) AS ?score)\nWHERE {\n {\n SELECT ?metric ?metric_lbl ?value\n WHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"MUTAG\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?eval orkgp:HAS_VALUE ?value.\n OPTIONAL {?eval orkgp:HAS_METRIC ?metric.\n ?metric rdfs:label ?metric_lbl.}\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n OPTIONAL {?cont orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.}\n }\n ORDER BY DESC(?value)\n }\n}\nGROUP BY ?metric ?metric_lbl"
} | T04 | Tree | WHICH-WHAT | true | 13 |
AQ1312 | non-factoid | {
"string": "What is the top benchmark result (metric and value) over the dataset Atari 2600 James Bond?"
} | [] | {
"sparql": "SELECT DISTINCT ?metric ?metric_lbl (MAX(?value) AS ?score)\nWHERE {\n {\n SELECT ?metric ?metric_lbl ?value\n WHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"Atari 2600 James Bond\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?eval orkgp:HAS_VALUE ?value.\n OPTIONAL {?eval orkgp:HAS_METRIC ?metric.\n ?metric rdfs:label ?metric_lbl.}\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n OPTIONAL {?cont orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.}\n }\n ORDER BY DESC(?value)\n }\n}\nGROUP BY ?metric ?metric_lbl"
} | T04 | Tree | WHICH-WHAT | true | 13 |
AQ2074 | Factoid | {
"string": "Where can I find code references in papers that have used the BigBird model for benchmarking purposes?"
} | [] | {
"sparql": "SELECT DISTINCT ?code\nWHERE {\n ?model a orkgc:Model;\n rdfs:label ?model_lbl.\n FILTER (str(?model_lbl) = \"BigBird\")\n ?benchmark orkgp:HAS_DATASET ?dataset.\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n ?cont orkgp:HAS_MODEL ?model;\n orkgp:HAS_SOURCE_CODE ?code.\n}"
} | T07 | Tree | WHICH-WHAT | true | 4 |
AQ1042 | non-factoid | {
"string": "What is the top benchmark result (metric and value) over the dataset CommonsenseQA?"
} | [] | {
"sparql": "SELECT DISTINCT ?metric ?metric_lbl (MAX(?value) AS ?score)\nWHERE {\n {\n SELECT ?metric ?metric_lbl ?value\n WHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"CommonsenseQA\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?eval orkgp:HAS_VALUE ?value.\n OPTIONAL {?eval orkgp:HAS_METRIC ?metric.\n ?metric rdfs:label ?metric_lbl.}\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n OPTIONAL {?cont orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.}\n }\n ORDER BY DESC(?value)\n }\n}\nGROUP BY ?metric ?metric_lbl"
} | T04 | Tree | WHICH-WHAT | true | 13 |
AQ0940 | Factoid | {
"string": "What are the metrics of evaluation over the Atari 2600 Krull dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?metric ?metric_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"Atari 2600 Krull\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n OPTIONAL {?eval orkgp:HAS_METRIC ?metric.\n ?metric rdfs:label ?metric_lbl.}\n}"
} | T03 | Tree | WHICH-WHAT | true | 6 |
AQ1035 | non-factoid | {
"string": "What is the top benchmark result (metric and value) over the dataset ARC-PDN?"
} | [] | {
"sparql": "SELECT DISTINCT ?metric ?metric_lbl (MAX(?value) AS ?score)\nWHERE {\n {\n SELECT ?metric ?metric_lbl ?value\n WHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"ARC-PDN\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?eval orkgp:HAS_VALUE ?value.\n OPTIONAL {?eval orkgp:HAS_METRIC ?metric.\n ?metric rdfs:label ?metric_lbl.}\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n OPTIONAL {?cont orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.}\n }\n ORDER BY DESC(?value)\n }\n}\nGROUP BY ?metric ?metric_lbl"
} | T04 | Tree | WHICH-WHAT | true | 13 |
AQ0805 | Factoid | {
"string": "List the metrics that are used to evaluate models on the COPA benchmark dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?metric ?metric_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"COPA\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n OPTIONAL {?eval orkgp:HAS_METRIC ?metric.\n ?metric rdfs:label ?metric_lbl.}\n}"
} | T03 | Tree | WHICH-WHAT | true | 6 |
AQ2049 | Factoid | {
"string": "Where can I find code references in papers that have used the AVTS model for benchmarking purposes?"
} | [] | {
"sparql": "SELECT DISTINCT ?code\nWHERE {\n ?model a orkgc:Model;\n rdfs:label ?model_lbl.\n FILTER (str(?model_lbl) = \"AVTS\")\n ?benchmark orkgp:HAS_DATASET ?dataset.\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n ?cont orkgp:HAS_MODEL ?model;\n orkgp:HAS_SOURCE_CODE ?code.\n}"
} | T07 | Tree | WHICH-WHAT | true | 4 |
AQ2089 | Factoid | {
"string": "Provide a list of papers that have utilized the Dynamic Coattention Networks (ensemble) model and include the links to their code?"
} | [] | {
"sparql": "SELECT DISTINCT ?code\nWHERE {\n ?model a orkgc:Model;\n rdfs:label ?model_lbl.\n FILTER (str(?model_lbl) = \"Dynamic Coattention Networks (ensemble)\")\n ?benchmark orkgp:HAS_DATASET ?dataset.\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n ?cont orkgp:HAS_MODEL ?model;\n orkgp:HAS_SOURCE_CODE ?code.\n}"
} | T07 | Tree | WHICH-WHAT | true | 4 |
AQ2310 | Factoid | {
"string": "List the code links in papers that use the DDRL A3C model in any benchmark?"
} | [] | {
"sparql": "SELECT DISTINCT ?code\nWHERE {\n ?model a orkgc:Model;\n rdfs:label ?model_lbl.\n FILTER (str(?model_lbl) = \"DDRL A3C\")\n ?benchmark orkgp:HAS_DATASET ?dataset.\n ?cont orkgp:HAS_BENCHMARK ?benchmark.\n ?cont orkgp:HAS_MODEL ?model;\n orkgp:HAS_SOURCE_CODE ?code.\n}"
} | T07 | Tree | WHICH-WHAT | true | 4 |
AQ1817 | Factoid | {
"string": "What is the best performing model benchmarking the STL-10, 1000 Labels dataset in terms of Accuracy metric?"
} | [] | {
"sparql": "SELECT DISTINCT ?model ?model_lbl\nWHERE {\n ?metric a orkgc:Metric;\n rdfs:label ?metric_lbl.\n FILTER (str(?metric_lbl) = \"Accuracy\")\n {\n SELECT ?model ?model_lbl\n WHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"STL-10, 1000 Labels\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?eval orkgp:HAS_VALUE ?value;\n orkgp:HAS_METRIC ?metric.\n ?cont orkgp:HAS_BENCHMARK ?benchmark;\n orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.\n }\n ORDER BY DESC(?value)\n LIMIT 1\n }\n}"
} | T05 | Tree | WHICH-WHAT | true | 12 |
AQ1579 | Factoid | {
"string": "What is the best performing model benchmarking the X-Sum dataset in terms of ROUGE-3 metric?"
} | [] | {
"sparql": "SELECT DISTINCT ?model ?model_lbl\nWHERE {\n ?metric a orkgc:Metric;\n rdfs:label ?metric_lbl.\n FILTER (str(?metric_lbl) = \"ROUGE-3\")\n {\n SELECT ?model ?model_lbl\n WHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"X-Sum\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n ?eval orkgp:HAS_VALUE ?value;\n orkgp:HAS_METRIC ?metric.\n ?cont orkgp:HAS_BENCHMARK ?benchmark;\n orkgp:HAS_MODEL ?model.\n ?model rdfs:label ?model_lbl.\n }\n ORDER BY DESC(?value)\n LIMIT 1\n }\n}"
} | T05 | Tree | WHICH-WHAT | true | 12 |
AQ0730 | Factoid | {
"string": "List the metrics that are used to evaluate models on the Pubmed benchmark dataset?"
} | [] | {
"sparql": "SELECT DISTINCT ?metric ?metric_lbl\nWHERE {\n ?dataset a orkgc:Dataset;\n rdfs:label ?dataset_lbl.\n FILTER (str(?dataset_lbl) = \"Pubmed\")\n ?benchmark orkgp:HAS_DATASET ?dataset;\n orkgp:HAS_EVALUATION ?eval.\n OPTIONAL {?eval orkgp:HAS_METRIC ?metric.\n ?metric rdfs:label ?metric_lbl.}\n}"
} | T03 | Tree | WHICH-WHAT | true | 6 |