albertvillanova HF staff commited on
Commit
15f7854
1 Parent(s): 09acc4b

Delete legacy dataset_infos.json

Browse files
Files changed (1) hide show
  1. dataset_infos.json +0 -91
dataset_infos.json DELETED
@@ -1,91 +0,0 @@
1
- {
2
- "default": {
3
- "description": "MedMCQA is a large-scale, Multiple-Choice Question Answering (MCQA) dataset designed to address real-world medical entrance exam questions.\nMedMCQA has more than 194k high-quality AIIMS & NEET PG entrance exam MCQs covering 2.4k healthcare topics and 21 medical subjects are collected with an average token length of 12.77 and high topical diversity.\nThe dataset contains questions about the following topics: Anesthesia, Anatomy, Biochemistry, Dental, ENT, Forensic Medicine (FM)\nObstetrics and Gynecology (O&G), Medicine, Microbiology, Ophthalmology, Orthopedics Pathology, Pediatrics, Pharmacology, Physiology,\nPsychiatry, Radiology Skin, Preventive & Social Medicine (PSM) and Surgery\n",
4
- "citation": "@InProceedings{pmlr-v174-pal22a,\n title = \t {MedMCQA: A Large-scale Multi-Subject Multi-Choice Dataset for Medical domain Question Answering},\n author = {Pal, Ankit and Umapathi, Logesh Kumar and Sankarasubbu, Malaikannan},\n booktitle = \t {Proceedings of the Conference on Health, Inference, and Learning},\n pages = \t {248--260},\n year = \t {2022},\n editor = \t {Flores, Gerardo and Chen, George H and Pollard, Tom and Ho, Joyce C and Naumann, Tristan},\n volume = \t {174},\n series = \t {Proceedings of Machine Learning Research},\n month = \t {07--08 Apr},\n publisher = {PMLR},\n pdf = \t {https://proceedings.mlr.press/v174/pal22a/pal22a.pdf},\n url = \t {https://proceedings.mlr.press/v174/pal22a.html},\n abstract = \t {This paper introduces MedMCQA, a new large-scale, Multiple-Choice Question Answering (MCQA) dataset designed to address real-world medical entrance exam questions. More than 194k high-quality AIIMS & NEET PG entrance exam MCQs covering 2.4k healthcare topics and 21 medical subjects are collected with an average token length of 12.77 and high topical diversity. Each sample contains a question, correct answer(s), and other options which requires a deeper language understanding as it tests the 10+ reasoning abilities of a model across a wide range of medical subjects & topics. A detailed explanation of the solution, along with the above information, is provided in this study.}\n}\n",
5
- "homepage": "https://medmcqa.github.io",
6
- "license": "Apache License 2.0",
7
- "features": {
8
- "id": {
9
- "dtype": "string",
10
- "_type": "Value"
11
- },
12
- "question": {
13
- "dtype": "string",
14
- "_type": "Value"
15
- },
16
- "opa": {
17
- "dtype": "string",
18
- "_type": "Value"
19
- },
20
- "opb": {
21
- "dtype": "string",
22
- "_type": "Value"
23
- },
24
- "opc": {
25
- "dtype": "string",
26
- "_type": "Value"
27
- },
28
- "opd": {
29
- "dtype": "string",
30
- "_type": "Value"
31
- },
32
- "cop": {
33
- "names": [
34
- "a",
35
- "b",
36
- "c",
37
- "d"
38
- ],
39
- "_type": "ClassLabel"
40
- },
41
- "choice_type": {
42
- "dtype": "string",
43
- "_type": "Value"
44
- },
45
- "exp": {
46
- "dtype": "string",
47
- "_type": "Value"
48
- },
49
- "subject_name": {
50
- "dtype": "string",
51
- "_type": "Value"
52
- },
53
- "topic_name": {
54
- "dtype": "string",
55
- "_type": "Value"
56
- }
57
- },
58
- "builder_name": "medmcqa",
59
- "dataset_name": "medmcqa",
60
- "config_name": "default",
61
- "version": {
62
- "version_str": "1.1.0",
63
- "major": 1,
64
- "minor": 1,
65
- "patch": 0
66
- },
67
- "splits": {
68
- "train": {
69
- "name": "train",
70
- "num_bytes": 131903297,
71
- "num_examples": 182822,
72
- "dataset_name": null
73
- },
74
- "test": {
75
- "name": "test",
76
- "num_bytes": 1399350,
77
- "num_examples": 6150,
78
- "dataset_name": null
79
- },
80
- "validation": {
81
- "name": "validation",
82
- "num_bytes": 2221428,
83
- "num_examples": 4183,
84
- "dataset_name": null
85
- }
86
- },
87
- "download_size": 88311487,
88
- "dataset_size": 135524075,
89
- "size_in_bytes": 223835562
90
- }
91
- }