root commited on
Commit
e99541e
1 Parent(s): 467d25a

update split naming

Browse files
Files changed (1) hide show
  1. xCodeEval.py +7 -7
xCodeEval.py CHANGED
@@ -107,8 +107,8 @@ _DESCRIPTIONS = {
107
  13. `apr_id`: A unique ID for the apr sample. It is not important for model training. If you find any issue with the sample, you can report it to us mentioning the `apr_id`.
108
  14. `difficulty`: Difficulty rating of the problem indicated by `src_uid`. The higher the harder.
109
  15. `tags`: List of potential algorithmic techniques required to write the program.
110
- 16. `bug_exec_outcome`: A pre-run execution outcome of `bug_source_code`. Follow [Section 4.1](https://arxiv.org/pdf/2303.03004.pdf) to know the potential list of outcomes. The `exec_outcome` flags in the training data comes from a pre-run environmeent. However, training data doesn't includes unit-test to avoid potential hacks. We provide unit test for only dev and test data.
111
- 17. `fix_exec_outcome`: A pre-run execution outcome of `fix_source_code`. Follow [Section 4.1](https://arxiv.org/pdf/2303.03004.pdf) to know the potential list of outcomes. The `exec_outcome` flags in the training data comes from a pre-run environmeent. However, training data doesn't includes unit-test to avoid potential hacks. We provide unit test for only dev and test data.
112
  18. `potential_dominant_fix_op`: A potential fix op recommended by difflib.
113
  19. `lang_cluster`: A generic programming language name the value of `lang` belongs to.
114
  20. `prob_desc_description`: Problem description in textual format, math operations are written in latex.
@@ -133,7 +133,7 @@ _DESCRIPTIONS = {
133
  3. `code_uid`: A unique ID for the sample. It is not important for model training. If you find any issue with the sample, you can report it to us mentioning the `code_uid`.
134
  4. `src_uid`: A specific identifier that shows which problem the code is associated with. This identifier is **important** for the training of the model. The problem referred to by the `src_uid` provides a natural description of the problem that the code successfully solved. Refer to [Structure of `problem_descriptions.jsonl`](./README.md#structure-of-problem_descriptionsjsonl)
135
  5. `difficulty`: Difficulty rating of the problem indicated by `src_uid`. The higher the harder.
136
- 6. `exec_outcome`: Execution outcome status. Follow [Section 4.1](https://arxiv.org/pdf/2303.03004.pdf) to know the potential list of outcomes. The `exec_outcome` flags in the training data comes from a pre-run environmeent. However, training data doesn't includes unit-test to avoid potential hacks. We provide unit test for only dev and test data.
137
  7. `lang_cluster`: A generic programming language name the value of `lang` belongs to.
138
  8. `prob_desc_description`: Problem description in textual format, math operations are written in latex.
139
  9. `prob_desc_input_from`: How the program should take the unit test.
@@ -159,7 +159,7 @@ _DESCRIPTIONS = {
159
  5. `src_uid`: A specific identifier that shows which problem the code is associated with. This identifier is **important** for the training of the model. The problem referred to by the `src_uid` provides a natural description of the problem that the code successfully solved. Refer to [Structure of `problem_descriptions.jsonl`](./README.md#structure-of-problem_descriptionsjsonl)
160
  6. `code_uid`: A unique ID for the sample. It is not important for model training. If you find any issue with the sample, you can report it to us mentioning the `code_uid`.
161
  7. `difficulty`: Difficulty rating of the problem indicated by `src_uid`. The higher the harder.
162
- 8. `exec_outcome`: Execution outcome status. Follow [Section 4.1](https://arxiv.org/pdf/2303.03004.pdf) to know the potential list of outcomes. The `exec_outcome` flags in the training data comes from a pre-run environmeent. However, training data doesn't includes unit-test to avoid potential hacks. We provide unit test for only dev and test data.
163
  9. `prob_desc_description`: Problem description in textual format, math operations are written in latex.
164
  10. `prob_desc_input_from`: How the program should take the unit test.
165
  11. `prob_desc_output_to`: Where the program should output the result of the unit test.
@@ -2437,7 +2437,7 @@ class xCodeEval(datasets.GeneratorBasedBuilder):
2437
  },
2438
  ),
2439
  datasets.SplitGenerator(
2440
- name=datasets.Split.VALIDATION,
2441
  gen_kwargs={
2442
  "filepaths": validation_downloaded_files,
2443
  "problem_description_file": prob_desc_file,
@@ -2445,7 +2445,7 @@ class xCodeEval(datasets.GeneratorBasedBuilder):
2445
  },
2446
  ),
2447
  datasets.SplitGenerator(
2448
- name=datasets.Split.TEST,
2449
  gen_kwargs={
2450
  "filepaths": test_downloaded_files,
2451
  "problem_description_file": prob_desc_file,
@@ -2456,7 +2456,7 @@ class xCodeEval(datasets.GeneratorBasedBuilder):
2456
  if task_name == "code_translation":
2457
  split_info.append(
2458
  datasets.SplitGenerator(
2459
- name="validation_small",
2460
  gen_kwargs={
2461
  "filepaths": validation_small_downloaded_files,
2462
  "problem_description_file": prob_desc_file,
 
107
  13. `apr_id`: A unique ID for the apr sample. It is not important for model training. If you find any issue with the sample, you can report it to us mentioning the `apr_id`.
108
  14. `difficulty`: Difficulty rating of the problem indicated by `src_uid`. The higher the harder.
109
  15. `tags`: List of potential algorithmic techniques required to write the program.
110
+ 16. `bug_exec_outcome`: A pre-run execution outcome of `bug_source_code`. Follow [Section 4.1](https://arxiv.org/pdf/2303.03004.pdf) to know the potential list of outcomes. The `exec_outcome` flags in the training data comes from a pre-run environmeent. However, training data doesn't includes unit-test to avoid potential hacks. We provide unit test for only compact and titan test data.
111
+ 17. `fix_exec_outcome`: A pre-run execution outcome of `fix_source_code`. Follow [Section 4.1](https://arxiv.org/pdf/2303.03004.pdf) to know the potential list of outcomes. The `exec_outcome` flags in the training data comes from a pre-run environmeent. However, training data doesn't includes unit-test to avoid potential hacks. We provide unit test for only compact and titan test data.
112
  18. `potential_dominant_fix_op`: A potential fix op recommended by difflib.
113
  19. `lang_cluster`: A generic programming language name the value of `lang` belongs to.
114
  20. `prob_desc_description`: Problem description in textual format, math operations are written in latex.
 
133
  3. `code_uid`: A unique ID for the sample. It is not important for model training. If you find any issue with the sample, you can report it to us mentioning the `code_uid`.
134
  4. `src_uid`: A specific identifier that shows which problem the code is associated with. This identifier is **important** for the training of the model. The problem referred to by the `src_uid` provides a natural description of the problem that the code successfully solved. Refer to [Structure of `problem_descriptions.jsonl`](./README.md#structure-of-problem_descriptionsjsonl)
135
  5. `difficulty`: Difficulty rating of the problem indicated by `src_uid`. The higher the harder.
136
+ 6. `exec_outcome`: Execution outcome status. Follow [Section 4.1](https://arxiv.org/pdf/2303.03004.pdf) to know the potential list of outcomes. The `exec_outcome` flags in the training data comes from a pre-run environmeent. However, training data doesn't includes unit-test to avoid potential hacks. We provide unit test for only compact and titan test data.
137
  7. `lang_cluster`: A generic programming language name the value of `lang` belongs to.
138
  8. `prob_desc_description`: Problem description in textual format, math operations are written in latex.
139
  9. `prob_desc_input_from`: How the program should take the unit test.
 
159
  5. `src_uid`: A specific identifier that shows which problem the code is associated with. This identifier is **important** for the training of the model. The problem referred to by the `src_uid` provides a natural description of the problem that the code successfully solved. Refer to [Structure of `problem_descriptions.jsonl`](./README.md#structure-of-problem_descriptionsjsonl)
160
  6. `code_uid`: A unique ID for the sample. It is not important for model training. If you find any issue with the sample, you can report it to us mentioning the `code_uid`.
161
  7. `difficulty`: Difficulty rating of the problem indicated by `src_uid`. The higher the harder.
162
+ 8. `exec_outcome`: Execution outcome status. Follow [Section 4.1](https://arxiv.org/pdf/2303.03004.pdf) to know the potential list of outcomes. The `exec_outcome` flags in the training data comes from a pre-run environmeent. However, training data doesn't includes unit-test to avoid potential hacks. We provide unit test for only compact and titan test data.
163
  9. `prob_desc_description`: Problem description in textual format, math operations are written in latex.
164
  10. `prob_desc_input_from`: How the program should take the unit test.
165
  11. `prob_desc_output_to`: Where the program should output the result of the unit test.
 
2437
  },
2438
  ),
2439
  datasets.SplitGenerator(
2440
+ name='compact',
2441
  gen_kwargs={
2442
  "filepaths": validation_downloaded_files,
2443
  "problem_description_file": prob_desc_file,
 
2445
  },
2446
  ),
2447
  datasets.SplitGenerator(
2448
+ name='titan',
2449
  gen_kwargs={
2450
  "filepaths": test_downloaded_files,
2451
  "problem_description_file": prob_desc_file,
 
2456
  if task_name == "code_translation":
2457
  split_info.append(
2458
  datasets.SplitGenerator(
2459
+ name="compact_small",
2460
  gen_kwargs={
2461
  "filepaths": validation_small_downloaded_files,
2462
  "problem_description_file": prob_desc_file,