pminervini commited on
Commit
b1a5839
1 Parent(s): 4a18d8a
src/backend/envs.py CHANGED
@@ -25,9 +25,9 @@ class Tasks(Enum):
25
  task2 = Task("truthfulqa_mc1", "mc1", "TruthfulQA MC1", 0)
26
  task3 = Task("truthfulqa_mc2", "mc2", "TruthfulQA MC2", 0) # TruthfulQA is intended as a zero-shot benchmark [5, 47]. https://owainevans.github.io/pdfs/truthfulQA_lin_evans.pdf
27
 
28
-
29
  # NUM_FEWSHOT = 64 # Change with your few shot
30
 
 
31
  EVAL_REQUESTS_PATH_BACKEND = os.path.join(CACHE_PATH, "eval-queue-bk")
32
  EVAL_RESULTS_PATH_BACKEND = os.path.join(CACHE_PATH, "eval-results-bk")
33
 
 
25
  task2 = Task("truthfulqa_mc1", "mc1", "TruthfulQA MC1", 0)
26
  task3 = Task("truthfulqa_mc2", "mc2", "TruthfulQA MC2", 0) # TruthfulQA is intended as a zero-shot benchmark [5, 47]. https://owainevans.github.io/pdfs/truthfulQA_lin_evans.pdf
27
 
 
28
  # NUM_FEWSHOT = 64 # Change with your few shot
29
 
30
+
31
  EVAL_REQUESTS_PATH_BACKEND = os.path.join(CACHE_PATH, "eval-queue-bk")
32
  EVAL_RESULTS_PATH_BACKEND = os.path.join(CACHE_PATH, "eval-results-bk")
33
 
src/leaderboard/read_evals.py CHANGED
@@ -201,9 +201,6 @@ def get_raw_eval_results(results_path: str, requests_path: str) -> list[EvalResu
201
  except dateutil.parser._parser.ParserError:
202
  files = [files[-1]]
203
 
204
- # if '125m' in root:
205
- # breakpoint()
206
-
207
  for file in files:
208
  model_result_filepaths.append(os.path.join(root, file))
209
 
 
201
  except dateutil.parser._parser.ParserError:
202
  files = [files[-1]]
203
 
 
 
 
204
  for file in files:
205
  model_result_filepaths.append(os.path.join(root, file))
206