Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
File size: 1,516 Bytes
08ae6c5 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 |
import json
import os
import logging
from datetime import datetime
from lighteval.main_accelerate import main
from src.envs import RESULTS_REPO, CACHE_PATH
from src.backend.manage_requests import EvalRequest
logging.getLogger("openai").setLevel(logging.WARNING)
def run_evaluation(eval_request: EvalRequest, task_names: str, batch_size: int, local_dir: str, accelerator: str, region: str, vendor: str, instance_size: str, instance_type: str, limit=None):
if limit:
print("WARNING: --limit SHOULD ONLY BE USED FOR TESTING. REAL METRICS SHOULD NOT BE COMPUTED USING LIMIT.")
results = main(
endpoint_model_name=f"{eval_request.model}_{eval_request.precision}".lower(),
accelerator= accelerator,
vendor= vendor,
region= region,
instance_size= instance_size,
instance_type= instance_type,
max_samples= limit,
job_id= str(datetime.now()),
push_results_to_hub= True,
save_details= True,
push_details_to_hub= True,
public_run= False,
cache_dir= CACHE_PATH,
results_org= RESULTS_REPO,
output_dir= local_dir,
override_batch_size= batch_size,
custom_tasks= "custom_tasks.py",
tasks= task_names
)
results["config"]["model_dtype"] = eval_request.precision
results["config"]["model_name"] = eval_request.model
results["config"]["model_sha"] = eval_request.revision
dumped = json.dumps(results, indent=2)
print(dumped)
return results
|