Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
import asyncio | |
from src.backend.manage_requests import EvalRequest | |
from src.envs import LIMIT, EVAL_RESULTS_PATH_BACKEND, RESULTS_REPO, DEVICE, LOCAL_MODEL_NAME | |
from src.backend.run_eval_suite import run_evaluation | |
from src.about import HarnessTasks | |
async def run_adhoc_eval(eval_request: EvalRequest): | |
# This job runs lamini tasks and harness tasks | |
TASKS_HARNESS = [task.value.benchmark for task in HarnessTasks] | |
await run_evaluation( | |
eval_request=eval_request, | |
task_names=TASKS_HARNESS, | |
num_fewshot=0, | |
local_dir=EVAL_RESULTS_PATH_BACKEND, | |
results_repo=RESULTS_REPO, | |
batch_size=1, | |
device=DEVICE, | |
no_cache=True, | |
limit=LIMIT | |
) | |
def main(): | |
# eval_request: EvalRequest(model='meta-llama/Llama-2-7b-chat-hf', private=False, status='FINISHED', json_filepath='', weight_type='Original', model_type='\ud83d\udfe2 : pretrained', precision='bfloat16', base_model='', revision='main', submitted_time='2023-11-21T18:10:08Z', likes=0, params=0.1, license='custom') | |
vals = {"model": LOCAL_MODEL_NAME, "json_filepath": "", "base_model": "", "revision": "main", | |
"private": False, | |
"precision": "bfloat16", "weight_type": "Original", "status": "PENDING", | |
"submitted_time": "2023-11-21T18:10:08Z", "model_type": "\ud83d\udfe2 : pretrained", "likes": 0, | |
"params": 0.1, "license": "custom"} | |
eval_request = EvalRequest(**vals) | |
print(f"eval_request: {eval_request}") | |
asyncio.run(run_adhoc_eval(eval_request)) | |
if __name__ == "__main__": | |
main() | |