| #!/bin/bash |
| |
| |
| |
| |
| |
|
|
| set -e |
|
|
| |
| |
| |
| EXPERIMENT_NAME_PREFIX="alphaevolve_ac_mmv1_3" |
| NUM_GENERATIONS=200 |
| MAX_PARALLEL_JOBS=5 |
| META_INTERVAL=10 |
|
|
| LLM_MODELS="native-gemini-3-flash-preview" |
| LLM_SELECTION="ucb1" |
| LLM_TEMPERATURES="0.0 0.5 1.0" |
| LLM_MAX_TOKENS=65536 |
|
|
| USE_EVAL_SERVICE="--use-eval-service" |
| EVAL_SERVICE_URL="http://localhost:8711" |
| EVAL_TRIGGER_MODE="periodic" |
| EVAL_TRIGGER_INTERVAL=10 |
|
|
| EXPERIMENT_NAME=${EXPERIMENT_NAME_PREFIX}_gen${NUM_GENERATIONS}_${EVAL_TRIGGER_MODE}${EVAL_TRIGGER_INTERVAL} |
|
|
| USE_WANDB="--use-wandb" |
| WANDB_PROJECT="ev2" |
| WANDB_TAGS="alphaevolve_ac eval-service full-experiment ${EVAL_TRIGGER_MODE}" |
|
|
| |
| |
| |
| .venv/bin/python tasks/alphaevolve_ac/run_experiment.py \ |
| --experiment-name "$EXPERIMENT_NAME" \ |
| --num-generations "$NUM_GENERATIONS" \ |
| --max-parallel-jobs "$MAX_PARALLEL_JOBS" \ |
| --meta-interval "$META_INTERVAL" \ |
| --llm-models $LLM_MODELS \ |
| --llm-selection "$LLM_SELECTION" \ |
| --llm-temperatures $LLM_TEMPERATURES \ |
| --llm-max-tokens "$LLM_MAX_TOKENS" \ |
| $USE_EVAL_SERVICE \ |
| --eval-service-url "$EVAL_SERVICE_URL" \ |
| --eval-trigger-mode "$EVAL_TRIGGER_MODE" \ |
| --eval-trigger-interval "$EVAL_TRIGGER_INTERVAL" \ |
| $USE_WANDB \ |
| --wandb-project "$WANDB_PROJECT" \ |
| --wandb-tags $WANDB_TAGS \ |
| --verbose |
|
|