esctr-environment / train_4b.py
musharraf7's picture
Upload folder using huggingface_hub
08a3b81 verified
#!/usr/bin/env python3
"""
ESCTR 4B Training β€” GRPO + LoRA on RTX 4090 (24 GB)
====================================================
Self-contained training script. No imports from other train scripts.
Memory budget (RTX 4090, 24 GB):
Qwen3-4B in bf16 β‰ˆ 8 GB
LoRA adapters β‰ˆ 0.05 GB
KV cache (K=2, 512 tok) β‰ˆ 3 GB
Grad checkpointing β‰ˆ 3 GB
Optimizer (LoRA only) β‰ˆ 0.2 GB
──────────────────────────────────
Total β‰ˆ 14 GB (plenty of headroom)
Usage on RunPod:
chmod +x setup_runpod.sh && ./setup_runpod.sh
python train_4b.py
"""
import os
import sys
import random
import time
# ── Memory + cache config (must be set BEFORE any torch import) ───────────
os.environ["PYTORCH_CUDA_ALLOC_CONF"] = "expandable_segments:True"
# Redirect HF + torch caches to the CONTAINER DISK (not /workspace volume,
# which is often only 5 GB on RunPod β€” too small for 4B model weights).
# The container disk is typically 50 GB at /.
cache_root = "/root" if os.path.isdir("/root") else "."
os.environ.setdefault("HF_HOME", os.path.join(cache_root, "hf_cache"))
os.environ.setdefault("TORCH_HOME", os.path.join(cache_root, "torch_cache"))
os.makedirs(os.environ["HF_HOME"], exist_ok=True)
os.makedirs(os.environ["TORCH_HOME"], exist_ok=True)
import torch
from datasets import Dataset
from peft import LoraConfig
from trl import GRPOConfig, GRPOTrainer
# ── Import ESCTR environment (in-process, no server) ─────────────────────
sys.path.insert(0, os.path.dirname(os.path.abspath(__file__)))
from server.environment import ESCTREnvironment
from server.models import ESCTRAction
# ── System prompt ─────────────────────────────────────────────────────────
SYSTEM_PROMPT = """You are an autonomous Financial Controller AI operating within an enterprise ERP system.
Your job is to investigate financial discrepancies in procurement records by using the available tools, then submit a precise monetary adjustment.
INVESTIGATION WORKFLOW:
1. Query databases to discover what records exist (purchase_orders, invoices, shipping_logs, sla_contracts, warehouse_logs)
2. Read specific documents to get full details
3. Compare line items, delivery dates, and contract terms
4. Calculate the exact adjustment amount
5. Submit your financial decision with the calculated amount and reasoning
CRITICAL RULES:
- Always query AND read documents before submitting. Never guess.
- Your adjustment_amount must be the EXACT monetary difference you calculated.
- Show your arithmetic in the adjustment_reason.
- If a vendor offers a settlement, verify their claims against internal records before accepting.
You have access to the following tools. Call them to interact with the ERP system."""
# ── Task config ───────────────────────────────────────────────────────────
TRAIN_TASKS = [
t.strip()
for t in os.environ.get(
"ESCTR_TASKS",
os.environ.get("ESCTR_TASK", "procurement_reconciliation")
).split(",")
if t.strip()
]
# ── TRL environment wrapper ──────────────────────────────────────────────
class ESCTRToolEnv:
"""TRL-compatible wrapper.
Public methods with docstrings become tools. TRL handles the multi-turn
loop automatically via environment_factory.
"""
def __init__(self):
self.env = ESCTREnvironment()
self.reward = 0.0
self.done = False
self._tasks = TRAIN_TASKS or ["procurement_reconciliation"]
def reset(self, **kwargs) -> str | None:
"""Reset the environment and return the initial briefing."""
seed = random.randint(0, 100_000)
task = random.choice(self._tasks)
obs = self.env.reset(task_name=task, seed=seed)
self.reward = 0.0
self.done = False
return obs.system_response
def query_database(self, table: str) -> str:
"""
Query a corporate database table to discover available records.
Args:
table: The database table to query. One of: 'purchase_orders', 'invoices', 'shipping_logs', 'sla_contracts', 'warehouse_logs'
Returns:
A summary of records found in the specified table.
"""
if self.done:
return "Episode is over."
action = ESCTRAction(action_type="query_database", query_parameters={"table": table})
obs = self.env.step(action)
self.reward = obs.reward
self.done = obs.done
return obs.system_response
def read_document(self, document_id: str) -> str:
"""
Read a specific document by its unique identifier to see full details.
Args:
document_id: The document ID to read, e.g. 'PO-2024-0055' or 'INV-2024-0055'
Returns:
The full contents of the requested document.
"""
if self.done:
return "Episode is over."
action = ESCTRAction(action_type="read_document", document_id=document_id)
obs = self.env.step(action)
self.reward = obs.reward
self.done = obs.done
return obs.system_response
def communicate_vendor(self, message_content: str) -> str:
"""
Send a message to the vendor during a dispute negotiation.
Args:
message_content: The message to send to the vendor, such as requesting clarification or rejecting a settlement offer.
Returns:
The vendor's response to your message.
"""
if self.done:
return "Episode is over."
action = ESCTRAction(action_type="communicate_vendor", message_content=message_content)
obs = self.env.step(action)
self.reward = obs.reward
self.done = obs.done
return obs.system_response
def submit_financial_decision(self, adjustment_amount: float, adjustment_reason: str) -> str:
"""
Submit the final financial adjustment. This is the terminal action that ends the episode.
Args:
adjustment_amount: The exact monetary adjustment amount as a float (e.g. 450.00). Must be calculated from the documents.
adjustment_reason: A brief explanation of why this adjustment is correct, including your arithmetic.
Returns:
The grading result with your score and feedback.
"""
if self.done:
return "Episode is over."
action = ESCTRAction(
action_type="submit_financial_decision",
adjustment_amount=adjustment_amount,
adjustment_reason=adjustment_reason,
)
obs = self.env.step(action)
self.reward = obs.reward
self.done = obs.done
return obs.system_response
# ── Reward function ───────────────────────────────────────────────────────
def reward_func(environments, **kwargs) -> list[float]:
"""Shaped reward for GRPO β€” gives partial credit for investigation progress.
Without shaping, the model must call submit_financial_decision to get ANY
reward. Qwen3-4B never calls submit (it stops after investigating), so
all rewards are 0, GRPO has zero gradient, and nothing is learned.
Shaping:
- Each tool call earns a small bonus (0.05)
- Calling submit earns a larger bonus (0.15) regardless of correctness
- The environment's graded reward (0-1) is added on top
- This creates variance between rollouts even without submission
"""
rewards = []
for env in environments:
# Base: the environment's graded reward (non-zero only if submitted)
r = env.reward
# Shaping: credit for investigation effort
step_count = env.env._state.step_count if hasattr(env.env, '_state') else 0
submitted = env.env._state.outcome_submitted if hasattr(env.env, '_state') else False
# Small per-step bonus for using tools (caps at 0.20)
investigation_bonus = min(step_count * 0.05, 0.20)
# Bonus for actually submitting (even with wrong amount)
submit_bonus = 0.15 if submitted else 0.0
rewards.append(r + investigation_bonus + submit_bonus)
return rewards
# ── Main ──────────────────────────────────────────────────────────────────
def main():
# ── User-configurable via env vars ────────────────────────────────────
model_name = os.environ.get("ESCTR_MODEL", "Qwen/Qwen3-4B")
num_episodes = int(os.environ.get("ESCTR_EPISODES", "300"))
max_len = int(os.environ.get("ESCTR_MAX_COMPLETION_LENGTH", "1024"))
lora_r = int(os.environ.get("ESCTR_LORA_R", "16"))
grad_accum = int(os.environ.get("ESCTR_GRAD_ACCUM", "4"))
# Output goes to container disk (not /workspace which may be tiny)
default_out = "/root/esctr-4b-lora" if os.path.isdir("/root") else "./esctr-4b-lora"
output_dir = os.environ.get("ESCTR_OUTPUT", default_out)
os.makedirs(output_dir, exist_ok=True)
# ── Preflight checks ─────────────────────────────────────────────────
assert torch.cuda.is_available(), "CUDA not available β€” this script requires a GPU."
gpu = torch.cuda.get_device_properties(0)
vram_gb = round(gpu.total_memory / 1024**3, 1)
print(f"\n{'='*60}")
print(f" GPU: {gpu.name} | VRAM: {vram_gb} GB")
print(f" Model: {model_name}")
print(f" Tasks: {', '.join(TRAIN_TASKS)}")
print(f" Episodes: {num_episodes}")
print(f" LoRA rank: {lora_r} | Max completion: {max_len} tokens")
print(f" Output: {output_dir}")
print(f"{'='*60}\n")
# ── Dataset ───────────────────────────────────────────────────────────
dataset = Dataset.from_dict({
"prompt": [[{"role": "user", "content": SYSTEM_PROMPT}]] * num_episodes
})
# ── LoRA config ───────────────────────────────────────────────────────
# Target all linear layers for maximum expressiveness within budget.
# TRL automatically uses base model as reference when peft_config is set,
# so we do NOT need a separate ref model copy in VRAM.
peft_config = LoraConfig(
task_type="CAUSAL_LM",
r=lora_r,
lora_alpha=lora_r * 2, # standard 2x multiplier
target_modules="all-linear", # TRL recommended
lora_dropout=0.05,
bias="none",
)
# ── Detect precision ──────────────────────────────────────────────────
# RTX 4090 (Ada Lovelace) has native bf16 support.
# Fall back to fp16 if something is wrong.
use_bf16 = os.environ.get("ESCTR_BF16", "1") == "1"
try:
if use_bf16:
torch.tensor([1.0], dtype=torch.bfloat16, device="cuda")
except Exception:
print("⚠️ bf16 not supported on this GPU, falling back to fp16")
use_bf16 = False
# ── GRPO config ───────────────────────────────────────────────────────
grpo_config = GRPOConfig(
# Schedule
num_train_epochs=1,
learning_rate=2e-5, # higher LR for LoRA (only adapter weights)
warmup_steps=10,
max_grad_norm=1.0,
optim="adamw_torch",
# Batching β€” keep batch=1, accumulate for effective batch
per_device_train_batch_size=1,
gradient_accumulation_steps=grad_accum,
# GRPO β€” exploration is CRITICAL
# Without high temperature, Qwen3-4B collapses to entropy ~0.0005
# (deterministic) and both K rollouts get identical reward β†’ zero gradient.
num_generations=4, # K=4: more rollouts = more reward variance
max_completion_length=max_len,
temperature=1.5, # Force exploration (default 1.0 is too greedy for 4B)
max_tool_calling_iterations=10, # Allow enough turns for investigate + submit
log_completions=True,
num_completions_to_print=1,
# CRITICAL for Qwen3: disable thinking mode or it burns all tokens
# on <think>...</think> blocks and never makes tool calls
chat_template_kwargs={"enable_thinking": False},
# Memory
gradient_checkpointing=True,
bf16=use_bf16,
fp16=not use_bf16,
# Logging β€” "none" avoids Trackio integration crashes (known issue)
output_dir=output_dir,
report_to="none",
logging_steps=1,
save_steps=50,
save_total_limit=2,
# Do NOT push to hub during training (avoids quota/auth crashes)
push_to_hub=False,
)
# ── Create trainer ────────────────────────────────────────────────────
print("Loading model + LoRA adapters...")
t0 = time.time()
trainer = GRPOTrainer(
model=model_name,
reward_funcs=reward_func,
train_dataset=dataset,
args=grpo_config,
environment_factory=ESCTRToolEnv,
peft_config=peft_config,
)
load_time = time.time() - t0
used = round(torch.cuda.max_memory_reserved() / 1024**3, 2)
print(f"βœ… Model loaded in {load_time:.0f}s | VRAM used: {used} GB / {vram_gb} GB\n")
# ── Train! ────────────────────────────────────────────────────────────
print("πŸš€ Starting GRPO training...\n")
t0 = time.time()
stats = trainer.train()
elapsed = time.time() - t0
peak = round(torch.cuda.max_memory_reserved() / 1024**3, 2)
print(f"\n{'='*60}")
print(f" βœ… Training complete!")
print(f" Wall time: {elapsed/60:.1f} minutes")
print(f" Peak VRAM: {peak} GB / {vram_gb} GB")
if hasattr(stats, 'metrics'):
rt = stats.metrics.get('train_runtime', elapsed)
print(f" Train runtime: {rt:.0f}s")
print(f"{'='*60}\n")
# ── Save ──────────────────────────────────────────────────────────────
save_path = os.path.join(output_dir, "final")
trainer.save_model(save_path)
print(f"πŸ’Ύ LoRA adapters saved to: {save_path}")
print(f" (To push to Hub later: trainer.push_to_hub())")
if __name__ == "__main__":
main()