| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| import os |
| import shutil |
| from dataclasses import dataclass, field |
| from enum import Enum |
| from pathlib import Path |
| from typing import Optional |
|
|
|
|
| class EvalTaskConfig(Enum): |
| NUTPOURING = ( |
| "Isaac-NutPour-GR1T2-ClosedLoop-v0", |
| "/home/gr00t/GR00T-N1-2B-tuned-Nut-Pouring-task", |
| ( |
| "Pick up the beaker and tilt it to pour out 1 metallic nut into the bowl. Pick up the bowl and place it on" |
| " the metallic measuring scale." |
| ), |
| "nut_pouring_task.hdf5", |
| 0 |
| ) |
| PIPESORTING = ( |
| "Isaac-ExhaustPipe-GR1T2-ClosedLoop-v0", |
| "/home/gr00t/GR00T-N1-2B-tuned-Exhaust-Pipe-Sorting-task", |
| "Pick up the blue pipe and place it into the blue bin.", |
| "exhaust_pipe_sorting_task.hdf5", |
| 2 |
| ) |
| PICKPLACE_LARGE = ( |
| "Isaac-PickPlace-Camera-G1-v0", |
| "~/IsaacLabEvalTasks/datasets/Isaac-PickPlace-Camera-G1-v0", |
| "Pick up the steering wheel and place it into the basket.", |
| "generated_dataset_pick_place_camera_g1.hdf5", |
| 3 |
| ) |
| APPLE_LARGE = ( |
| "Isaac-Apple-PickPlace-G1-v0", |
| "~/IsaacLabEvalTasks/datasets/Isaac-Apple-PickPlace-G1-v0", |
| "Pick up the apple and place it on the plate.", |
| "apple_pick_place_generated.hdf5", |
| 4 |
| ) |
| |
| APPLE_5 = ( |
| "Isaac-Apple-PickPlace-G1-v0", |
| "~/isaaclabevaltasks/datasets", |
| "Pick up the apple and place it on the plate.", |
| "apple_pick_place_annotated.hdf5", |
| 5 |
| ) |
| |
| APPLE_20 = ( |
| "Isaac-Apple-PickPlace-G1-v0", |
| "~/isaaclabevaltasks/datasets", |
| "Pick up the apple and place it on the plate.", |
| "apple_pick_place_generated_small.hdf5", |
| 5 |
| ) |
| |
| STEERING_WHEEL = ( |
| "Isaac-PickPlace-Camera-G1-Mimic-v0", |
| "~/isaaclab/datasets", |
| "Pick up the steering wheel and place it on the basket.", |
| "steering_wheel_generated.hdf5", |
| 6 |
| ) |
|
|
| def __init__(self, task: str, model_path: str, language_instruction: str, hdf5_name: str, task_index: int): |
| self.task = task |
| self.model_path = model_path |
| self.language_instruction = language_instruction |
| self.hdf5_name = hdf5_name |
| assert task_index != 1, "task_index must not be 1. (Use 0 for nutpouring, 2 for exhaustpipe, etc.)" |
| self.task_index = task_index |
|
|
| @dataclass |
| class Gr00tN1ClosedLoopArguments: |
| |
| headless: bool = field( |
| default=False, metadata={"description": "Whether to run the simulator in headless (no GUI) mode."} |
| ) |
| num_envs: int = field(default=10, metadata={"description": "Number of environments to run in parallel."}) |
| enable_pinocchio: bool = field( |
| default=True, |
| metadata={ |
| "description": ( |
| "Whether to use Pinocchio for physics simulation. Required for NutPouring and ExhaustPipe tasks." |
| ) |
| }, |
| ) |
| record_camera: bool = field( |
| default=False, |
| metadata={"description": "Whether to record the camera images as videos during evaluation."}, |
| ) |
| record_video_output_path: str = field( |
| default="videos/", |
| metadata={"description": "Path to save the recorded videos."}, |
| ) |
|
|
| |
| task_name: str = field( |
| default="nutpouring", metadata={"description": "Short name of the task to run (e.g., nutpouring, exhaustpipe)."} |
| ) |
| task: str = field(default="", metadata={"description": "Full task name for the gym-registered environment."}) |
| language_instruction: str = field( |
| default="", metadata={"description": "Instruction given to the policy in natural language."} |
| ) |
| model_path: str = field(default="", metadata={"description": "Full path to the tuned model checkpoint directory."}) |
| action_horizon: int = field( |
| default=16, metadata={"description": "Number of actions in the policy's predictionhorizon."} |
| ) |
| embodiment_tag: str = field( |
| default="g1", |
| metadata={ |
| "description": ( |
| "Identifier for the robot embodiment used in the policy inference (e.g., 'g1' or 'new_embodiment')." |
| ) |
| }, |
| ) |
| denoising_steps: int = field( |
| default=4, metadata={"description": "Number of denoising steps used in the policy inference."} |
| ) |
| data_config: str = field( |
| default="g1", metadata={"description": "Name of the data configuration to use for the policy."} |
| ) |
| original_image_size: tuple[int, int, int] = field( |
| default=(160, 256, 3), metadata={"description": "Original size of input images as (height, width, channels)."} |
| ) |
| target_image_size: tuple[int, int, int] = field( |
| default=(256, 256, 3), |
| metadata={"description": "Target size for images after resizing and padding as (height, width, channels)."}, |
| ) |
| gr00t_joints_config_path: Path = field( |
| default=Path(__file__).parent.resolve() / "gr00t_g1" / "gr00t_joint_space.yaml", |
| metadata={"description": "Path to the YAML file specifying the joint ordering configuration for GR00T policy."}, |
| ) |
|
|
| |
| action_joints_config_path: Path = field( |
| default=Path(__file__).parent.resolve() / "g1" / "action_joint_space.yaml", |
| metadata={ |
| "description": ( |
| "Path to the YAML file specifying the joint ordering configuration for G1 action space in Lab." |
| ) |
| }, |
| ) |
| state_joints_config_path: Path = field( |
| default=Path(__file__).parent.resolve() / "g1" / "state_joint_space.yaml", |
| metadata={ |
| "description": ( |
| "Path to the YAML file specifying the joint ordering configuration for G1 state space in Lab." |
| ) |
| }, |
| ) |
|
|
| |
| policy_device: str = field( |
| default="cuda", metadata={"description": "Device to run the policy model on (e.g., 'cuda' or 'cpu')."} |
| ) |
| simulation_device: str = field( |
| default="cpu", metadata={"description": "Device to run the physics simulation on (e.g., 'cpu' or 'cuda')."} |
| ) |
|
|
| |
| max_num_rollouts: int = field( |
| default=100, metadata={"description": "Maximum number of rollouts to perform during evaluation."} |
| ) |
| checkpoint_name: str = field( |
| default="gr00t-n1-2b-tuned", metadata={"description": "Name of the model checkpoint used for evaluation."} |
| ) |
| eval_file_path: Optional[str] = field( |
| default=None, metadata={"description": "Path to the file where evaluation results will be saved."} |
| ) |
|
|
| |
| num_feedback_actions: int = field( |
| default=16, |
| metadata={ |
| "description": "Number of feedback actions to execute per rollout (can be less than action_horizon)." |
| }, |
| ) |
| rollout_length: int = field(default=30, metadata={"description": "Number of steps in each rollout episode."}) |
| seed: int = field(default=10, metadata={"description": "Random seed for reproducibility."}) |
|
|
| def __post_init__(self): |
| |
| if self.task_name.upper() not in EvalTaskConfig.__members__: |
| raise ValueError(f"task_name must be one of: {', '.join(EvalTaskConfig.__members__.keys())}") |
| config = EvalTaskConfig[self.task_name.upper()] |
| if self.task == "": |
| self.task = config.task |
| if self.model_path == "": |
| self.model_path = config.model_path |
| if self.language_instruction == "": |
| self.language_instruction = config.language_instruction |
| |
| if not os.path.isabs(self.model_path): |
| raise ValueError("model_path must be an absolute path. Do not use relative paths.") |
| assert ( |
| self.num_feedback_actions <= self.action_horizon |
| ), "num_feedback_actions must be less than or equal to action_horizon" |
| |
| assert Path(self.gr00t_joints_config_path).exists(), "gr00t_joints_config_path does not exist" |
| assert Path(self.action_joints_config_path).exists(), "action_joints_config_path does not exist" |
| assert Path(self.state_joints_config_path).exists(), "state_joints_config_path does not exist" |
| assert Path(self.model_path).exists(), "model_path does not exist." |
| |
| assert self.embodiment_tag in [ |
| "g1", |
| "new_embodiment", |
| ], "embodiment_tag must be one of the following: " + ", ".join(["g1", "new_embodiment"]) |
|
|
|
|
| @dataclass |
| class Gr00tN1DatasetConfig: |
| |
| data_root: Path = field( |
| default=Path("/mnt/datab/PhysicalAI-GR00T-Tuned-Tasks"), |
| metadata={"description": "Root directory for all data storage."}, |
| ) |
| task_name: str = field( |
| default="nutpouring", metadata={"description": "Short name of the task to run (e.g., nutpouring, exhaustpipe)."} |
| ) |
| language_instruction: str = field( |
| default="", metadata={"description": "Instruction given to the policy in natural language."} |
| ) |
| hdf5_name: str = field(default="", metadata={"description": "Name of the HDF5 file to use for the dataset."}) |
|
|
| |
| state_name_sim: str = field( |
| default="robot_joint_pos", metadata={"description": "Name of the state in the HDF5 file."} |
| ) |
| action_name_sim: str = field( |
| default="processed_actions", metadata={"description": "Name of the action in the HDF5 file."} |
| ) |
| pov_cam_name_sim: str = field( |
| default="robot_pov_cam", metadata={"description": "Name of the POV camera in the HDF5 file."} |
| ) |
| |
| state_name_lerobot: str = field( |
| default="observation.state", metadata={"description": "Name of the state in the LeRobot file."} |
| ) |
| action_name_lerobot: str = field( |
| default="action", metadata={"description": "Name of the action in the LeRobot file."} |
| ) |
| video_name_lerobot: str = field( |
| default="observation.images.ego_view", metadata={"description": "Name of the video in the LeRobot file."} |
| ) |
| task_description_lerobot: str = field( |
| default="annotation.human.action.task_description", |
| metadata={"description": "Name of the task description in the LeRobot file."}, |
| ) |
| valid_lerobot: str = field( |
| default="annotation.human.action.valid", metadata={"description": "Name of the validity in the LeRobot file."} |
| ) |
|
|
| |
| chunks_size: int = field(default=1000, metadata={"description": "Number of episodes per data chunk."}) |
| |
| fps: int = field(default=20, metadata={"description": "Frames per second for video recording."}) |
| |
| data_path: str = field( |
| default="data/chunk-{episode_chunk:03d}/episode_{episode_index:06d}.parquet", |
| metadata={"description": "Template path for storing episode data files."}, |
| ) |
| video_path: str = field( |
| default="videos/chunk-{episode_chunk:03d}/{video_key}/episode_{episode_index:06d}.mp4", |
| metadata={"description": "Template path for storing episode video files."}, |
| ) |
| modality_template_path: Path = field( |
| default=Path(__file__).parent.resolve() / "gr00t_g1" / "modality.json", |
| metadata={"description": "Path to the modality template JSON file."}, |
| ) |
| modality_fname: str = field( |
| default="modality.json", metadata={"description": "Filename for the modality JSON file."} |
| ) |
| episodes_fname: str = field( |
| default="episodes.jsonl", metadata={"description": "Filename for the episodes JSONL file."} |
| ) |
| tasks_fname: str = field(default="tasks.jsonl", metadata={"description": "Filename for the tasks JSONL file."}) |
| info_template_path: Path = field( |
| default=Path(__file__).parent.resolve() / "gr00t_g1" / "info.json", |
| metadata={"description": "Path to the info template JSON file."}, |
| ) |
| info_fname: str = field(default="info.json", metadata={"description": "Filename for the info JSON file."}) |
| |
| gr00t_joints_config_path: Path = field( |
| default=Path(__file__).parent.resolve() / "gr00t_g1" / "gr00t_joint_space.yaml", |
| metadata={"description": "Path to the YAML file specifying the joint ordering configuration for GR00T policy."}, |
| ) |
| robot_type: str = field( |
| default="g1", metadata={"description": "Type of robot embodiment used in the policy fine-tuning."} |
| ) |
| |
| action_joints_config_path: Path = field( |
| default=Path(__file__).parent.resolve() / "g1" / "action_joint_space.yaml", |
| metadata={ |
| "description": ( |
| "Path to the YAML file specifying the joint ordering configuration for G1 action space in Lab." |
| ) |
| }, |
| ) |
| state_joints_config_path: Path = field( |
| default=Path(__file__).parent.resolve() / "g1" / "state_joint_space.yaml", |
| metadata={ |
| "description": ( |
| "Path to the YAML file specifying the joint ordering configuration for G1 state space in Lab." |
| ) |
| }, |
| ) |
| original_image_size: tuple[int, int, int] = field( |
| default=(160, 256, 3), metadata={"description": "Original size of input images as (height, width, channels)."} |
| ) |
| target_image_size: tuple[int, int, int] = field( |
| default=(256, 256, 3), metadata={"description": "Target size for images after resizing and padding."} |
| ) |
|
|
| hdf5_file_path: Path = field(init=False) |
| lerobot_data_dir: Path = field(init=False) |
| task_index: int = field(init=False) |
|
|
| def __post_init__(self): |
|
|
| |
| if self.task_name.upper() not in EvalTaskConfig.__members__: |
| raise ValueError(f"task_name must be one of: {', '.join(EvalTaskConfig.__members__.keys())}") |
| config = EvalTaskConfig[self.task_name.upper()] |
| self.language_instruction = config.language_instruction |
| self.hdf5_name = config.hdf5_name |
| self.task_index = config.task_index |
|
|
| self.hdf5_file_path = self.data_root / self.hdf5_name |
| self.lerobot_data_dir = self.data_root / self.hdf5_name.replace(".hdf5", "") / "lerobot" |
|
|
| |
| assert self.hdf5_file_path.exists(), "hdf5_file_path does not exist" |
| assert Path(self.gr00t_joints_config_path).exists(), "gr00t_joints_config_path does not exist" |
| assert Path(self.action_joints_config_path).exists(), "action_joints_config_path does not exist" |
| assert Path(self.state_joints_config_path).exists(), "state_joints_config_path does not exist" |
| assert Path(self.info_template_path).exists(), "info_template_path does not exist" |
| assert Path(self.modality_template_path).exists(), "modality_template_path does not exist" |
| |
| if self.lerobot_data_dir.exists(): |
| print(f"Warning: lerobot_data_dir {self.lerobot_data_dir} already exists. Removing it.") |
| |
| shutil.rmtree(self.lerobot_data_dir) |
| |
| self.hdf5_keys = { |
| "state": self.state_name_sim, |
| "action": self.action_name_sim, |
| } |
| |
| self.lerobot_keys = { |
| "state": self.state_name_lerobot, |
| "action": self.action_name_lerobot, |
| "video": self.video_name_lerobot, |
| "annotation": ( |
| self.task_description_lerobot, |
| self.valid_lerobot, |
| ), |
| } |
|
|