|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| import math
|
| import time
|
| from dataclasses import dataclass
|
| from typing import Any, Protocol, TypeVar, runtime_checkable
|
|
|
| import numpy as np
|
| import torch
|
| import torchvision.transforms.functional as F
|
|
|
| from lerobot.configs.types import PipelineFeatureType, PolicyFeature
|
| from lerobot.teleoperators.teleoperator import Teleoperator
|
| from lerobot.teleoperators.utils import TeleopEvents
|
|
|
| from .core import EnvTransition, PolicyAction, TransitionKey
|
| from .pipeline import (
|
| ComplementaryDataProcessorStep,
|
| InfoProcessorStep,
|
| ObservationProcessorStep,
|
| ProcessorStep,
|
| ProcessorStepRegistry,
|
| TruncatedProcessorStep,
|
| )
|
|
|
| GRIPPER_KEY = "gripper"
|
| DISCRETE_PENALTY_KEY = "discrete_penalty"
|
| TELEOP_ACTION_KEY = "teleop_action"
|
|
|
|
|
| @runtime_checkable
|
| class HasTeleopEvents(Protocol):
|
| """
|
| Minimal protocol for objects that provide teleoperation events.
|
|
|
| This protocol defines the `get_teleop_events()` method, allowing processor
|
| steps to interact with teleoperators that support event-based controls
|
| (like episode termination or success flagging) without needing to know the
|
| teleoperator's specific class.
|
| """
|
|
|
| def get_teleop_events(self) -> dict[str, Any]:
|
| """
|
| Get extra control events from the teleoperator.
|
|
|
| Returns:
|
| A dictionary containing control events such as:
|
| - `is_intervention`: bool - Whether the human is currently intervening.
|
| - `terminate_episode`: bool - Whether to terminate the current episode.
|
| - `success`: bool - Whether the episode was successful.
|
| - `rerecord_episode`: bool - Whether to rerecord the episode.
|
| """
|
| ...
|
|
|
|
|
|
|
| TeleopWithEvents = TypeVar("TeleopWithEvents", bound=Teleoperator)
|
|
|
|
|
| def _check_teleop_with_events(teleop: Teleoperator) -> None:
|
| """
|
| Runtime check that a teleoperator implements the `HasTeleopEvents` protocol.
|
|
|
| Args:
|
| teleop: The teleoperator instance to check.
|
|
|
| Raises:
|
| TypeError: If the teleoperator does not have a `get_teleop_events` method.
|
| """
|
| if not isinstance(teleop, HasTeleopEvents):
|
| raise TypeError(
|
| f"Teleoperator {type(teleop).__name__} must implement get_teleop_events() method. "
|
| f"Compatible teleoperators: GamepadTeleop, KeyboardEndEffectorTeleop"
|
| )
|
|
|
|
|
| @ProcessorStepRegistry.register("add_teleop_action_as_complementary_data")
|
| @dataclass
|
| class AddTeleopActionAsComplimentaryDataStep(ComplementaryDataProcessorStep):
|
| """
|
| Adds the raw action from a teleoperator to the transition's complementary data.
|
|
|
| This is useful for human-in-the-loop scenarios where the human's input needs to
|
| be available to downstream processors, for example, to override a policy's action
|
| during an intervention.
|
|
|
| Attributes:
|
| teleop_device: The teleoperator instance to get the action from.
|
| """
|
|
|
| teleop_device: Teleoperator
|
|
|
| def complementary_data(self, complementary_data: dict) -> dict:
|
| """
|
| Retrieves the teleoperator's action and adds it to the complementary data.
|
|
|
| Args:
|
| complementary_data: The incoming complementary data dictionary.
|
|
|
| Returns:
|
| A new dictionary with the teleoperator action added under the
|
| `teleop_action` key.
|
| """
|
| new_complementary_data = dict(complementary_data)
|
| new_complementary_data[TELEOP_ACTION_KEY] = self.teleop_device.get_action()
|
| return new_complementary_data
|
|
|
| def transform_features(
|
| self, features: dict[PipelineFeatureType, dict[str, PolicyFeature]]
|
| ) -> dict[PipelineFeatureType, dict[str, PolicyFeature]]:
|
| return features
|
|
|
|
|
| @ProcessorStepRegistry.register("add_teleop_action_as_info")
|
| @dataclass
|
| class AddTeleopEventsAsInfoStep(InfoProcessorStep):
|
| """
|
| Adds teleoperator control events (e.g., terminate, success) to the transition's info.
|
|
|
| This step extracts control events from teleoperators that support event-based
|
| interaction, making these signals available to other parts of the system.
|
|
|
| Attributes:
|
| teleop_device: An instance of a teleoperator that implements the
|
| `HasTeleopEvents` protocol.
|
| """
|
|
|
| teleop_device: TeleopWithEvents
|
|
|
| def __post_init__(self):
|
| """Validates that the provided teleoperator supports events after initialization."""
|
| _check_teleop_with_events(self.teleop_device)
|
|
|
| def info(self, info: dict) -> dict:
|
| """
|
| Retrieves teleoperator events and updates the info dictionary.
|
|
|
| Args:
|
| info: The incoming info dictionary.
|
|
|
| Returns:
|
| A new dictionary including the teleoperator events.
|
| """
|
| new_info = dict(info)
|
|
|
| teleop_events = self.teleop_device.get_teleop_events()
|
| new_info.update(teleop_events)
|
| return new_info
|
|
|
| def transform_features(
|
| self, features: dict[PipelineFeatureType, dict[str, PolicyFeature]]
|
| ) -> dict[PipelineFeatureType, dict[str, PolicyFeature]]:
|
| return features
|
|
|
|
|
| @ProcessorStepRegistry.register("image_crop_resize_processor")
|
| @dataclass
|
| class ImageCropResizeProcessorStep(ObservationProcessorStep):
|
| """
|
| Crops and/or resizes image observations.
|
|
|
| This step iterates through all image keys in an observation dictionary and applies
|
| the specified transformations. It handles device placement, moving tensors to the
|
| CPU if necessary for operations not supported on certain accelerators like MPS.
|
|
|
| Attributes:
|
| crop_params_dict: A dictionary mapping image keys to cropping parameters
|
| (top, left, height, width).
|
| resize_size: A tuple (height, width) to resize all images to.
|
| """
|
|
|
| crop_params_dict: dict[str, tuple[int, int, int, int]] | None = None
|
| resize_size: tuple[int, int] | None = None
|
|
|
| def observation(self, observation: dict) -> dict:
|
| """
|
| Applies cropping and resizing to all images in the observation dictionary.
|
|
|
| Args:
|
| observation: The observation dictionary, potentially containing image tensors.
|
|
|
| Returns:
|
| A new observation dictionary with transformed images.
|
| """
|
| if self.resize_size is None and not self.crop_params_dict:
|
| return observation
|
|
|
| new_observation = dict(observation)
|
|
|
|
|
| for key in observation:
|
| if "image" not in key:
|
| continue
|
|
|
| image = observation[key]
|
| device = image.device
|
|
|
| if device.type == "mps":
|
| image = image.cpu()
|
|
|
| if self.crop_params_dict is not None and key in self.crop_params_dict:
|
| crop_params = self.crop_params_dict[key]
|
| image = F.crop(image, *crop_params)
|
| if self.resize_size is not None:
|
| image = F.resize(image, self.resize_size)
|
| image = image.clamp(0.0, 1.0)
|
| new_observation[key] = image.to(device)
|
|
|
| return new_observation
|
|
|
| def get_config(self) -> dict[str, Any]:
|
| """
|
| Returns the configuration of the step for serialization.
|
|
|
| Returns:
|
| A dictionary with the crop parameters and resize dimensions.
|
| """
|
| return {
|
| "crop_params_dict": self.crop_params_dict,
|
| "resize_size": self.resize_size,
|
| }
|
|
|
| def transform_features(
|
| self, features: dict[PipelineFeatureType, dict[str, PolicyFeature]]
|
| ) -> dict[PipelineFeatureType, dict[str, PolicyFeature]]:
|
| """
|
| Updates the image feature shapes in the policy features dictionary if resizing is applied.
|
|
|
| Args:
|
| features: The policy features dictionary.
|
|
|
| Returns:
|
| The updated policy features dictionary with new image shapes.
|
| """
|
| if self.resize_size is None:
|
| return features
|
| for key in features[PipelineFeatureType.OBSERVATION]:
|
| if "image" in key:
|
| nb_channel = features[PipelineFeatureType.OBSERVATION][key].shape[0]
|
| features[PipelineFeatureType.OBSERVATION][key] = PolicyFeature(
|
| type=features[PipelineFeatureType.OBSERVATION][key].type,
|
| shape=(nb_channel, *self.resize_size),
|
| )
|
| return features
|
|
|
|
|
| @dataclass
|
| @ProcessorStepRegistry.register("time_limit_processor")
|
| class TimeLimitProcessorStep(TruncatedProcessorStep):
|
| """
|
| Tracks episode steps and enforces a time limit by truncating the episode.
|
|
|
| Attributes:
|
| max_episode_steps: The maximum number of steps allowed per episode.
|
| current_step: The current step count for the active episode.
|
| """
|
|
|
| max_episode_steps: int
|
| current_step: int = 0
|
|
|
| def truncated(self, truncated: bool) -> bool:
|
| """
|
| Increments the step counter and sets the truncated flag if the time limit is reached.
|
|
|
| Args:
|
| truncated: The incoming truncated flag.
|
|
|
| Returns:
|
| True if the episode step limit is reached, otherwise the incoming value.
|
| """
|
| self.current_step += 1
|
| if self.current_step >= self.max_episode_steps:
|
| truncated = True
|
|
|
| return truncated
|
|
|
| def get_config(self) -> dict[str, Any]:
|
| """
|
| Returns the configuration of the step for serialization.
|
|
|
| Returns:
|
| A dictionary containing the `max_episode_steps`.
|
| """
|
| return {
|
| "max_episode_steps": self.max_episode_steps,
|
| }
|
|
|
| def reset(self) -> None:
|
| """Resets the step counter, typically called at the start of a new episode."""
|
| self.current_step = 0
|
|
|
| def transform_features(
|
| self, features: dict[PipelineFeatureType, dict[str, PolicyFeature]]
|
| ) -> dict[PipelineFeatureType, dict[str, PolicyFeature]]:
|
| return features
|
|
|
|
|
| @dataclass
|
| @ProcessorStepRegistry.register("gripper_penalty_processor")
|
| class GripperPenaltyProcessorStep(ComplementaryDataProcessorStep):
|
| """
|
| Applies a penalty for inefficient gripper usage.
|
|
|
| This step penalizes actions that attempt to close an already closed gripper or
|
| open an already open one, based on position thresholds.
|
|
|
| Attributes:
|
| penalty: The negative reward value to apply.
|
| max_gripper_pos: The maximum position value for the gripper, used for normalization.
|
| """
|
|
|
| penalty: float = -0.01
|
| max_gripper_pos: float = 30.0
|
|
|
| def complementary_data(self, complementary_data: dict) -> dict:
|
| """
|
| Calculates the gripper penalty and adds it to the complementary data.
|
|
|
| Args:
|
| complementary_data: The incoming complementary data, which should contain
|
| raw joint positions.
|
|
|
| Returns:
|
| A new complementary data dictionary with the `discrete_penalty` key added.
|
| """
|
| action = self.transition.get(TransitionKey.ACTION)
|
|
|
| raw_joint_positions = complementary_data.get("raw_joint_positions")
|
| if raw_joint_positions is None:
|
| return complementary_data
|
|
|
| current_gripper_pos = raw_joint_positions.get(GRIPPER_KEY, None)
|
| if current_gripper_pos is None:
|
| return complementary_data
|
|
|
|
|
| gripper_action = action[-1].item()
|
| gripper_action_normalized = gripper_action / self.max_gripper_pos
|
|
|
|
|
| gripper_state_normalized = current_gripper_pos / self.max_gripper_pos
|
|
|
|
|
| gripper_penalty_bool = (gripper_state_normalized < 0.5 and gripper_action_normalized > 0.5) or (
|
| gripper_state_normalized > 0.75 and gripper_action_normalized < 0.5
|
| )
|
|
|
| gripper_penalty = self.penalty * int(gripper_penalty_bool)
|
|
|
|
|
| new_complementary_data = dict(complementary_data)
|
| new_complementary_data[DISCRETE_PENALTY_KEY] = gripper_penalty
|
|
|
| return new_complementary_data
|
|
|
| def get_config(self) -> dict[str, Any]:
|
| """
|
| Returns the configuration of the step for serialization.
|
|
|
| Returns:
|
| A dictionary containing the penalty value and max gripper position.
|
| """
|
| return {
|
| "penalty": self.penalty,
|
| "max_gripper_pos": self.max_gripper_pos,
|
| }
|
|
|
| def reset(self) -> None:
|
| """Resets the processor's internal state."""
|
| pass
|
|
|
| def transform_features(
|
| self, features: dict[PipelineFeatureType, dict[str, PolicyFeature]]
|
| ) -> dict[PipelineFeatureType, dict[str, PolicyFeature]]:
|
| return features
|
|
|
|
|
| @dataclass
|
| @ProcessorStepRegistry.register("intervention_action_processor")
|
| class InterventionActionProcessorStep(ProcessorStep):
|
| """
|
| Handles human intervention, overriding policy actions and managing episode termination.
|
|
|
| When an intervention is detected (via teleoperator events in the `info` dict),
|
| this step replaces the policy's action with the human's teleoperated action.
|
| It also processes signals to terminate the episode or flag success.
|
|
|
| Attributes:
|
| use_gripper: Whether to include the gripper in the teleoperated action.
|
| terminate_on_success: If True, automatically sets the `done` flag when a
|
| `success` event is received.
|
| """
|
|
|
| use_gripper: bool = False
|
| terminate_on_success: bool = True
|
|
|
| def __call__(self, transition: EnvTransition) -> EnvTransition:
|
| """
|
| Processes the transition to handle interventions.
|
|
|
| Args:
|
| transition: The incoming environment transition.
|
|
|
| Returns:
|
| The modified transition, potentially with an overridden action, updated
|
| reward, and termination status.
|
| """
|
| action = transition.get(TransitionKey.ACTION)
|
| if not isinstance(action, PolicyAction):
|
| raise ValueError(f"Action should be a PolicyAction type got {type(action)}")
|
|
|
|
|
| info = transition.get(TransitionKey.INFO, {})
|
| complementary_data = transition.get(TransitionKey.COMPLEMENTARY_DATA, {})
|
| teleop_action = complementary_data.get(TELEOP_ACTION_KEY, {})
|
| is_intervention = info.get(TeleopEvents.IS_INTERVENTION, False)
|
| terminate_episode = info.get(TeleopEvents.TERMINATE_EPISODE, False)
|
| success = info.get(TeleopEvents.SUCCESS, False)
|
| rerecord_episode = info.get(TeleopEvents.RERECORD_EPISODE, False)
|
|
|
| new_transition = transition.copy()
|
|
|
|
|
| if is_intervention and teleop_action is not None:
|
| if isinstance(teleop_action, dict):
|
|
|
| action_list = [
|
| teleop_action.get("delta_x", 0.0),
|
| teleop_action.get("delta_y", 0.0),
|
| teleop_action.get("delta_z", 0.0),
|
| ]
|
| if self.use_gripper:
|
| action_list.append(teleop_action.get(GRIPPER_KEY, 1.0))
|
| elif isinstance(teleop_action, np.ndarray):
|
| action_list = teleop_action.tolist()
|
| else:
|
| action_list = teleop_action
|
|
|
| teleop_action_tensor = torch.tensor(action_list, dtype=action.dtype, device=action.device)
|
| new_transition[TransitionKey.ACTION] = teleop_action_tensor
|
|
|
|
|
| new_transition[TransitionKey.DONE] = bool(terminate_episode) or (
|
| self.terminate_on_success and success
|
| )
|
| new_transition[TransitionKey.REWARD] = float(success)
|
|
|
|
|
| info = new_transition.get(TransitionKey.INFO, {})
|
| info[TeleopEvents.IS_INTERVENTION] = is_intervention
|
| info[TeleopEvents.RERECORD_EPISODE] = rerecord_episode
|
| info[TeleopEvents.SUCCESS] = success
|
| new_transition[TransitionKey.INFO] = info
|
|
|
|
|
| complementary_data = new_transition.get(TransitionKey.COMPLEMENTARY_DATA, {})
|
| complementary_data[TELEOP_ACTION_KEY] = new_transition.get(TransitionKey.ACTION)
|
| new_transition[TransitionKey.COMPLEMENTARY_DATA] = complementary_data
|
|
|
| return new_transition
|
|
|
| def get_config(self) -> dict[str, Any]:
|
| """
|
| Returns the configuration of the step for serialization.
|
|
|
| Returns:
|
| A dictionary containing the step's configuration attributes.
|
| """
|
| return {
|
| "use_gripper": self.use_gripper,
|
| "terminate_on_success": self.terminate_on_success,
|
| }
|
|
|
| def transform_features(
|
| self, features: dict[PipelineFeatureType, dict[str, PolicyFeature]]
|
| ) -> dict[PipelineFeatureType, dict[str, PolicyFeature]]:
|
| return features
|
|
|
|
|
| @dataclass
|
| @ProcessorStepRegistry.register("reward_classifier_processor")
|
| class RewardClassifierProcessorStep(ProcessorStep):
|
| """
|
| Applies a pretrained reward classifier to image observations to predict success.
|
|
|
| This step uses a model to determine if the current state is successful, updating
|
| the reward and potentially terminating the episode.
|
|
|
| Attributes:
|
| pretrained_path: Path to the pretrained reward classifier model.
|
| device: The device to run the classifier on.
|
| success_threshold: The probability threshold to consider a prediction as successful.
|
| success_reward: The reward value to assign on success.
|
| terminate_on_success: If True, terminates the episode upon successful classification.
|
| reward_classifier: The loaded classifier model instance.
|
| """
|
|
|
| pretrained_path: str | None = None
|
| device: str = "cpu"
|
| success_threshold: float = 0.5
|
| success_reward: float = 1.0
|
| terminate_on_success: bool = True
|
|
|
| reward_classifier: Any = None
|
|
|
| def __post_init__(self):
|
| """Initializes the reward classifier model after the dataclass is created."""
|
| if self.pretrained_path is not None:
|
| from lerobot.policies.sac.reward_model.modeling_classifier import Classifier
|
|
|
| self.reward_classifier = Classifier.from_pretrained(self.pretrained_path)
|
| self.reward_classifier.to(self.device)
|
| self.reward_classifier.eval()
|
|
|
| def __call__(self, transition: EnvTransition) -> EnvTransition:
|
| """
|
| Processes a transition, applying the reward classifier to its image observations.
|
|
|
| Args:
|
| transition: The incoming environment transition.
|
|
|
| Returns:
|
| The modified transition with an updated reward and done flag based on the
|
| classifier's prediction.
|
| """
|
| new_transition = transition.copy()
|
| observation = new_transition.get(TransitionKey.OBSERVATION)
|
| if observation is None or self.reward_classifier is None:
|
| return new_transition
|
|
|
|
|
| images = {key: value for key, value in observation.items() if "image" in key}
|
|
|
| if not images:
|
| return new_transition
|
|
|
|
|
| start_time = time.perf_counter()
|
| with torch.inference_mode():
|
| success = self.reward_classifier.predict_reward(images, threshold=self.success_threshold)
|
|
|
| classifier_frequency = 1 / (time.perf_counter() - start_time)
|
|
|
|
|
| reward = new_transition.get(TransitionKey.REWARD, 0.0)
|
| terminated = new_transition.get(TransitionKey.DONE, False)
|
|
|
| if math.isclose(success, 1, abs_tol=1e-2):
|
| reward = self.success_reward
|
| if self.terminate_on_success:
|
| terminated = True
|
|
|
|
|
| new_transition[TransitionKey.REWARD] = reward
|
| new_transition[TransitionKey.DONE] = terminated
|
|
|
|
|
| info = new_transition.get(TransitionKey.INFO, {})
|
| info["reward_classifier_frequency"] = classifier_frequency
|
| new_transition[TransitionKey.INFO] = info
|
|
|
| return new_transition
|
|
|
| def get_config(self) -> dict[str, Any]:
|
| """
|
| Returns the configuration of the step for serialization.
|
|
|
| Returns:
|
| A dictionary containing the step's configuration attributes.
|
| """
|
| return {
|
| "device": self.device,
|
| "success_threshold": self.success_threshold,
|
| "success_reward": self.success_reward,
|
| "terminate_on_success": self.terminate_on_success,
|
| }
|
|
|
| def transform_features(
|
| self, features: dict[PipelineFeatureType, dict[str, PolicyFeature]]
|
| ) -> dict[PipelineFeatureType, dict[str, PolicyFeature]]:
|
| return features
|
|
|