| | import abc |
| | import time |
| | from typing import ( |
| | TYPE_CHECKING, |
| | Any, |
| | Callable, |
| | Dict, |
| | Iterable, |
| | Iterator, |
| | List, |
| | Optional, |
| | Tuple, |
| | TypeVar, |
| | Union, |
| | ) |
| |
|
| | import numpy as np |
| |
|
| | from ray.data._internal.block_batching.iter_batches import iter_batches |
| | from ray.data._internal.execution.interfaces import RefBundle |
| | from ray.data._internal.logical.operators.input_data_operator import InputData |
| | from ray.data._internal.logical.optimizers import LogicalPlan |
| | from ray.data._internal.plan import ExecutionPlan |
| | from ray.data._internal.stats import DatasetStats, StatsManager |
| | from ray.data.block import BlockAccessor, DataBatch, _apply_batch_format |
| | from ray.util.annotations import PublicAPI |
| |
|
| | if TYPE_CHECKING: |
| | import tensorflow as tf |
| | import torch |
| |
|
| | from ray.data.dataset import ( |
| | CollatedData, |
| | MaterializedDataset, |
| | Schema, |
| | TensorFlowTensorBatchType, |
| | TorchBatchType, |
| | ) |
| |
|
| |
|
| | T = TypeVar("T") |
| |
|
| |
|
| | class _IterableFromIterator(Iterable[T]): |
| | def __init__(self, iterator_gen: Callable[[], Iterator[T]]): |
| | """Constructs an Iterable from an iterator generator. |
| | |
| | Args: |
| | iterator_gen: A function that returns an iterator each time it |
| | is called. For example, this can be a generator function. |
| | """ |
| | self.iterator_gen = iterator_gen |
| |
|
| | def __iter__(self): |
| | return self.iterator_gen() |
| |
|
| |
|
| | @PublicAPI(stability="beta") |
| | class DataIterator(abc.ABC): |
| | """An iterator for reading records from a :class:`~Dataset`. |
| | |
| | For Datasets, each iteration call represents a complete read of all items in the |
| | Dataset. |
| | |
| | If using Ray Train, each trainer actor should get its own iterator by calling |
| | :meth:`ray.train.get_dataset_shard("train") |
| | <ray.train.get_dataset_shard>`. |
| | |
| | Examples: |
| | >>> import ray |
| | >>> ds = ray.data.range(5) |
| | >>> ds |
| | Dataset(num_rows=5, schema={id: int64}) |
| | >>> ds.iterator() |
| | DataIterator(Dataset(num_rows=5, schema={id: int64})) |
| | """ |
| |
|
| | @abc.abstractmethod |
| | def _to_ref_bundle_iterator( |
| | self, |
| | ) -> Tuple[Iterator[RefBundle], Optional[DatasetStats], bool]: |
| | """Returns the iterator to use for `iter_batches`. |
| | |
| | Returns: |
| | A tuple. The first item of the tuple is an iterator over RefBundles. |
| | The second item of the tuple is a DatasetStats object used for recording |
| | stats during iteration. |
| | The third item is a boolean indicating if the blocks can be safely cleared |
| | after use. |
| | """ |
| | raise NotImplementedError |
| |
|
| | @PublicAPI(stability="beta") |
| | def iter_batches( |
| | self, |
| | *, |
| | prefetch_batches: int = 1, |
| | batch_size: int = 256, |
| | batch_format: Optional[str] = "default", |
| | drop_last: bool = False, |
| | local_shuffle_buffer_size: Optional[int] = None, |
| | local_shuffle_seed: Optional[int] = None, |
| | _collate_fn: Optional[Callable[[DataBatch], "CollatedData"]] = None, |
| | _finalize_fn: Optional[Callable[[Any], Any]] = None, |
| | ) -> Iterable[DataBatch]: |
| | """Return a batched iterable over the dataset. |
| | |
| | Examples: |
| | >>> import ray |
| | >>> for batch in ray.data.range( |
| | ... 1000000 |
| | ... ).iterator().iter_batches(): # doctest: +SKIP |
| | ... print(batch) # doctest: +SKIP |
| | |
| | Time complexity: O(1) |
| | |
| | Args: |
| | prefetch_batches: The number of batches to fetch ahead of the current batch |
| | to fetch. If set to greater than 0, a separate threadpool will be used |
| | to fetch the objects to the local node, format the batches, and apply |
| | the collate_fn. Defaults to 1. |
| | batch_size: The number of rows in each batch, or None to use entire blocks |
| | as batches (blocks may contain different number of rows). |
| | The final batch may include fewer than ``batch_size`` rows if |
| | ``drop_last`` is ``False``. Defaults to 256. |
| | batch_format: Specify ``"default"`` to use the default block format |
| | (NumPy), ``"pandas"`` to select ``pandas.DataFrame``, "pyarrow" to |
| | select ``pyarrow.Table``, or ``"numpy"`` to select |
| | ``Dict[str, numpy.ndarray]``, or None to return the underlying block |
| | exactly as is with no additional formatting. |
| | drop_last: Whether to drop the last batch if it's incomplete. |
| | local_shuffle_buffer_size: If non-None, the data will be randomly shuffled |
| | using a local in-memory shuffle buffer, and this value will serve as the |
| | minimum number of rows that must be in the local in-memory shuffle |
| | buffer in order to yield a batch. When there are no more rows to add to |
| | the buffer, the remaining rows in the buffer will be drained. |
| | local_shuffle_seed: The seed to use for the local random shuffle. |
| | |
| | Returns: |
| | An iterable over record batches. |
| | """ |
| | batch_format = _apply_batch_format(batch_format) |
| |
|
| | def _create_iterator() -> Iterator[DataBatch]: |
| | time_start = time.perf_counter() |
| | |
| | |
| | |
| | |
| | ( |
| | ref_bundles_iterator, |
| | stats, |
| | blocks_owned_by_consumer, |
| | ) = self._to_ref_bundle_iterator() |
| |
|
| | iterator = iter( |
| | iter_batches( |
| | ref_bundles_iterator, |
| | stats=stats, |
| | clear_block_after_read=blocks_owned_by_consumer, |
| | batch_size=batch_size, |
| | batch_format=batch_format, |
| | drop_last=drop_last, |
| | collate_fn=_collate_fn, |
| | finalize_fn=_finalize_fn, |
| | shuffle_buffer_min_size=local_shuffle_buffer_size, |
| | shuffle_seed=local_shuffle_seed, |
| | prefetch_batches=prefetch_batches, |
| | ) |
| | ) |
| |
|
| | dataset_tag = self._get_dataset_tag() |
| |
|
| | if stats: |
| | stats.iter_initialize_s.add(time.perf_counter() - time_start) |
| |
|
| | for batch in iterator: |
| | yield batch |
| | StatsManager.update_iteration_metrics(stats, dataset_tag) |
| | StatsManager.clear_iteration_metrics(dataset_tag) |
| |
|
| | if stats: |
| | stats.iter_total_s.add(time.perf_counter() - time_start) |
| |
|
| | return _IterableFromIterator(_create_iterator) |
| |
|
| | def _get_dataset_tag(self) -> str: |
| | return "unknown_dataset" |
| |
|
| | def iter_rows(self) -> Iterable[Dict[str, Any]]: |
| | """Return a local row iterable over the dataset. |
| | |
| | If the dataset is a tabular dataset (Arrow/Pandas blocks), dicts |
| | are yielded for each row by the iterator. If the dataset is not tabular, |
| | the raw row is yielded. |
| | |
| | Examples: |
| | >>> import ray |
| | >>> dataset = ray.data.range(10) |
| | >>> next(iter(dataset.iterator().iter_rows())) |
| | {'id': 0} |
| | |
| | Time complexity: O(1) |
| | |
| | Returns: |
| | An iterable over rows of the dataset. |
| | """ |
| | batch_iterable = self.iter_batches( |
| | batch_size=None, batch_format=None, prefetch_batches=1 |
| | ) |
| |
|
| | def _wrapped_iterator(): |
| | for batch in batch_iterable: |
| | batch = BlockAccessor.for_block(BlockAccessor.batch_to_block(batch)) |
| | for row in batch.iter_rows(public_row_format=True): |
| | yield row |
| |
|
| | return _IterableFromIterator(_wrapped_iterator) |
| |
|
| | @abc.abstractmethod |
| | @PublicAPI(stability="beta") |
| | def stats(self) -> str: |
| | """Returns a string containing execution timing information.""" |
| | raise NotImplementedError |
| |
|
| | @abc.abstractmethod |
| | def schema(self) -> "Schema": |
| | """Return the schema of the dataset iterated over.""" |
| | raise NotImplementedError |
| |
|
| | @PublicAPI(stability="beta") |
| | def iter_torch_batches( |
| | self, |
| | *, |
| | prefetch_batches: int = 1, |
| | batch_size: Optional[int] = 256, |
| | dtypes: Optional[Union["torch.dtype", Dict[str, "torch.dtype"]]] = None, |
| | device: str = "auto", |
| | collate_fn: Optional[Callable[[Dict[str, np.ndarray]], "CollatedData"]] = None, |
| | drop_last: bool = False, |
| | local_shuffle_buffer_size: Optional[int] = None, |
| | local_shuffle_seed: Optional[int] = None, |
| | ) -> Iterable["TorchBatchType"]: |
| | """Return a batched iterable of Torch Tensors over the dataset. |
| | |
| | This iterable yields a dictionary of column-tensors. If you are looking for |
| | more flexibility in the tensor conversion (e.g. casting dtypes) or the batch |
| | format, try using :meth:`~ray.data.DataIterator.iter_batches` directly. |
| | |
| | Examples: |
| | >>> import ray |
| | >>> for batch in ray.data.range( |
| | ... 12, |
| | ... ).iterator().iter_torch_batches(batch_size=4): |
| | ... print(batch) |
| | {'id': tensor([0, 1, 2, 3])} |
| | {'id': tensor([4, 5, 6, 7])} |
| | {'id': tensor([ 8, 9, 10, 11])} |
| | |
| | Use the ``collate_fn`` to customize how the tensor batch is created. |
| | |
| | >>> from typing import Any, Dict |
| | >>> import torch |
| | >>> import numpy as np |
| | >>> import ray |
| | >>> def collate_fn(batch: Dict[str, np.ndarray]) -> Any: |
| | ... return torch.stack( |
| | ... [torch.as_tensor(array) for array in batch.values()], |
| | ... axis=1 |
| | ... ) |
| | >>> iterator = ray.data.from_items([ |
| | ... {"col_1": 1, "col_2": 2}, |
| | ... {"col_1": 3, "col_2": 4}]).iterator() |
| | >>> for batch in iterator.iter_torch_batches(collate_fn=collate_fn): |
| | ... print(batch) |
| | tensor([[1, 2], |
| | [3, 4]]) |
| | |
| | Time complexity: O(1) |
| | |
| | Args: |
| | prefetch_batches: The number of batches to fetch ahead of the current batch |
| | to fetch. If set to greater than 0, a separate threadpool will be used |
| | to fetch the objects to the local node, format the batches, and apply |
| | the collate_fn. Defaults to 1. |
| | batch_size: The number of rows in each batch, or None to use entire blocks |
| | as batches (blocks may contain different number of rows). |
| | The final batch may include fewer than ``batch_size`` rows if |
| | ``drop_last`` is ``False``. Defaults to 256. |
| | dtypes: The Torch dtype(s) for the created tensor(s); if None, the dtype |
| | will be inferred from the tensor data. You can't use this parameter |
| | with ``collate_fn``. |
| | device: The device on which the tensor should be placed. Defaults to |
| | "auto" which moves the tensors to the appropriate device when the |
| | Dataset is passed to Ray Train and ``collate_fn`` is not provided. |
| | Otherwise, defaults to CPU. You can't use this parameter with |
| | ``collate_fn``. |
| | collate_fn: A function to convert a Numpy batch to a PyTorch tensor batch. |
| | When this parameter is specified, the user should manually handle the |
| | host to device data transfer outside of ``collate_fn``. |
| | This is useful for further processing the data after it has been |
| | batched. Potential use cases include collating along a dimension other |
| | than the first, padding sequences of various lengths, or generally |
| | handling batches of different length tensors. If not provided, the |
| | default collate function is used which simply converts the batch of |
| | numpy arrays to a batch of PyTorch tensors. This API is still |
| | experimental and is subject to change. You can't use this parameter in |
| | conjunction with ``dtypes`` or ``device``. |
| | drop_last: Whether to drop the last batch if it's incomplete. |
| | local_shuffle_buffer_size: If non-None, the data will be randomly shuffled |
| | using a local in-memory shuffle buffer, and this value will serve as the |
| | minimum number of rows that must be in the local in-memory shuffle |
| | buffer in order to yield a batch. When there are no more rows to add to |
| | the buffer, the remaining rows in the buffer will be drained. This |
| | buffer size must be greater than or equal to ``batch_size``, and |
| | therefore ``batch_size`` must also be specified when using local |
| | shuffling. |
| | local_shuffle_seed: The seed to use for the local random shuffle. |
| | |
| | Returns: |
| | An iterable over Torch Tensor batches. |
| | """ |
| |
|
| | from ray.air._internal.torch_utils import ( |
| | convert_ndarray_batch_to_torch_tensor_batch, |
| | ) |
| | from ray.train.torch import get_device |
| |
|
| | if collate_fn is not None and (dtypes is not None or device != "auto"): |
| | raise ValueError( |
| | "collate_fn cannot be used with dtypes and device." |
| | "You should manually move the output Torch tensors to the" |
| | "desired dtype and device outside of collate_fn." |
| | ) |
| |
|
| | if device == "auto": |
| | |
| | |
| | device = get_device() |
| |
|
| | if collate_fn is None: |
| | |
| | |
| | |
| | def collate_fn(batch: Union[np.ndarray, Dict[str, np.ndarray]]): |
| | return convert_ndarray_batch_to_torch_tensor_batch( |
| | batch, |
| | dtypes=dtypes, |
| | device=None, |
| | ) |
| |
|
| | |
| | |
| | |
| | def finalize_fn(batch: Union["torch.Tensor", Dict[str, "torch.Tensor"]]): |
| | if device is not None: |
| | if isinstance(batch, dict): |
| | for k, t in batch.items(): |
| | batch[k] = t.to(device=device) |
| | else: |
| | batch = batch.to(device=device) |
| | return batch |
| |
|
| | else: |
| | finalize_fn = None |
| |
|
| | return self.iter_batches( |
| | prefetch_batches=prefetch_batches, |
| | batch_size=batch_size, |
| | drop_last=drop_last, |
| | local_shuffle_buffer_size=local_shuffle_buffer_size, |
| | local_shuffle_seed=local_shuffle_seed, |
| | _collate_fn=collate_fn, |
| | _finalize_fn=finalize_fn, |
| | ) |
| |
|
| | def iter_tf_batches( |
| | self, |
| | *, |
| | prefetch_batches: int = 1, |
| | batch_size: Optional[int] = 256, |
| | dtypes: Optional[Union["tf.dtypes.DType", Dict[str, "tf.dtypes.DType"]]] = None, |
| | drop_last: bool = False, |
| | local_shuffle_buffer_size: Optional[int] = None, |
| | local_shuffle_seed: Optional[int] = None, |
| | ) -> Iterable["TensorFlowTensorBatchType"]: |
| | """Return a batched iterable of TensorFlow Tensors over the dataset. |
| | |
| | This iterable will yield single-tensor batches of the underlying dataset |
| | consists of a single column; otherwise, it will yield a dictionary of |
| | column-tensors. |
| | |
| | .. tip:: |
| | If you don't need the additional flexibility provided by this method, |
| | consider using :meth:`~ray.data.Dataset.to_tf` instead. It's easier |
| | to use. |
| | |
| | Examples: |
| | >>> import ray |
| | >>> for batch in ray.data.range( # doctest: +SKIP |
| | ... 12, |
| | ... ).iter_tf_batches(batch_size=4): |
| | ... print(batch.shape) # doctest: +SKIP |
| | (4, 1) |
| | (4, 1) |
| | (4, 1) |
| | |
| | Time complexity: O(1) |
| | |
| | Args: |
| | prefetch_batches: The number of batches to fetch ahead of the current batch |
| | to fetch. If set to greater than 0, a separate threadpool will be used |
| | to fetch the objects to the local node, format the batches, and apply |
| | the collate_fn. Defaults to 1. |
| | batch_size: The number of rows in each batch, or None to use entire blocks |
| | as batches (blocks may contain different number of rows). |
| | The final batch may include fewer than ``batch_size`` rows if |
| | ``drop_last`` is ``False``. Defaults to 256. |
| | dtypes: The TensorFlow dtype(s) for the created tensor(s); if None, the |
| | dtype will be inferred from the tensor data. |
| | drop_last: Whether to drop the last batch if it's incomplete. |
| | local_shuffle_buffer_size: If non-None, the data will be randomly shuffled |
| | using a local in-memory shuffle buffer, and this value will serve as the |
| | minimum number of rows that must be in the local in-memory shuffle |
| | buffer in order to yield a batch. When there are no more rows to add to |
| | the buffer, the remaining rows in the buffer will be drained. This |
| | buffer size must be greater than or equal to ``batch_size``, and |
| | therefore ``batch_size`` must also be specified when using local |
| | shuffling. |
| | local_shuffle_seed: The seed to use for the local random shuffle. |
| | |
| | Returns: |
| | An iterator over TensorFlow Tensor batches. |
| | """ |
| | from ray.air._internal.tensorflow_utils import ( |
| | convert_ndarray_batch_to_tf_tensor_batch, |
| | ) |
| |
|
| | batch_iterable = self.iter_batches( |
| | prefetch_batches=prefetch_batches, |
| | batch_size=batch_size, |
| | drop_last=drop_last, |
| | local_shuffle_buffer_size=local_shuffle_buffer_size, |
| | local_shuffle_seed=local_shuffle_seed, |
| | ) |
| | mapped_iterable = map( |
| | lambda batch: convert_ndarray_batch_to_tf_tensor_batch( |
| | batch, dtypes=dtypes |
| | ), |
| | batch_iterable, |
| | ) |
| |
|
| | return mapped_iterable |
| |
|
| | def to_torch( |
| | self, |
| | *, |
| | label_column: Optional[str] = None, |
| | feature_columns: Optional[ |
| | Union[List[str], List[List[str]], Dict[str, List[str]]] |
| | ] = None, |
| | label_column_dtype: Optional["torch.dtype"] = None, |
| | feature_column_dtypes: Optional[ |
| | Union["torch.dtype", List["torch.dtype"], Dict[str, "torch.dtype"]] |
| | ] = None, |
| | batch_size: int = 1, |
| | prefetch_batches: int = 1, |
| | drop_last: bool = False, |
| | local_shuffle_buffer_size: Optional[int] = None, |
| | local_shuffle_seed: Optional[int] = None, |
| | unsqueeze_label_tensor: bool = True, |
| | unsqueeze_feature_tensors: bool = True, |
| | ) -> "torch.utils.data.IterableDataset": |
| | """Return a Torch IterableDataset over this dataset. |
| | |
| | This is only supported for datasets convertible to Arrow records. |
| | |
| | It is recommended to use the returned ``IterableDataset`` directly |
| | instead of passing it into a torch ``DataLoader``. |
| | |
| | Each element in IterableDataset will be a tuple consisting of 2 |
| | elements. The first item contains the feature tensor(s), and the |
| | second item is the label tensor. Those can take on different |
| | forms, depending on the specified arguments. |
| | |
| | For the features tensor (N is the ``batch_size`` and n, m, k |
| | are the number of features per tensor): |
| | |
| | * If ``feature_columns`` is a ``List[str]``, the features will be |
| | a tensor of shape (N, n), with columns corresponding to |
| | ``feature_columns`` |
| | |
| | * If ``feature_columns`` is a ``List[List[str]]``, the features will be |
| | a list of tensors of shape [(N, m),...,(N, k)], with columns of each |
| | tensor corresponding to the elements of ``feature_columns`` |
| | |
| | * If ``feature_columns`` is a ``Dict[str, List[str]]``, the features |
| | will be a dict of key-tensor pairs of shape |
| | {key1: (N, m),..., keyN: (N, k)}, with columns of each |
| | tensor corresponding to the value of ``feature_columns`` under the |
| | key. |
| | |
| | If ``unsqueeze_label_tensor=True`` (default), the label tensor will be |
| | of shape (N, 1). Otherwise, it will be of shape (N,). |
| | If ``label_column`` is specified as ``None``, then no column from the |
| | ``Dataset`` will be treated as the label, and the output label tensor |
| | will be ``None``. |
| | |
| | Note that you probably want to call ``.split()`` on this dataset if |
| | there are to be multiple Torch workers consuming the data. |
| | |
| | Time complexity: O(1) |
| | |
| | Args: |
| | label_column: The name of the column used as the |
| | label (second element of the output list). Can be None for |
| | prediction, in which case the second element of returned |
| | tuple will also be None. |
| | feature_columns: The names of the columns |
| | to use as the features. Can be a list of lists or |
| | a dict of string-list pairs for multi-tensor output. |
| | If None, then use all columns except the label column as |
| | the features. |
| | label_column_dtype: The torch dtype to |
| | use for the label column. If None, then automatically infer |
| | the dtype. |
| | feature_column_dtypes: The dtypes to use for the feature |
| | tensors. This should match the format of ``feature_columns``, |
| | or be a single dtype, in which case it will be applied to |
| | all tensors. If None, then automatically infer the dtype. |
| | batch_size: How many samples per batch to yield at a time. |
| | Defaults to 1. |
| | prefetch_batches: The number of batches to fetch ahead of the current batch |
| | to fetch. If set to greater than 0, a separate threadpool will be used |
| | to fetch the objects to the local node, format the batches, and apply |
| | the collate_fn. Defaults to 1. |
| | drop_last: Set to True to drop the last incomplete batch, |
| | if the dataset size is not divisible by the batch size. If |
| | False and the size of dataset is not divisible by the batch |
| | size, then the last batch will be smaller. Defaults to False. |
| | local_shuffle_buffer_size: If non-None, the data will be randomly shuffled |
| | using a local in-memory shuffle buffer, and this value will serve as the |
| | minimum number of rows that must be in the local in-memory shuffle |
| | buffer in order to yield a batch. When there are no more rows to add to |
| | the buffer, the remaining rows in the buffer will be drained. This |
| | buffer size must be greater than or equal to ``batch_size``, and |
| | therefore ``batch_size`` must also be specified when using local |
| | shuffling. |
| | local_shuffle_seed: The seed to use for the local random shuffle. |
| | unsqueeze_label_tensor: If set to True, the label tensor |
| | will be unsqueezed (reshaped to (N, 1)). Otherwise, it will |
| | be left as is, that is (N, ). In general, regression loss |
| | functions expect an unsqueezed tensor, while classification |
| | loss functions expect a squeezed one. Defaults to True. |
| | unsqueeze_feature_tensors: If set to True, the features tensors |
| | will be unsqueezed (reshaped to (N, 1)) before being concatenated into |
| | the final features tensor. Otherwise, they will be left as is, that is |
| | (N, ). Defaults to True. |
| | |
| | Returns: |
| | A torch IterableDataset. |
| | """ |
| | import torch |
| |
|
| | from ray.air._internal.torch_utils import convert_pandas_to_torch_tensor |
| | from ray.data._internal.torch_iterable_dataset import TorchIterableDataset |
| |
|
| | |
| | if not feature_columns: |
| | feature_columns = None |
| |
|
| | if feature_column_dtypes and not isinstance(feature_column_dtypes, torch.dtype): |
| | if isinstance(feature_columns, dict): |
| | if not isinstance(feature_column_dtypes, dict): |
| | raise TypeError( |
| | "If `feature_columns` is a dict, " |
| | "`feature_column_dtypes` must be None, `torch.dtype`," |
| | f" or dict, got {type(feature_column_dtypes)}." |
| | ) |
| | if set(feature_columns) != set(feature_column_dtypes): |
| | raise ValueError( |
| | "`feature_columns` and `feature_column_dtypes` " |
| | "must have the same keys." |
| | ) |
| | if any(not subcolumns for subcolumns in feature_columns.values()): |
| | raise ValueError("column list may not be empty") |
| | elif isinstance(feature_columns[0], (list, tuple)): |
| | if not isinstance(feature_column_dtypes, (list, tuple)): |
| | raise TypeError( |
| | "If `feature_columns` is a list of lists, " |
| | "`feature_column_dtypes` must be None, `torch.dtype`," |
| | f" or a sequence, got {type(feature_column_dtypes)}." |
| | ) |
| | if len(feature_columns) != len(feature_column_dtypes): |
| | raise ValueError( |
| | "`feature_columns` and `feature_column_dtypes` " |
| | "must have the same length." |
| | ) |
| | if any(not subcolumns for subcolumns in feature_columns): |
| | raise ValueError("column list may not be empty") |
| |
|
| | def make_generator(): |
| | for batch in self.iter_batches( |
| | batch_size=batch_size, |
| | batch_format="pandas", |
| | prefetch_batches=prefetch_batches, |
| | drop_last=drop_last, |
| | local_shuffle_buffer_size=local_shuffle_buffer_size, |
| | local_shuffle_seed=local_shuffle_seed, |
| | ): |
| | if label_column: |
| | label_tensor = convert_pandas_to_torch_tensor( |
| | batch, |
| | [label_column], |
| | label_column_dtype, |
| | unsqueeze=unsqueeze_label_tensor, |
| | ) |
| | batch.pop(label_column) |
| | else: |
| | label_tensor = None |
| |
|
| | if isinstance(feature_columns, dict): |
| | features_tensor = { |
| | key: convert_pandas_to_torch_tensor( |
| | batch, |
| | feature_columns[key], |
| | ( |
| | feature_column_dtypes[key] |
| | if isinstance(feature_column_dtypes, dict) |
| | else feature_column_dtypes |
| | ), |
| | unsqueeze=unsqueeze_feature_tensors, |
| | ) |
| | for key in feature_columns |
| | } |
| | else: |
| | features_tensor = convert_pandas_to_torch_tensor( |
| | batch, |
| | columns=feature_columns, |
| | column_dtypes=feature_column_dtypes, |
| | unsqueeze=unsqueeze_feature_tensors, |
| | ) |
| |
|
| | yield (features_tensor, label_tensor) |
| |
|
| | return TorchIterableDataset(make_generator) |
| |
|
| | @PublicAPI(stability="beta") |
| | def to_tf( |
| | self, |
| | feature_columns: Union[str, List[str]], |
| | label_columns: Union[str, List[str]], |
| | *, |
| | additional_columns: Union[Optional[str], Optional[List[str]]] = None, |
| | prefetch_batches: int = 1, |
| | batch_size: int = 1, |
| | drop_last: bool = False, |
| | local_shuffle_buffer_size: Optional[int] = None, |
| | local_shuffle_seed: Optional[int] = None, |
| | feature_type_spec: Union["tf.TypeSpec", Dict[str, "tf.TypeSpec"]] = None, |
| | label_type_spec: Union["tf.TypeSpec", Dict[str, "tf.TypeSpec"]] = None, |
| | additional_type_spec: Union[ |
| | Optional["tf.TypeSpec"], Optional[Dict[str, "tf.TypeSpec"]] |
| | ] = None, |
| | ) -> "tf.data.Dataset": |
| | """Return a TF Dataset over this dataset. |
| | |
| | .. warning:: |
| | If your dataset contains ragged tensors, this method errors. To prevent |
| | errors, :ref:`resize your tensors <transforming_tensors>`. |
| | |
| | Examples: |
| | >>> import ray |
| | >>> ds = ray.data.read_csv( |
| | ... "s3://anonymous@air-example-data/iris.csv" |
| | ... ) |
| | >>> it = ds.iterator(); it |
| | DataIterator(Dataset( |
| | num_rows=?, |
| | schema={ |
| | sepal length (cm): double, |
| | sepal width (cm): double, |
| | petal length (cm): double, |
| | petal width (cm): double, |
| | target: int64 |
| | } |
| | )) |
| | |
| | If your model accepts a single tensor as input, specify a single feature column. |
| | |
| | >>> it.to_tf(feature_columns="sepal length (cm)", label_columns="target") |
| | <_OptionsDataset element_spec=(TensorSpec(shape=(None,), dtype=tf.float64, name='sepal length (cm)'), TensorSpec(shape=(None,), dtype=tf.int64, name='target'))> |
| | |
| | If your model accepts a dictionary as input, specify a list of feature columns. |
| | |
| | >>> it.to_tf(["sepal length (cm)", "sepal width (cm)"], "target") |
| | <_OptionsDataset element_spec=({'sepal length (cm)': TensorSpec(shape=(None,), dtype=tf.float64, name='sepal length (cm)'), 'sepal width (cm)': TensorSpec(shape=(None,), dtype=tf.float64, name='sepal width (cm)')}, TensorSpec(shape=(None,), dtype=tf.int64, name='target'))> |
| | |
| | If your dataset contains multiple features but your model accepts a single |
| | tensor as input, combine features with |
| | :class:`~ray.data.preprocessors.Concatenator`. |
| | |
| | >>> from ray.data.preprocessors import Concatenator |
| | >>> columns_to_concat = ["sepal length (cm)", "sepal width (cm)", "petal length (cm)", "petal width (cm)"] |
| | >>> preprocessor = Concatenator(columns=columns_to_concat, output_column_name="features") |
| | >>> it = preprocessor.transform(ds).iterator() |
| | >>> it |
| | DataIterator(Concatenator |
| | +- Dataset( |
| | num_rows=?, |
| | schema={ |
| | sepal length (cm): double, |
| | sepal width (cm): double, |
| | petal length (cm): double, |
| | petal width (cm): double, |
| | target: int64 |
| | } |
| | )) |
| | >>> it.to_tf("features", "target") |
| | <_OptionsDataset element_spec=(TensorSpec(shape=(None, 4), dtype=tf.float64, name='features'), TensorSpec(shape=(None,), dtype=tf.int64, name='target'))> |
| | |
| | If your model accepts different types, shapes, or names of tensors as input, specify the type spec. |
| | If type specs are not specified, they are automatically inferred from the schema of the iterator. |
| | |
| | >>> import tensorflow as tf |
| | >>> it.to_tf( |
| | ... feature_columns="features", |
| | ... label_columns="target", |
| | ... feature_type_spec=tf.TensorSpec(shape=(None, 4), dtype=tf.float32, name="features"), |
| | ... label_type_spec=tf.TensorSpec(shape=(None,), dtype=tf.float32, name="label") |
| | ... ) |
| | <_OptionsDataset element_spec=(TensorSpec(shape=(None, 4), dtype=tf.float32, name='features'), TensorSpec(shape=(None,), dtype=tf.float32, name='label'))> |
| | |
| | If your model accepts additional metadata aside from features and label, specify a single additional column or a list of additional columns. |
| | A common use case is to include sample weights in the data samples and train a ``tf.keras.Model`` with ``tf.keras.Model.fit``. |
| | |
| | >>> import pandas as pd |
| | >>> ds = ds.add_column("sample weights", lambda df: pd.Series([1] * len(df))) |
| | >>> it = ds.iterator() |
| | >>> it.to_tf(feature_columns="sepal length (cm)", label_columns="target", additional_columns="sample weights") |
| | <_OptionsDataset element_spec=(TensorSpec(shape=(None,), dtype=tf.float64, name='sepal length (cm)'), TensorSpec(shape=(None,), dtype=tf.int64, name='target'), TensorSpec(shape=(None,), dtype=tf.int64, name='sample weights'))> |
| | |
| | If your model accepts different types, shapes, or names for the additional metadata, specify the type spec of the additional column. |
| | |
| | >>> it.to_tf( |
| | ... feature_columns="sepal length (cm)", |
| | ... label_columns="target", |
| | ... additional_columns="sample weights", |
| | ... additional_type_spec=tf.TensorSpec(shape=(None,), dtype=tf.float32, name="weight") |
| | ... ) |
| | <_OptionsDataset element_spec=(TensorSpec(shape=(None,), dtype=tf.float64, name='sepal length (cm)'), TensorSpec(shape=(None,), dtype=tf.int64, name='target'), TensorSpec(shape=(None,), dtype=tf.float32, name='weight'))> |
| | |
| | Args: |
| | feature_columns: Columns that correspond to model inputs. If this is a |
| | string, the input data is a tensor. If this is a list, the input data |
| | is a ``dict`` that maps column names to their tensor representation. |
| | label_columns: Columns that correspond to model targets. If this is a |
| | string, the target data is a tensor. If this is a list, the target data |
| | is a ``dict`` that maps column names to their tensor representation. |
| | additional_columns: Columns that correspond to sample weights or other metadata. |
| | If this is a string, the weight data is a tensor. If this is a list, the |
| | weight data is a ``dict`` that maps column names to their tensor representation. |
| | prefetch_batches: The number of batches to fetch ahead of the current batch |
| | to fetch. If set to greater than 0, a separate threadpool will be used |
| | to fetch the objects to the local node, format the batches, and apply |
| | the collate_fn. Defaults to 1. |
| | batch_size: Record batch size. Defaults to 1. |
| | drop_last: Set to True to drop the last incomplete batch, |
| | if the dataset size is not divisible by the batch size. If |
| | False and the size of dataset is not divisible by the batch |
| | size, then the last batch will be smaller. Defaults to False. |
| | local_shuffle_buffer_size: If non-None, the data will be randomly shuffled |
| | using a local in-memory shuffle buffer, and this value will serve as the |
| | minimum number of rows that must be in the local in-memory shuffle |
| | buffer in order to yield a batch. When there are no more rows to add to |
| | the buffer, the remaining rows in the buffer will be drained. This |
| | buffer size must be greater than or equal to ``batch_size``, and |
| | therefore ``batch_size`` must also be specified when using local |
| | shuffling. |
| | local_shuffle_seed: The seed to use for the local random shuffle. |
| | feature_type_spec: The `tf.TypeSpec` of `feature_columns`. If there is |
| | only one column, specify a `tf.TypeSpec`. If there are multiple columns, |
| | specify a ``dict`` that maps column names to their `tf.TypeSpec`. |
| | Default is `None` to automatically infer the type of each column. |
| | label_type_spec: The `tf.TypeSpec` of `label_columns`. If there is |
| | only one column, specify a `tf.TypeSpec`. If there are multiple columns, |
| | specify a ``dict`` that maps column names to their `tf.TypeSpec`. |
| | Default is `None` to automatically infer the type of each column. |
| | additional_type_spec: The `tf.TypeSpec` of `additional_columns`. If there |
| | is only one column, specify a `tf.TypeSpec`. If there are multiple |
| | columns, specify a ``dict`` that maps column names to their `tf.TypeSpec`. |
| | Default is `None` to automatically infer the type of each column. |
| | |
| | Returns: |
| | A ``tf.data.Dataset`` that yields inputs and targets. |
| | """ |
| |
|
| | from ray.air._internal.tensorflow_utils import ( |
| | convert_ndarray_to_tf_tensor, |
| | get_type_spec, |
| | ) |
| |
|
| | try: |
| | import tensorflow as tf |
| | except ImportError: |
| | raise ValueError("tensorflow must be installed!") |
| |
|
| | def validate_column(column: str) -> None: |
| | if column not in valid_columns: |
| | raise ValueError( |
| | f"You specified '{column}' in `feature_columns`, " |
| | f"`label_columns`, or `additional_columns`, but there's no " |
| | f"column named '{column}' in the dataset. " |
| | f"Valid column names are: {valid_columns}." |
| | ) |
| |
|
| | def validate_columns(columns: Union[str, List]) -> None: |
| | if isinstance(columns, list): |
| | for column in columns: |
| | validate_column(column) |
| | else: |
| | validate_column(columns) |
| |
|
| | def convert_batch_to_tensors( |
| | batch: Dict[str, np.ndarray], |
| | *, |
| | columns: Union[str, List[str]], |
| | type_spec: Union[tf.TypeSpec, Dict[str, tf.TypeSpec]], |
| | ) -> Union[tf.Tensor, Dict[str, tf.Tensor]]: |
| | if isinstance(columns, str): |
| | return convert_ndarray_to_tf_tensor(batch[columns], type_spec=type_spec) |
| | return { |
| | column: convert_ndarray_to_tf_tensor( |
| | batch[column], type_spec=type_spec[column] |
| | ) |
| | for column in columns |
| | } |
| |
|
| | def generator(): |
| | for batch in self.iter_batches( |
| | prefetch_batches=prefetch_batches, |
| | batch_size=batch_size, |
| | drop_last=drop_last, |
| | local_shuffle_buffer_size=local_shuffle_buffer_size, |
| | local_shuffle_seed=local_shuffle_seed, |
| | ): |
| | assert isinstance(batch, dict) |
| | features = convert_batch_to_tensors( |
| | batch, columns=feature_columns, type_spec=feature_type_spec |
| | ) |
| | labels = convert_batch_to_tensors( |
| | batch, columns=label_columns, type_spec=label_type_spec |
| | ) |
| |
|
| | if additional_columns is None: |
| | yield features, labels |
| | else: |
| | additional_metadata = convert_batch_to_tensors( |
| | batch, |
| | columns=additional_columns, |
| | type_spec=additional_type_spec, |
| | ) |
| | yield features, labels, additional_metadata |
| |
|
| | if feature_type_spec is None or label_type_spec is None: |
| | schema = self.schema() |
| | valid_columns = set(schema.names) |
| | validate_columns(feature_columns) |
| | validate_columns(label_columns) |
| | feature_type_spec = get_type_spec(schema, columns=feature_columns) |
| | label_type_spec = get_type_spec(schema, columns=label_columns) |
| |
|
| | if additional_columns is not None and additional_type_spec is None: |
| | schema = self.schema() |
| | valid_columns = set(schema.names) |
| | validate_columns(additional_columns) |
| | additional_type_spec = get_type_spec(schema, columns=additional_columns) |
| |
|
| | if additional_columns is not None: |
| | dataset = tf.data.Dataset.from_generator( |
| | generator, |
| | output_signature=( |
| | feature_type_spec, |
| | label_type_spec, |
| | additional_type_spec, |
| | ), |
| | ) |
| | else: |
| | dataset = tf.data.Dataset.from_generator( |
| | generator, output_signature=(feature_type_spec, label_type_spec) |
| | ) |
| |
|
| | options = tf.data.Options() |
| | options.experimental_distribute.auto_shard_policy = ( |
| | tf.data.experimental.AutoShardPolicy.OFF |
| | ) |
| | return dataset.with_options(options) |
| |
|
| | @PublicAPI(stability="beta") |
| | def materialize(self) -> "MaterializedDataset": |
| | """Execute and materialize this data iterator into object store memory. |
| | |
| | .. note:: |
| | This method triggers the execution and materializes all blocks |
| | of the iterator, returning its contents as a |
| | :class:`~ray.data.dataset.MaterializedDataset` for further processing. |
| | """ |
| |
|
| | from ray.data.dataset import MaterializedDataset |
| |
|
| | ref_bundles_iter, stats, _ = self._to_ref_bundle_iterator() |
| |
|
| | ref_bundles = list(ref_bundles_iter) |
| | execution_plan = ExecutionPlan(stats) |
| | logical_plan = LogicalPlan( |
| | InputData(input_data=ref_bundles), |
| | execution_plan._context, |
| | ) |
| | return MaterializedDataset( |
| | execution_plan, |
| | logical_plan, |
| | ) |
| |
|
| | def __del__(self): |
| | |
| | StatsManager.clear_iteration_metrics(self._get_dataset_tag()) |
| |
|
| |
|
| | |
| | DatasetIterator = DataIterator |
| |
|