| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | """Placeholder docstring""" |
| | from __future__ import absolute_import |
| |
|
| | import io |
| | import logging |
| | import struct |
| | import sys |
| |
|
| | import numpy as np |
| |
|
| | from sagemaker.amazon.record_pb2 import Record |
| | from sagemaker.deprecations import deprecated_class |
| | from sagemaker.deserializers import SimpleBaseDeserializer |
| | from sagemaker.serializers import SimpleBaseSerializer |
| | from sagemaker.utils import DeferredError |
| |
|
| |
|
| | class RecordSerializer(SimpleBaseSerializer): |
| | """Serialize a NumPy array for an inference request.""" |
| |
|
| | def __init__(self, content_type="application/x-recordio-protobuf"): |
| | """Initialize a ``RecordSerializer`` instance. |
| | |
| | Args: |
| | content_type (str): The MIME type to signal to the inference endpoint when sending |
| | request data (default: "application/x-recordio-protobuf"). |
| | """ |
| | super(RecordSerializer, self).__init__(content_type=content_type) |
| |
|
| | def serialize(self, data): |
| | """Serialize a NumPy array into a buffer containing RecordIO records. |
| | |
| | Args: |
| | data (numpy.ndarray): The data to serialize. |
| | |
| | Returns: |
| | io.BytesIO: A buffer containing the data serialized as records. |
| | """ |
| | if len(data.shape) == 1: |
| | data = data.reshape(1, data.shape[0]) |
| |
|
| | if len(data.shape) != 2: |
| | raise ValueError( |
| | "Expected a 1D or 2D array, but got a %dD array instead." % len(data.shape) |
| | ) |
| |
|
| | buffer = io.BytesIO() |
| | write_numpy_to_dense_tensor(buffer, data) |
| | buffer.seek(0) |
| |
|
| | return buffer |
| |
|
| |
|
| | class RecordDeserializer(SimpleBaseDeserializer): |
| | """Deserialize RecordIO Protobuf data from an inference endpoint.""" |
| |
|
| | def __init__(self, accept="application/x-recordio-protobuf"): |
| | """Initialize a ``RecordDeserializer`` instance. |
| | |
| | Args: |
| | accept (union[str, tuple[str]]): The MIME type (or tuple of allowable MIME types) that |
| | is expected from the inference endpoint (default: |
| | "application/x-recordio-protobuf"). |
| | """ |
| | super(RecordDeserializer, self).__init__(accept=accept) |
| |
|
| | def deserialize(self, data, content_type): |
| | """Deserialize RecordIO Protobuf data from an inference endpoint. |
| | |
| | Args: |
| | data (object): The protobuf message to deserialize. |
| | content_type (str): The MIME type of the data. |
| | Returns: |
| | list: A list of records. |
| | """ |
| | try: |
| | return read_records(data) |
| | finally: |
| | data.close() |
| |
|
| |
|
| | def _write_feature_tensor(resolved_type, record, vector): |
| | """Placeholder Docstring""" |
| | if resolved_type == "Int32": |
| | record.features["values"].int32_tensor.values.extend(vector) |
| | elif resolved_type == "Float64": |
| | record.features["values"].float64_tensor.values.extend(vector) |
| | elif resolved_type == "Float32": |
| | record.features["values"].float32_tensor.values.extend(vector) |
| |
|
| |
|
| | def _write_label_tensor(resolved_type, record, scalar): |
| | """Placeholder Docstring""" |
| | if resolved_type == "Int32": |
| | record.label["values"].int32_tensor.values.extend([scalar]) |
| | elif resolved_type == "Float64": |
| | record.label["values"].float64_tensor.values.extend([scalar]) |
| | elif resolved_type == "Float32": |
| | record.label["values"].float32_tensor.values.extend([scalar]) |
| |
|
| |
|
| | def _write_keys_tensor(resolved_type, record, vector): |
| | """Placeholder Docstring""" |
| | if resolved_type == "Int32": |
| | record.features["values"].int32_tensor.keys.extend(vector) |
| | elif resolved_type == "Float64": |
| | record.features["values"].float64_tensor.keys.extend(vector) |
| | elif resolved_type == "Float32": |
| | record.features["values"].float32_tensor.keys.extend(vector) |
| |
|
| |
|
| | def _write_shape(resolved_type, record, scalar): |
| | """Placeholder Docstring""" |
| | if resolved_type == "Int32": |
| | record.features["values"].int32_tensor.shape.extend([scalar]) |
| | elif resolved_type == "Float64": |
| | record.features["values"].float64_tensor.shape.extend([scalar]) |
| | elif resolved_type == "Float32": |
| | record.features["values"].float32_tensor.shape.extend([scalar]) |
| |
|
| |
|
| | def write_numpy_to_dense_tensor(file, array, labels=None): |
| | """Writes a numpy array to a dense tensor |
| | |
| | Args: |
| | file: |
| | array: |
| | labels: |
| | """ |
| |
|
| | |
| | if not len(array.shape) == 2: |
| | raise ValueError("Array must be a Matrix") |
| | if labels is not None: |
| | if not len(labels.shape) == 1: |
| | raise ValueError("Labels must be a Vector") |
| | if labels.shape[0] not in array.shape: |
| | raise ValueError( |
| | "Label shape {} not compatible with array shape {}".format( |
| | labels.shape, array.shape |
| | ) |
| | ) |
| | resolved_label_type = _resolve_type(labels.dtype) |
| | resolved_type = _resolve_type(array.dtype) |
| |
|
| | |
| | record = Record() |
| | for index, vector in enumerate(array): |
| | record.Clear() |
| | _write_feature_tensor(resolved_type, record, vector) |
| | if labels is not None: |
| | _write_label_tensor(resolved_label_type, record, labels[index]) |
| | _write_recordio(file, record.SerializeToString()) |
| |
|
| |
|
| | def write_spmatrix_to_sparse_tensor(file, array, labels=None): |
| | """Writes a scipy sparse matrix to a sparse tensor |
| | |
| | Args: |
| | file: |
| | array: |
| | labels: |
| | """ |
| | try: |
| | import scipy |
| | except ImportError as e: |
| | logging.warning( |
| | "scipy failed to import. Sparse matrix functions will be impaired or broken." |
| | ) |
| | |
| | scipy = DeferredError(e) |
| |
|
| | if not scipy.sparse.issparse(array): |
| | raise TypeError("Array must be sparse") |
| |
|
| | |
| | if not len(array.shape) == 2: |
| | raise ValueError("Array must be a Matrix") |
| | if labels is not None: |
| | if not len(labels.shape) == 1: |
| | raise ValueError("Labels must be a Vector") |
| | if labels.shape[0] not in array.shape: |
| | raise ValueError( |
| | "Label shape {} not compatible with array shape {}".format( |
| | labels.shape, array.shape |
| | ) |
| | ) |
| | resolved_label_type = _resolve_type(labels.dtype) |
| | resolved_type = _resolve_type(array.dtype) |
| |
|
| | csr_array = array.tocsr() |
| | n_rows, n_cols = csr_array.shape |
| |
|
| | record = Record() |
| | for row_idx in range(n_rows): |
| | record.Clear() |
| | row = csr_array.getrow(row_idx) |
| | |
| | _write_feature_tensor(resolved_type, record, row.data) |
| | |
| | _write_keys_tensor(resolved_type, record, row.indices.astype(np.uint64)) |
| |
|
| | |
| | if labels is not None: |
| | _write_label_tensor(resolved_label_type, record, labels[row_idx]) |
| |
|
| | |
| | _write_shape(resolved_type, record, n_cols) |
| |
|
| | _write_recordio(file, record.SerializeToString()) |
| |
|
| |
|
| | def read_records(file): |
| | """Eagerly read a collection of amazon Record protobuf objects from file. |
| | |
| | Args: |
| | file: |
| | """ |
| | records = [] |
| | for record_data in read_recordio(file): |
| | record = Record() |
| | record.ParseFromString(record_data) |
| | records.append(record) |
| | return records |
| |
|
| |
|
| | |
| | |
| | |
| | padding = {} |
| | for amount in range(4): |
| | if sys.version_info >= (3,): |
| | padding[amount] = bytes([0x00 for _ in range(amount)]) |
| | else: |
| | padding[amount] = bytearray([0x00 for _ in range(amount)]) |
| |
|
| | _kmagic = 0xCED7230A |
| |
|
| |
|
| | def _write_recordio(f, data): |
| | """Writes a single data point as a RecordIO record to the given file. |
| | |
| | Args: |
| | f: |
| | data: |
| | """ |
| | length = len(data) |
| | f.write(struct.pack("I", _kmagic)) |
| | f.write(struct.pack("I", length)) |
| | pad = (((length + 3) >> 2) << 2) - length |
| | f.write(data) |
| | f.write(padding[pad]) |
| |
|
| |
|
| | def read_recordio(f): |
| | """Placeholder Docstring""" |
| | while True: |
| | try: |
| | (read_kmagic,) = struct.unpack("I", f.read(4)) |
| | except struct.error: |
| | return |
| | assert read_kmagic == _kmagic |
| | (len_record,) = struct.unpack("I", f.read(4)) |
| | pad = (((len_record + 3) >> 2) << 2) - len_record |
| | yield f.read(len_record) |
| | if pad: |
| | f.read(pad) |
| |
|
| |
|
| | def _resolve_type(dtype): |
| | """Placeholder Docstring""" |
| | if dtype == np.dtype(int): |
| | return "Int32" |
| | if dtype == np.dtype(float): |
| | return "Float64" |
| | if dtype == np.dtype("float32"): |
| | return "Float32" |
| | raise ValueError("Unsupported dtype {} on array".format(dtype)) |
| |
|
| |
|
| | numpy_to_record_serializer = deprecated_class(RecordSerializer, "numpy_to_record_serializer") |
| | record_deserializer = deprecated_class(RecordDeserializer, "record_deserializer") |
| |
|