|
from abc import abstractmethod |
|
from dataclasses import field |
|
from typing import Any, Dict, Generator, List, Optional, Union |
|
|
|
from .artifact import Artifact |
|
from .dataclass import InternalField, NonPositionalField |
|
from .stream import DynamicStream, EmptyStreamError, MultiStream, Stream |
|
from .utils import is_module_available |
|
|
|
|
|
class Operator(Artifact): |
|
pass |
|
|
|
|
|
class PackageRequirementsMixin(Artifact): |
|
"""Base class used to automatically check for the existence of required python dependencies for an artifact (e.g. Operator or Metric). |
|
|
|
The _requirement list is either a list of required packages |
|
(e.g. ["torch","sentence_transformers"]) or a dictionary between required packages |
|
and detailed installation instructions on how how to install each package. |
|
(e.g. {"torch" : "Install Torch using `pip install torch`", "sentence_transformers" : Install Sentence Transformers using `pip install sentence-transformers`}) |
|
Note that the package names should be specified as they are used in the python import statement for the package. |
|
""" |
|
|
|
_requirements_list: Union[List[str], Dict[str, str]] = InternalField( |
|
default_factory=list |
|
) |
|
|
|
def verify(self): |
|
super().verify() |
|
self.check_missing_requirements() |
|
|
|
def check_missing_requirements(self, requirements=None): |
|
if requirements is None: |
|
requirements = self._requirements_list |
|
if isinstance(requirements, List): |
|
requirements = {package: "" for package in requirements} |
|
|
|
missing_packages = [] |
|
installation_instructions = [] |
|
for package, installation_instruction in requirements.items(): |
|
if not is_module_available(package): |
|
missing_packages.append(package) |
|
installation_instructions.append(installation_instruction) |
|
if missing_packages: |
|
raise MissingRequirementsError( |
|
self.__class__.__name__, missing_packages, installation_instructions |
|
) |
|
|
|
|
|
class MissingRequirementsError(Exception): |
|
def __init__(self, class_name, missing_packages, installation_instructions): |
|
self.class_name = class_name |
|
self.missing_packages = missing_packages |
|
self.installation_instruction = installation_instructions |
|
self.message = ( |
|
f"{self.class_name} requires the following missing package(s): {', '.join(self.missing_packages)}. " |
|
+ "\n".join(self.installation_instruction) |
|
) |
|
super().__init__(self.message) |
|
|
|
|
|
class OperatorError(Exception): |
|
def __init__(self, exception: Exception, operators: List[Operator]): |
|
super().__init__( |
|
"This error was raised by the following operators: " |
|
+ ",\n".join([str(operator) for operator in operators]) |
|
+ "." |
|
) |
|
self.exception = exception |
|
self.operators = operators |
|
|
|
@classmethod |
|
def from_operator_error(cls, exception: Exception, operator: Operator): |
|
return cls(exception.exception, [*exception.operators, operator]) |
|
|
|
@classmethod |
|
def from_exception(cls, exception: Exception, operator: Operator): |
|
return cls(exception, [operator]) |
|
|
|
|
|
class StreamingOperator(Operator, PackageRequirementsMixin): |
|
"""Base class for all stream operators in the streaming model. |
|
|
|
Stream operators are a key component of the streaming model and are responsible for processing continuous data streams. |
|
They perform operations such as transformations, aggregations, joins, windowing and more on these streams. |
|
There are several types of stream operators, including source operators, processing operators, etc. |
|
|
|
As a `StreamingOperator`, this class is responsible for performing operations on a stream, and must be implemented by all other specific types of stream operators in the system. |
|
When called, a `StreamingOperator` must return a MultiStream. |
|
|
|
As a subclass of `Artifact`, every `StreamingOperator` can be saved in a catalog for further usage or reference. |
|
|
|
""" |
|
|
|
@abstractmethod |
|
def __call__(self, streams: Optional[MultiStream] = None) -> MultiStream: |
|
"""Abstract method that performs operations on the stream. |
|
|
|
Args: |
|
streams (Optional[MultiStream]): The input MultiStream, which can be None. |
|
|
|
Returns: |
|
MultiStream: The output MultiStream resulting from the operations performed on the input. |
|
""" |
|
|
|
|
|
class SideEffectOperator(StreamingOperator): |
|
"""Base class for operators that does not affect the stream.""" |
|
|
|
def __call__(self, streams: Optional[MultiStream] = None) -> MultiStream: |
|
self.process() |
|
return streams |
|
|
|
@abstractmethod |
|
def process() -> None: |
|
pass |
|
|
|
|
|
def instance_generator(instance): |
|
yield instance |
|
|
|
|
|
def stream_single(instance: Dict[str, Any]) -> Stream: |
|
return DynamicStream( |
|
generator=instance_generator, gen_kwargs={"instance": instance} |
|
) |
|
|
|
|
|
class MultiStreamOperator(StreamingOperator): |
|
"""A class representing a multi-stream operator in the streaming system. |
|
|
|
A multi-stream operator is a type of `StreamingOperator` that operates on an entire MultiStream object at once. It takes a `MultiStream` as input and produces a `MultiStream` as output. The `process` method should be implemented by subclasses to define the specific operations to be performed on the input `MultiStream`. |
|
""" |
|
|
|
caching: bool = NonPositionalField(default=None) |
|
|
|
def __call__(self, multi_stream: Optional[MultiStream] = None) -> MultiStream: |
|
self.before_process_multi_stream() |
|
result = self._process_multi_stream(multi_stream) |
|
if self.caching is not None: |
|
result.set_caching(self.caching) |
|
return result |
|
|
|
def before_process_multi_stream(self): |
|
pass |
|
|
|
def _process_multi_stream( |
|
self, multi_stream: Optional[MultiStream] = None |
|
) -> MultiStream: |
|
result = self.process(multi_stream) |
|
assert isinstance( |
|
result, MultiStream |
|
), "MultiStreamOperator must return a MultiStream" |
|
return result |
|
|
|
@abstractmethod |
|
def process(self, multi_stream: MultiStream) -> MultiStream: |
|
pass |
|
|
|
def process_instance(self, instance, stream_name="tmp"): |
|
instance = self.verify_instance(instance) |
|
multi_stream = MultiStream({stream_name: stream_single(instance)}) |
|
processed_multi_stream = self(multi_stream) |
|
return next(iter(processed_multi_stream[stream_name])) |
|
|
|
|
|
class SourceOperator(MultiStreamOperator): |
|
"""A class representing a source operator in the streaming system. |
|
|
|
A source operator is responsible for generating the data stream from some source, such as a database or a file. |
|
This is the starting point of a stream processing pipeline. |
|
The `SourceOperator` class is a type of `SourceOperator`, which is a special type of `StreamingOperator` |
|
that generates an output stream but does not take any input streams. |
|
|
|
When called, a `SourceOperator` invokes its `process` method, which should be implemented by all subclasses |
|
to generate the required `MultiStream`. |
|
|
|
""" |
|
|
|
def _process_multi_stream( |
|
self, multi_stream: Optional[MultiStream] = None |
|
) -> MultiStream: |
|
result = self.process() |
|
assert isinstance( |
|
result, MultiStream |
|
), "MultiStreamOperator must return a MultiStream" |
|
return result |
|
|
|
@abstractmethod |
|
def process(self) -> MultiStream: |
|
pass |
|
|
|
|
|
class StreamInitializerOperator(SourceOperator): |
|
"""A class representing a stream initializer operator in the streaming system. |
|
|
|
A stream initializer operator is a special type of `SourceOperator` that is capable of taking parameters during the stream generation process. This can be useful in situations where the stream generation process needs to be customized or configured based on certain parameters. |
|
|
|
When called, a `StreamInitializerOperator` invokes its `process` method, passing any supplied arguments and keyword arguments. The `process` method should be implemented by all subclasses to generate the required `MultiStream` based on the given arguments and keyword arguments. |
|
|
|
""" |
|
|
|
caching: bool = NonPositionalField(default=None) |
|
|
|
def __call__(self, *args, **kwargs) -> MultiStream: |
|
multi_stream = self.process(*args, **kwargs) |
|
if self.caching is not None: |
|
multi_stream.set_caching(self.caching) |
|
return self.process(*args, **kwargs) |
|
|
|
@abstractmethod |
|
def process(self, *args, **kwargs) -> MultiStream: |
|
pass |
|
|
|
|
|
class StreamOperator(MultiStreamOperator): |
|
"""A class representing a single-stream operator in the streaming system. |
|
|
|
A single-stream operator is a type of `MultiStreamOperator` that operates on individual |
|
`Stream` objects within a `MultiStream`. It iterates through each `Stream` in the `MultiStream` |
|
and applies the `process` method. |
|
The `process` method should be implemented by subclasses to define the specific operations |
|
to be performed on each `Stream`. |
|
|
|
""" |
|
|
|
apply_to_streams: List[str] = NonPositionalField( |
|
default=None |
|
) |
|
dont_apply_to_streams: List[str] = NonPositionalField(default=None) |
|
|
|
def _process_multi_stream(self, multi_stream: MultiStream) -> MultiStream: |
|
result = {} |
|
for stream_name, stream in multi_stream.items(): |
|
if self._is_should_be_processed(stream_name): |
|
stream = self._process_single_stream(stream, stream_name) |
|
else: |
|
stream = stream |
|
assert isinstance(stream, Stream), "StreamOperator must return a Stream" |
|
result[stream_name] = stream |
|
|
|
return MultiStream(result) |
|
|
|
def _process_single_stream( |
|
self, stream: Stream, stream_name: Optional[str] = None |
|
) -> Stream: |
|
return DynamicStream( |
|
self._process_stream, |
|
gen_kwargs={"stream": stream, "stream_name": stream_name}, |
|
) |
|
|
|
def _is_should_be_processed(self, stream_name): |
|
if ( |
|
self.apply_to_streams is not None |
|
and self.dont_apply_to_streams is not None |
|
and stream_name in self.apply_to_streams |
|
and stream_name in self.dont_apply_to_streams |
|
): |
|
raise ValueError( |
|
f"Stream '{stream_name}' can be in either apply_to_streams or dont_apply_to_streams not both." |
|
) |
|
|
|
return ( |
|
self.apply_to_streams is None or stream_name in self.apply_to_streams |
|
) and ( |
|
self.dont_apply_to_streams is None |
|
or stream_name not in self.dont_apply_to_streams |
|
) |
|
|
|
def _process_stream( |
|
self, stream: Stream, stream_name: Optional[str] = None |
|
) -> Generator: |
|
yield from self.process(stream, stream_name) |
|
|
|
@abstractmethod |
|
def process(self, stream: Stream, stream_name: Optional[str] = None) -> Generator: |
|
pass |
|
|
|
def process_instance(self, instance, stream_name="tmp"): |
|
instance = self.verify_instance(instance) |
|
processed_stream = self._process_single_stream( |
|
stream_single(instance), stream_name |
|
) |
|
return next(iter(processed_stream)) |
|
|
|
|
|
class SingleStreamOperator(StreamOperator): |
|
pass |
|
|
|
|
|
class PagedStreamOperator(StreamOperator): |
|
"""A class representing a paged-stream operator in the streaming system. |
|
|
|
A paged-stream operator is a type of `StreamOperator` that operates on a page of instances |
|
in a `Stream` at a time, where a page is a subset of instances. |
|
The `process` method should be implemented by subclasses to define the specific operations |
|
to be performed on each page. |
|
|
|
Args: |
|
page_size (int): The size of each page in the stream. Defaults to 1000. |
|
""" |
|
|
|
page_size: int = 1000 |
|
|
|
def _process_stream( |
|
self, stream: Stream, stream_name: Optional[str] = None |
|
) -> Generator: |
|
page = [] |
|
for instance in stream: |
|
page.append(instance) |
|
if len(page) >= self.page_size: |
|
yield from self.process(page, stream_name) |
|
page = [] |
|
yield from self._process_page(page, stream_name) |
|
|
|
def _process_page( |
|
self, page: List[Dict], stream_name: Optional[str] = None |
|
) -> Generator: |
|
yield from self.process(page, stream_name) |
|
|
|
@abstractmethod |
|
def process(self, page: List[Dict], stream_name: Optional[str] = None) -> Generator: |
|
pass |
|
|
|
def process_instance(self, instance, stream_name="tmp"): |
|
instance = self.verify_instance(instance) |
|
processed_stream = self._process_page([instance], stream_name) |
|
return next(iter(processed_stream)) |
|
|
|
|
|
class SingleStreamReducer(StreamingOperator): |
|
"""A class representing a single-stream reducer in the streaming system. |
|
|
|
A single-stream reducer is a type of `StreamingOperator` that operates on individual `Stream` objects within a `MultiStream` and reduces each `Stream` to a single output value. The `process` method should be implemented by subclasses to define the specific reduction operation to be performed on each `Stream`. |
|
""" |
|
|
|
def __call__(self, multi_stream: Optional[MultiStream] = None) -> Dict[str, Any]: |
|
result = {} |
|
for stream_name, stream in multi_stream.items(): |
|
stream = self.process(stream) |
|
result[stream_name] = stream |
|
|
|
return result |
|
|
|
@abstractmethod |
|
def process(self, stream: Stream) -> Stream: |
|
pass |
|
|
|
|
|
class InstanceOperator(StreamOperator): |
|
"""A class representing a stream instance operator in the streaming system. |
|
|
|
A stream instance operator is a type of `StreamOperator` that operates on individual instances within a `Stream`. It iterates through each instance in the `Stream` and applies the `process` method. The `process` method should be implemented by subclasses to define the specific operations to be performed on each instance. |
|
""" |
|
|
|
def _process_stream( |
|
self, stream: Stream, stream_name: Optional[str] = None |
|
) -> Generator: |
|
try: |
|
_index = None |
|
for _index, instance in enumerate(stream): |
|
yield self._process_instance(instance, stream_name) |
|
except Exception as e: |
|
if _index is None: |
|
raise e |
|
else: |
|
raise ValueError( |
|
f"Error processing instance '{_index}' from stream '{stream_name}' in {self.__class__.__name__} due to: {e}" |
|
) from e |
|
|
|
def _process_instance( |
|
self, instance: Dict[str, Any], stream_name: Optional[str] = None |
|
) -> Dict[str, Any]: |
|
instance = self.verify_instance(instance) |
|
return self.process(instance, stream_name) |
|
|
|
@abstractmethod |
|
def process( |
|
self, instance: Dict[str, Any], stream_name: Optional[str] = None |
|
) -> Dict[str, Any]: |
|
pass |
|
|
|
def process_instance(self, instance, stream_name="tmp"): |
|
return self._process_instance(instance, stream_name) |
|
|
|
|
|
class InstanceOperatorValidator(InstanceOperator): |
|
"""A class representing a stream instance operator validator in the streaming system. |
|
|
|
A stream instance operator validator is a type of `InstanceOperator` that includes a validation step. It operates on individual instances within a `Stream` and validates the result of processing each instance. |
|
""" |
|
|
|
@abstractmethod |
|
def validate(self, instance): |
|
pass |
|
|
|
def _process_stream( |
|
self, stream: Stream, stream_name: Optional[str] = None |
|
) -> Generator: |
|
iterator = iter(stream) |
|
try: |
|
first_instance = next(iterator) |
|
except StopIteration as e: |
|
raise EmptyStreamError(f"Stream '{stream_name}' is empty") from e |
|
result = self._process_instance(first_instance, stream_name) |
|
self.validate(result) |
|
yield result |
|
yield from ( |
|
self._process_instance(instance, stream_name) for instance in iterator |
|
) |
|
|
|
|
|
class BaseFieldOperator(Artifact): |
|
"""A class representing a field operator in the streaming system. |
|
|
|
A field operator is a type of `Artifact` that operates on a single field within an instance. It takes an instance and a field name as input, processes the field, and updates the field in the instance with the processed value. |
|
""" |
|
|
|
def __call__(self, data: Dict[str, Any], field: str) -> dict: |
|
data = self.verify_instance(data) |
|
value = self.process(data[field]) |
|
data[field] = value |
|
return data |
|
|
|
@abstractmethod |
|
def process(self, value: Any) -> Any: |
|
pass |
|
|
|
|
|
class InstanceOperatorWithMultiStreamAccess(StreamingOperator): |
|
"""A class representing an instance operator with global access in the streaming system. |
|
|
|
An instance operator with global access is a type of `StreamingOperator` that operates on individual instances within a `Stream` and can also access other streams. |
|
It uses the `accessible_streams` attribute to determine which other streams it has access to. |
|
In order to make this efficient and to avoid qudratic complexity, it caches the accessible streams by default. |
|
""" |
|
|
|
def __call__(self, multi_stream: Optional[MultiStream] = None) -> MultiStream: |
|
result = {} |
|
|
|
for stream_name, stream in multi_stream.items(): |
|
stream = DynamicStream( |
|
self.generator, |
|
gen_kwargs={"stream": stream, "multi_stream": multi_stream}, |
|
) |
|
result[stream_name] = stream |
|
|
|
return MultiStream(result) |
|
|
|
def generator(self, stream, multi_stream): |
|
yield from ( |
|
self.process(self.verify_instance(instance), multi_stream) |
|
for instance in stream |
|
) |
|
|
|
@abstractmethod |
|
def process(self, instance: dict, multi_stream: MultiStream) -> dict: |
|
pass |
|
|
|
|
|
class SequentialMixin(Artifact): |
|
max_steps: Optional[int] = None |
|
steps: List[StreamingOperator] = field(default_factory=list) |
|
|
|
def num_steps(self) -> int: |
|
return len(self.steps) |
|
|
|
def set_max_steps(self, max_steps): |
|
assert ( |
|
max_steps <= self.num_steps() |
|
), f"Max steps requested ({max_steps}) is larger than defined steps {self.num_steps()}" |
|
assert max_steps >= 1, f"Max steps requested ({max_steps}) is less than 1" |
|
self.max_steps = max_steps |
|
|
|
def get_last_step_description(self): |
|
last_step = ( |
|
self.max_steps - 1 if self.max_steps is not None else len(self.steps) - 1 |
|
) |
|
return self.steps[last_step].__description__ |
|
|
|
def _get_max_steps(self): |
|
return self.max_steps if self.max_steps is not None else len(self.steps) |
|
|
|
|
|
class SequentialOperator(MultiStreamOperator, SequentialMixin): |
|
"""A class representing a sequential operator in the streaming system. |
|
|
|
A sequential operator is a type of `MultiStreamOperator` that applies a sequence of other operators to a |
|
`MultiStream`. It maintains a list of `StreamingOperator`s and applies them in order to the `MultiStream`. |
|
""" |
|
|
|
def process(self, multi_stream: Optional[MultiStream] = None) -> MultiStream: |
|
for operator in self.steps[0 : self._get_max_steps()]: |
|
multi_stream = operator(multi_stream) |
|
return multi_stream |
|
|
|
|
|
class SourceSequentialOperator(SourceOperator, SequentialMixin): |
|
"""A class representing a source sequential operator in the streaming system. |
|
|
|
A source sequential operator is a type of `SequentialOperator` that starts with a source operator. |
|
The first operator in its list of steps is a `SourceOperator`, which generates the initial `MultiStream` |
|
that the other operators then process. |
|
""" |
|
|
|
def process(self, multi_stream: Optional[MultiStream] = None) -> MultiStream: |
|
assert ( |
|
self.num_steps() > 0 |
|
), "Calling process on a SourceSequentialOperator without any steps" |
|
multi_stream = self.steps[0]() |
|
for operator in self.steps[1 : self._get_max_steps()]: |
|
multi_stream = operator(multi_stream) |
|
return multi_stream |
|
|
|
|
|
class SequentialOperatorInitializer(SequentialOperator): |
|
"""A class representing a sequential operator initializer in the streaming system. |
|
|
|
A sequential operator initializer is a type of `SequntialOperator` that starts with a stream initializer operator. The first operator in its list of steps is a `StreamInitializerOperator`, which generates the initial `MultiStream` based on the provided arguments and keyword arguments. |
|
""" |
|
|
|
def __call__(self, *args, **kwargs) -> MultiStream: |
|
return self.process(*args, **kwargs) |
|
|
|
def process(self, *args, **kwargs) -> MultiStream: |
|
assert ( |
|
self.num_steps() > 0 |
|
), "Calling process on a SequentialOperatorInitializer without any steps" |
|
|
|
assert isinstance( |
|
self.steps[0], StreamInitializerOperator |
|
), "The first step in a SequentialOperatorInitializer must be a StreamInitializerOperator" |
|
multi_stream = self.steps[0](*args, **kwargs) |
|
for operator in self.steps[1 : self._get_max_steps()]: |
|
multi_stream = operator(multi_stream) |
|
return multi_stream |
|
|