zero2story / modules /llms /llm_factory_abstracts.py
chansung's picture
.
3332aa4
from abc import ABCMeta, abstractmethod
class LLMFactory(metaclass=ABCMeta):
@abstractmethod
def create_prompt_format(self):
pass
@abstractmethod
def create_prompt_manager(self):
pass
@abstractmethod
def create_pp_manager(self):
pass
@abstractmethod
def create_ui_pp_manager(self):
pass
@abstractmethod
def create_llm_service(self):
pass
class PromptFmt(metaclass=ABCMeta):
@classmethod
@abstractmethod
def ctx(cls, context):
pass
@classmethod
@abstractmethod
def prompt(cls, pingpong, truncate_size):
pass
class PromptManager(metaclass=ABCMeta):
@abstractmethod
def load_prompts(self):
pass
@abstractmethod
def reload_prompts(self):
pass
@property
@abstractmethod
def prompts_path(self):
pass
@prompts_path.setter
@abstractmethod
def prompts_path(self, prompts_path):
pass
@property
@abstractmethod
def prompts(self):
pass
class PPManager(metaclass=ABCMeta):
@abstractmethod
def build_prompts(self, from_idx: int=0, to_idx: int=-1, fmt: PromptFmt=None, truncate_size: int=None):
pass
class UIPPManager(PPManager, metaclass=ABCMeta):
@abstractmethod
def build_uis(self, from_idx: int=0, to_idx: int=-1):
pass
class LLMService(metaclass=ABCMeta):
@abstractmethod
def make_params(self, mode="chat",
temperature=None,
candidate_count=None,
top_k=None,
top_p=None,
max_output_tokens=None,
use_filter=True):
pass
@abstractmethod
async def gen_text(self, prompt, mode="chat", parameters=None, use_filter=True):
pass