import logging

import openai

from core.model_providers.error import LLMBadRequestError
from core.model_providers.providers.base import BaseModelProvider
from core.model_providers.providers.hosted import hosted_config, hosted_model_providers
from models.provider import ProviderType


def check_moderation(model_provider: BaseModelProvider, text: str) -> bool:
    if hosted_config.moderation.enabled is True and hosted_model_providers.openai:
        if model_provider.provider.provider_type == ProviderType.SYSTEM.value \
                and model_provider.provider_name in hosted_config.moderation.providers:
            # 2000 text per chunk
            length = 2000
            text_chunks = [text[i:i + length] for i in range(0, len(text), length)]

            max_text_chunks = 32
            chunks = [text_chunks[i:i + max_text_chunks] for i in range(0, len(text_chunks), max_text_chunks)]

            for text_chunk in chunks:
                try:
                    moderation_result = openai.Moderation.create(input=text_chunk,
                                                                 api_key=hosted_model_providers.openai.api_key)
                except Exception as ex:
                    logging.exception(ex)
                    raise LLMBadRequestError('Rate limit exceeded, please try again later.')

                for result in moderation_result.results:
                    if result['flagged'] is True:
                        return False

    return True
