|
import uuid |
|
|
|
import requests |
|
|
|
from ..typing import Any, CreateResult |
|
from .base_provider import BaseProvider |
|
|
|
|
|
class I207m(BaseProvider): |
|
url = "https://gpt3.i207m.top" |
|
supports_stream = True |
|
needs_auth = False |
|
supports_gpt_4 = True |
|
supports_gpt_4_32k = True |
|
|
|
@staticmethod |
|
def create_completion( |
|
model: str, |
|
messages: list[dict[str, str]], |
|
stream: bool, |
|
**kwargs: Any, |
|
) -> CreateResult: |
|
headers = { |
|
"authority": "gpt3.i207m.top", |
|
"content-type": "application/json", |
|
"origin": "https://gpt3.i207m.top", |
|
"referer": "https://gpt3.i207m.top/", |
|
"user-agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/112.0.0.0 Safari/537.36", |
|
"x-auth-code": str(kwargs.get("auth")), |
|
} |
|
models = { |
|
"gpt-4": { |
|
"id": "gpt-4", |
|
"name": "GPT-4", |
|
"maxLength": 24000, |
|
"tokenLimit": 8000, |
|
}, |
|
"gpt-4-32k": { |
|
"id":"gpt-4-32k", |
|
"name":"GPT-4-32k", |
|
"maxLength":96000, |
|
"tokenLimit":32000 |
|
}, |
|
} |
|
json_data = { |
|
"conversationId": str(uuid.uuid4()), |
|
"model": models[model], |
|
"messages": messages, |
|
"key": "", |
|
"prompt": "You are ChatGPT, a large language model trained by OpenAI. Follow the user's instructions carefully. Respond using markdown.", |
|
} |
|
|
|
response = requests.post( |
|
"https://gpt3.i207m.top/api/chat", |
|
headers=headers, |
|
json=json_data, |
|
stream=True, |
|
) |
|
response.raise_for_status() |
|
for token in response.iter_content(chunk_size=2046): |
|
yield token.decode("utf-8") |
|
|
|
@classmethod |
|
@property |
|
def params(cls): |
|
params = [ |
|
("model", "str"), |
|
("messages", "list[dict[str, str]]"), |
|
("stream", "bool"), |
|
("auth", "str"), |
|
] |
|
param = ", ".join([": ".join(p) for p in params]) |
|
return f"g4f.provider.{cls.__name__} supports: ({param})" |
|
|