hf-llm-api / networks /message_streamer.py
Hansimov's picture
:pencil: [Doc] Readme for features, and comments for references
e820e51
raw
history blame
3.58 kB
import json
import re
import requests
from messagers.message_outputer import OpenaiStreamOutputer
from utils.logger import logger
from utils.enver import enver
class MessageStreamer:
MODEL_MAP = {
"mixtral-8x7b": "mistralai/Mixtral-8x7B-Instruct-v0.1", # 72.62, fast [Recommended]
"mistral-7b": "mistralai/Mistral-7B-Instruct-v0.2", # 65.71, fast
"openchat-3.5": "openchat/openchat_3.5", # 61.24, fast
# "zephyr-7b-alpha": "HuggingFaceH4/zephyr-7b-alpha", # 59.5, fast
# "zephyr-7b-beta": "HuggingFaceH4/zephyr-7b-beta", # 61.95, slow
"default": "mistralai/Mixtral-8x7B-Instruct-v0.1",
}
def __init__(self, model: str):
if model in self.MODEL_MAP.keys():
self.model = model
else:
self.model = "default"
self.model_fullname = self.MODEL_MAP[self.model]
self.message_outputer = OpenaiStreamOutputer()
def parse_line(self, line):
line = line.decode("utf-8")
line = re.sub(r"data:\s*", "", line)
data = json.loads(line)
content = data["token"]["text"]
return content
def chat(
self,
prompt: str = None,
temperature: float = 0.01,
max_new_tokens: int = 8192,
stream: bool = True,
yield_output: bool = False,
):
# https://huggingface.co/docs/api-inference/detailed_parameters?code=curl
# curl --proxy http://<server>:<port> https://api-inference.huggingface.co/models/<org>/<model_name> -X POST -d '{"inputs":"who are you?","parameters":{"max_new_token":64}}' -H 'Content-Type: application/json' -H 'Authorization: Bearer <HF_TOKEN>'
self.request_url = (
f"https://api-inference.huggingface.co/models/{self.model_fullname}"
)
self.request_headers = {
"Content-Type": "application/json",
}
# References:
# huggingface_hub/inference/_client.py:
# class InferenceClient > def text_generation()
# huggingface_hub/inference/_text_generation.py:
# class TextGenerationRequest > param `stream`
# https://huggingface.co/docs/text-generation-inference/conceptual/streaming#streaming-with-curl
self.request_body = {
"inputs": prompt,
"parameters": {
"temperature": max(temperature, 0.01), # must be positive
"max_new_tokens": max_new_tokens,
"return_full_text": False,
},
"stream": stream,
}
logger.back(self.request_url)
enver.set_envs(proxies=True)
stream = requests.post(
self.request_url,
headers=self.request_headers,
json=self.request_body,
proxies=enver.requests_proxies,
stream=stream,
)
status_code = stream.status_code
if status_code == 200:
logger.success(status_code)
else:
logger.err(status_code)
for line in stream.iter_lines():
if not line:
continue
content = self.parse_line(line)
if content.strip() == "</s>":
content_type = "Finished"
logger.success("\n[Finished]")
else:
content_type = "Completions"
logger.back(content, end="")
if yield_output:
output = self.message_outputer.output(
content=content, content_type=content_type
)
yield output