Spaces:
Runtime error
Runtime error
import time | |
from enum import Enum, unique | |
from typing import List, Optional | |
from pydantic import BaseModel, Field | |
from typing_extensions import Literal | |
class Role(str, Enum): | |
USER = "user" | |
ASSISTANT = "assistant" | |
SYSTEM = "system" | |
FUNCTION = "function" | |
TOOL = "tool" | |
class Finish(str, Enum): | |
STOP = "stop" | |
LENGTH = "length" | |
TOOL = "tool_calls" | |
class ModelCard(BaseModel): | |
id: str | |
object: Literal["model"] = "model" | |
created: int = Field(default_factory=lambda: int(time.time())) | |
owned_by: Literal["owner"] = "owner" | |
class ModelList(BaseModel): | |
object: Literal["list"] = "list" | |
data: List[ModelCard] = [] | |
class Function(BaseModel): | |
name: str | |
arguments: str | |
class FunctionCall(BaseModel): | |
id: Literal["call_default"] = "call_default" | |
type: Literal["function"] = "function" | |
function: Function | |
class ChatMessage(BaseModel): | |
role: Role | |
content: str | |
class ChatCompletionMessage(BaseModel): | |
role: Optional[Role] = None | |
content: Optional[str] = None | |
tool_calls: Optional[List[FunctionCall]] = None | |
class ChatCompletionRequest(BaseModel): | |
model: str | |
messages: List[ChatMessage] | |
tools: Optional[list] = [] | |
do_sample: bool = True | |
temperature: Optional[float] = None | |
top_p: Optional[float] = None | |
n: int = 1 | |
max_tokens: Optional[int] = None | |
stream: bool = False | |
class ChatCompletionResponseChoice(BaseModel): | |
index: int | |
message: ChatCompletionMessage | |
finish_reason: Finish | |
class ChatCompletionResponseStreamChoice(BaseModel): | |
index: int | |
delta: ChatCompletionMessage | |
finish_reason: Optional[Finish] = None | |
class ChatCompletionResponseUsage(BaseModel): | |
prompt_tokens: int | |
completion_tokens: int | |
total_tokens: int | |
class ChatCompletionResponse(BaseModel): | |
id: Literal["chatcmpl-default"] = "chatcmpl-default" | |
object: Literal["chat.completion"] = "chat.completion" | |
created: int = Field(default_factory=lambda: int(time.time())) | |
model: str | |
choices: List[ChatCompletionResponseChoice] | |
usage: ChatCompletionResponseUsage | |
class ChatCompletionStreamResponse(BaseModel): | |
id: Literal["chatcmpl-default"] = "chatcmpl-default" | |
object: Literal["chat.completion.chunk"] = "chat.completion.chunk" | |
created: int = Field(default_factory=lambda: int(time.time())) | |
model: str | |
choices: List[ChatCompletionResponseStreamChoice] | |
class ScoreEvaluationRequest(BaseModel): | |
model: str | |
messages: List[str] | |
max_length: Optional[int] = None | |
class ScoreEvaluationResponse(BaseModel): | |
id: Literal["scoreeval-default"] = "scoreeval-default" | |
object: Literal["score.evaluation"] = "score.evaluation" | |
model: str | |
scores: List[float] | |