gpt4free/g4f/client/stubs.py
2025-05-16 00:27:27 +02:00

301 lines
No EOL
9 KiB
Python

from __future__ import annotations
import os
from typing import Optional, List
from time import time
from ..image import extract_data_uri
from ..image.copy_images import get_media_dir
from ..client.helper import filter_markdown
from .helper import filter_none
try:
from pydantic import BaseModel, field_serializer
except ImportError:
class BaseModel():
@classmethod
def model_construct(cls, **data):
new = cls()
for key, value in data.items():
setattr(new, key, value)
return new
class field_serializer():
def __init__(self, field_name):
self.field_name = field_name
def __call__(self, *args, **kwargs):
return args[0]
class BaseModel(BaseModel):
@classmethod
def model_construct(cls, **data):
if hasattr(super(), "model_construct"):
return super().model_construct(**data)
return cls.construct(**data)
class TokenDetails(BaseModel):
cached_tokens: int
class UsageModel(BaseModel):
prompt_tokens: int
completion_tokens: int
total_tokens: int
prompt_tokens_details: TokenDetails
completion_tokens_details: TokenDetails
@classmethod
def model_construct(cls, prompt_tokens=0, completion_tokens=0, total_tokens=0, prompt_tokens_details=None, completion_tokens_details=None, **kwargs):
return super().model_construct(
prompt_tokens=prompt_tokens,
completion_tokens=completion_tokens,
total_tokens=total_tokens,
prompt_tokens_details=TokenDetails.model_construct(**prompt_tokens_details if prompt_tokens_details else {"cached_tokens": 0}),
completion_tokens_details=TokenDetails.model_construct(**completion_tokens_details if completion_tokens_details else {}),
**kwargs
)
class ToolFunctionModel(BaseModel):
name: str
arguments: str
class ToolCallModel(BaseModel):
id: str
type: str
function: ToolFunctionModel
@classmethod
def model_construct(cls, function=None, **kwargs):
return super().model_construct(
**kwargs,
function=ToolFunctionModel.model_construct(**function),
)
class ChatCompletionChunk(BaseModel):
id: str
object: str
created: int
model: str
provider: Optional[str]
choices: List[ChatCompletionDeltaChoice]
usage: UsageModel
conversation: dict
@classmethod
def model_construct(
cls,
content: str,
finish_reason: str,
completion_id: str = None,
created: int = None,
usage: UsageModel = None,
conversation: dict = None
):
return super().model_construct(
id=f"chatcmpl-{completion_id}" if completion_id else None,
object="chat.completion.chunk",
created=created,
model=None,
provider=None,
choices=[ChatCompletionDeltaChoice.model_construct(
ChatCompletionDelta.model_construct(content),
finish_reason
)],
**filter_none(usage=usage, conversation=conversation)
)
@field_serializer('conversation')
def serialize_conversation(self, conversation: dict):
if hasattr(conversation, "get_dict"):
return conversation.get_dict()
return conversation
class ResponseMessage(BaseModel):
type: str = "message"
role: str
content: list[ResponseMessageContent]
@classmethod
def model_construct(cls, content: str):
return super().model_construct(role="assistant", content=[ResponseMessageContent.model_construct(content)])
class ResponseMessageContent(BaseModel):
type: str
text: str
@classmethod
def model_construct(cls, text: str):
return super().model_construct(type="output_text", text=text)
@field_serializer('text')
def serialize_text(self, text: str):
return str(text)
class ChatCompletionMessage(BaseModel):
role: str
content: str
tool_calls: list[ToolCallModel] = None
@classmethod
def model_construct(cls, content: str, tool_calls: list = None):
return super().model_construct(role="assistant", content=content, **filter_none(tool_calls=tool_calls))
@field_serializer('content')
def serialize_content(self, content: str):
return str(content)
def save(self, filepath: str, allowed_types = None):
if hasattr(self.content, "data"):
os.rename(self.content.data.replace("/media", get_media_dir()), filepath)
return
if self.content.startswith("data:"):
with open(filepath, "wb") as f:
f.write(extract_data_uri(self.content))
return
content = filter_markdown(self.content, allowed_types)
if content is not None:
with open(filepath, "w") as f:
f.write(content)
class ChatCompletionChoice(BaseModel):
index: int
message: ChatCompletionMessage
finish_reason: str
@classmethod
def model_construct(cls, message: ChatCompletionMessage, finish_reason: str):
return super().model_construct(index=0, message=message, finish_reason=finish_reason)
class ChatCompletion(BaseModel):
id: str
object: str
created: int
model: str
provider: Optional[str]
choices: list[ChatCompletionChoice]
usage: UsageModel
conversation: dict
@classmethod
def model_construct(
cls,
content: str,
finish_reason: str,
completion_id: str = None,
created: int = None,
tool_calls: list[ToolCallModel] = None,
usage: UsageModel = None,
conversation: dict = None
):
return super().model_construct(
id=f"chatcmpl-{completion_id}" if completion_id else None,
object="chat.completion",
created=created,
model=None,
provider=None,
choices=[ChatCompletionChoice.model_construct(
ChatCompletionMessage.model_construct(content, tool_calls),
finish_reason,
)],
**filter_none(usage=usage, conversation=conversation)
)
@field_serializer('conversation')
def serialize_conversation(self, conversation: dict):
if hasattr(conversation, "get_dict"):
return conversation.get_dict()
return conversation
class ClientResponse(BaseModel):
id: str
object: str
created_at: int
model: str
provider: Optional[str]
output: list[ResponseMessage]
usage: UsageModel
conversation: dict
@classmethod
def model_construct(
cls,
content: str,
response_id: str = None,
created_at: int = None,
usage: UsageModel = None,
conversation: dict = None
) -> ClientResponse:
return super().model_construct(
id=f"resp-{response_id}" if response_id else None,
object="response",
created_at=created_at,
model=None,
provider=None,
output=[
ResponseMessage.model_construct(content),
],
**filter_none(usage=usage, conversation=conversation)
)
@field_serializer('conversation')
def serialize_conversation(self, conversation: dict):
if hasattr(conversation, "get_dict"):
return conversation.get_dict()
return conversation
class ChatCompletionDelta(BaseModel):
role: str
content: Optional[str]
@classmethod
def model_construct(cls, content: Optional[str]):
return super().model_construct(role="assistant", content=content)
@field_serializer('content')
def serialize_content(self, content: Optional[str]):
if content is None:
return ""
return str(content)
class ChatCompletionDeltaChoice(BaseModel):
index: int
delta: ChatCompletionDelta
finish_reason: Optional[str]
@classmethod
def model_construct(cls, delta: ChatCompletionDelta, finish_reason: Optional[str]):
return super().model_construct(index=0, delta=delta, finish_reason=finish_reason)
class Image(BaseModel):
url: Optional[str]
b64_json: Optional[str]
revised_prompt: Optional[str]
@classmethod
def model_construct(cls, url: str = None, b64_json: str = None, revised_prompt: str = None):
return super().model_construct(**filter_none(
url=url,
b64_json=b64_json,
revised_prompt=revised_prompt
))
def save(self, path: str):
if self.url is not None and self.url.startswith("/media/"):
os.rename(self.url.replace("/media", get_media_dir()), path)
class ImagesResponse(BaseModel):
data: List[Image]
model: str
provider: str
created: int
@classmethod
def model_construct(cls, data: List[Image], created: int = None, model: str = None, provider: str = None):
if created is None:
created = int(time())
return super().model_construct(
data=data,
model=model,
provider=provider,
created=created
)