mirror of
https://github.com/xtekky/gpt4free.git
synced 2026-01-01 06:41:09 -08:00
Updates for the response of the BackendApi Update of the demo model list Improve web search tool Moved copy_images to /image
74 lines
3.3 KiB
Python
74 lines
3.3 KiB
Python
from __future__ import annotations
|
|
|
|
import json
|
|
from aiohttp import ClientSession
|
|
|
|
from ...typing import AsyncResult, Messages
|
|
from ...providers.response import ImageResponse, ImagePreview
|
|
from ...errors import ResponseError
|
|
from ..base_provider import AsyncGeneratorProvider, ProviderModelMixin
|
|
from ..helper import format_image_prompt
|
|
|
|
class StableDiffusion35Large(AsyncGeneratorProvider, ProviderModelMixin):
|
|
url = "https://stabilityai-stable-diffusion-3-5-large.hf.space"
|
|
api_endpoint = "/gradio_api/call/infer"
|
|
|
|
working = True
|
|
|
|
default_model = 'stabilityai-stable-diffusion-3-5-large'
|
|
default_image_model = default_model
|
|
image_models = [default_model]
|
|
models = image_models
|
|
model_aliases = {"sd-3.5": default_model}
|
|
|
|
@classmethod
|
|
async def create_async_generator(
|
|
cls, model: str, messages: Messages,
|
|
prompt: str = None,
|
|
negative_prompt: str = None,
|
|
api_key: str = None,
|
|
proxy: str = None,
|
|
width: int = 1024,
|
|
height: int = 1024,
|
|
guidance_scale: float = 4.5,
|
|
num_inference_steps: int = 50,
|
|
seed: int = 0,
|
|
randomize_seed: bool = True,
|
|
**kwargs
|
|
) -> AsyncResult:
|
|
headers = {
|
|
"Content-Type": "application/json",
|
|
"Accept": "application/json",
|
|
}
|
|
if api_key is not None:
|
|
headers["Authorization"] = f"Bearer {api_key}"
|
|
async with ClientSession(headers=headers) as session:
|
|
prompt = format_image_prompt(messages, prompt)
|
|
data = {
|
|
"data": [prompt, negative_prompt, seed, randomize_seed, width, height, guidance_scale, num_inference_steps]
|
|
}
|
|
async with session.post(f"{cls.url}{cls.api_endpoint}", json=data, proxy=proxy) as response:
|
|
response.raise_for_status()
|
|
event_id = (await response.json()).get("event_id")
|
|
async with session.get(f"{cls.url}{cls.api_endpoint}/{event_id}") as event_response:
|
|
event_response.raise_for_status()
|
|
event = None
|
|
async for chunk in event_response.content:
|
|
if chunk.startswith(b"event: "):
|
|
event = chunk[7:].decode(errors="replace").strip()
|
|
if chunk.startswith(b"data: "):
|
|
if event == "error":
|
|
raise ResponseError(f"GPU token limit exceeded: {chunk.decode(errors='replace')}")
|
|
if event in ("complete", "generating"):
|
|
try:
|
|
data = json.loads(chunk[6:])
|
|
if data is None:
|
|
continue
|
|
url = data[0]["url"]
|
|
except (json.JSONDecodeError, KeyError, TypeError) as e:
|
|
raise RuntimeError(f"Failed to parse image URL: {chunk.decode(errors='replace')}", e)
|
|
if event == "generating":
|
|
yield ImagePreview(url, prompt)
|
|
else:
|
|
yield ImageResponse(url, prompt)
|
|
break
|