mirror of
https://github.com/xtekky/gpt4free.git
synced 2025-12-06 02:30:41 -08:00
* feat: introduce AnyProvider & LM Arena, overhaul model/provider logic
- **Provider additions & removals**
- Added `Provider/LMArenaProvider.py` with full async stream implementation and vision model support
- Registered `LMArenaProvider` in `Provider/__init__.py`; removed old `hf_space/LMArenaProvider.py`
- Created `providers/any_provider.py`; registers `AnyProvider` dynamically in `Provider`
- **Provider framework enhancements**
- `providers/base_provider.py`
- Added `video_models` and `audio_models` attributes
- `providers/retry_provider.py`
- Introduced `is_content()` helper; now treats `AudioResponse` as stream content
- **Cloudflare provider refactor**
- `Provider/Cloudflare.py`
- Re‑implemented `get_models()` with `read_models()` helper, `fallback_models`, robust nodriver/curl handling and model‑name cleaning
- **Other provider tweaks**
- `Provider/Copilot.py` – removed `"reasoning"` alias and initial `setOptions` WS message
- `Provider/PollinationsAI.py` & `PollinationsImage.py`
- Converted `audio_models` from list to dict, adjusted usage checks and labels
- `Provider/hf/__init__.py` – applies `model_aliases` remap before dispatch
- `Provider/hf_space/DeepseekAI_JanusPro7b.py` – now merges media before upload
- `needs_auth/Gemini.py` – dropped obsolete Gemini model entries
- `needs_auth/GigaChat.py` – added lowercase `"gigachat"` alias
- **API & client updates**
- Replaced `ProviderUtils` with new `Provider` map usage throughout API and GUI server
- Integrated `AnyProvider` as default fallback in `g4f/client` sync & async flows
- API endpoints now return counts of providers per model and filter by `x_ignored` header
- **GUI improvements**
- Updated JS labels with emoji icons, provider ignore logic, model count display
- **Model registry**
- Renamed base model `"GigaChat:latest"` ➜ `"gigachat"` in `models.py`
- **Miscellaneous**
- Added audio/video flags to GUI provider list
- Tightened error propagation in `retry_provider.raise_exceptions`
* Fix unittests
* fix: handle None conversation when accessing provider-specific data
- Modified `AnyProvider` class in `g4f/providers/any_provider.py`
- Updated logic to check if `conversation` is not None before accessing `provider.__name__` attribute
- Wrapped `getattr(conversation, provider.__name__, None)` block in an additional `if conversation is not None` condition
- Changed `setattr(conversation, provider.__name__, chunk)` to use `chunk.get_dict()` instead of the object directly
- Ensured consistent use of `JsonConversation` when modifying or assigning `conversation` data
* ```
feat: add provider string conversion & update IterListProvider call
- In g4f/client/__init__.py, within both Completions and AsyncCompletions, added a check to convert the provider from a string using convert_to_provider(provider) when applicable.
- In g4f/providers/any_provider.py, removed the second argument (False) from the IterListProvider constructor call in the async for loop.
```
---------
Co-authored-by: hlohaus <983577+hlohaus@users.noreply.github.com>
245 lines
No EOL
9.4 KiB
Python
245 lines
No EOL
9.4 KiB
Python
from __future__ import annotations
|
|
|
|
import random
|
|
|
|
from ..typing import Type, List, CreateResult, Messages, AsyncResult
|
|
from .types import BaseProvider, BaseRetryProvider, ProviderType
|
|
from .response import MediaResponse, AudioResponse, ProviderInfo
|
|
from .. import debug
|
|
from ..errors import RetryProviderError, RetryNoProviderError
|
|
|
|
def is_content(chunk):
|
|
return isinstance(chunk, (str, MediaResponse, AudioResponse))
|
|
|
|
class IterListProvider(BaseRetryProvider):
|
|
def __init__(
|
|
self,
|
|
providers: List[Type[BaseProvider]],
|
|
shuffle: bool = True
|
|
) -> None:
|
|
"""
|
|
Initialize the BaseRetryProvider.
|
|
Args:
|
|
providers (List[Type[BaseProvider]]): List of providers to use.
|
|
shuffle (bool): Whether to shuffle the providers list.
|
|
single_provider_retry (bool): Whether to retry a single provider if it fails.
|
|
max_retries (int): Maximum number of retries for a single provider.
|
|
"""
|
|
self.providers = providers
|
|
self.shuffle = shuffle
|
|
self.working = True
|
|
self.last_provider: Type[BaseProvider] = None
|
|
|
|
def create_completion(
|
|
self,
|
|
model: str,
|
|
messages: Messages,
|
|
stream: bool = False,
|
|
ignore_stream: bool = False,
|
|
ignored: list[str] = [],
|
|
**kwargs,
|
|
) -> CreateResult:
|
|
"""
|
|
Create a completion using available providers, with an option to stream the response.
|
|
Args:
|
|
model (str): The model to be used for completion.
|
|
messages (Messages): The messages to be used for generating completion.
|
|
stream (bool, optional): Flag to indicate if the response should be streamed. Defaults to False.
|
|
Yields:
|
|
CreateResult: Tokens or results from the completion.
|
|
Raises:
|
|
Exception: Any exception encountered during the completion process.
|
|
"""
|
|
exceptions = {}
|
|
started: bool = False
|
|
|
|
for provider in self.get_providers(stream and not ignore_stream, ignored):
|
|
self.last_provider = provider
|
|
debug.log(f"Using {provider.__name__} provider")
|
|
yield ProviderInfo(**provider.get_dict(), model=model if model else getattr(provider, "default_model"))
|
|
try:
|
|
response = provider.get_create_function()(model, messages, stream=stream, **kwargs)
|
|
for chunk in response:
|
|
if chunk:
|
|
yield chunk
|
|
if is_content(chunk):
|
|
started = True
|
|
if started:
|
|
return
|
|
except Exception as e:
|
|
exceptions[provider.__name__] = e
|
|
debug.error(f"{provider.__name__} {type(e).__name__}: {e}")
|
|
if started:
|
|
raise e
|
|
yield e
|
|
|
|
raise_exceptions(exceptions)
|
|
|
|
async def create_async_generator(
|
|
self,
|
|
model: str,
|
|
messages: Messages,
|
|
stream: bool = True,
|
|
ignore_stream: bool = False,
|
|
ignored: list[str] = [],
|
|
**kwargs
|
|
) -> AsyncResult:
|
|
exceptions = {}
|
|
started: bool = False
|
|
|
|
for provider in self.get_providers(stream and not ignore_stream, ignored):
|
|
self.last_provider = provider
|
|
debug.log(f"Using {provider.__name__} provider")
|
|
yield ProviderInfo(**provider.get_dict(), model=model if model else getattr(provider, "default_model"))
|
|
try:
|
|
response = provider.get_async_create_function()(model, messages, stream=stream, **kwargs)
|
|
if hasattr(response, "__aiter__"):
|
|
async for chunk in response:
|
|
if chunk:
|
|
yield chunk
|
|
if is_content(chunk):
|
|
started = True
|
|
elif response:
|
|
response = await response
|
|
if response:
|
|
yield response
|
|
started = True
|
|
if started:
|
|
return
|
|
except Exception as e:
|
|
exceptions[provider.__name__] = e
|
|
debug.error(f"{provider.__name__} {type(e).__name__}: {e}")
|
|
if started:
|
|
raise e
|
|
yield e
|
|
|
|
raise_exceptions(exceptions)
|
|
|
|
def get_create_function(self) -> callable:
|
|
return self.create_completion
|
|
|
|
def get_async_create_function(self) -> callable:
|
|
return self.create_async_generator
|
|
|
|
def get_providers(self, stream: bool, ignored: list[str]) -> list[ProviderType]:
|
|
providers = [p for p in self.providers if (p.supports_stream or not stream) and p.__name__ not in ignored]
|
|
if self.shuffle:
|
|
random.shuffle(providers)
|
|
return providers
|
|
|
|
class RetryProvider(IterListProvider):
|
|
def __init__(
|
|
self,
|
|
providers: List[Type[BaseProvider]],
|
|
shuffle: bool = True,
|
|
single_provider_retry: bool = False,
|
|
max_retries: int = 3,
|
|
) -> None:
|
|
"""
|
|
Initialize the BaseRetryProvider.
|
|
Args:
|
|
providers (List[Type[BaseProvider]]): List of providers to use.
|
|
shuffle (bool): Whether to shuffle the providers list.
|
|
single_provider_retry (bool): Whether to retry a single provider if it fails.
|
|
max_retries (int): Maximum number of retries for a single provider.
|
|
"""
|
|
super().__init__(providers, shuffle)
|
|
self.single_provider_retry = single_provider_retry
|
|
self.max_retries = max_retries
|
|
|
|
def create_completion(
|
|
self,
|
|
model: str,
|
|
messages: Messages,
|
|
stream: bool = False,
|
|
**kwargs,
|
|
) -> CreateResult:
|
|
"""
|
|
Create a completion using available providers, with an option to stream the response.
|
|
Args:
|
|
model (str): The model to be used for completion.
|
|
messages (Messages): The messages to be used for generating completion.
|
|
stream (bool, optional): Flag to indicate if the response should be streamed. Defaults to False.
|
|
Yields:
|
|
CreateResult: Tokens or results from the completion.
|
|
Raises:
|
|
Exception: Any exception encountered during the completion process.
|
|
"""
|
|
if self.single_provider_retry:
|
|
exceptions = {}
|
|
started: bool = False
|
|
provider = self.providers[0]
|
|
self.last_provider = provider
|
|
for attempt in range(self.max_retries):
|
|
try:
|
|
if debug.logging:
|
|
print(f"Using {provider.__name__} provider (attempt {attempt + 1})")
|
|
response = provider.get_create_function()(model, messages, stream=stream, **kwargs)
|
|
for chunk in response:
|
|
yield chunk
|
|
if is_content(chunk):
|
|
started = True
|
|
if started:
|
|
return
|
|
except Exception as e:
|
|
exceptions[provider.__name__] = e
|
|
if debug.logging:
|
|
print(f"{provider.__name__}: {e.__class__.__name__}: {e}")
|
|
if started:
|
|
raise e
|
|
raise_exceptions(exceptions)
|
|
else:
|
|
yield from super().create_completion(model, messages, stream, **kwargs)
|
|
|
|
async def create_async_generator(
|
|
self,
|
|
model: str,
|
|
messages: Messages,
|
|
stream: bool = True,
|
|
**kwargs
|
|
) -> AsyncResult:
|
|
exceptions = {}
|
|
started = False
|
|
|
|
if self.single_provider_retry:
|
|
provider = self.providers[0]
|
|
self.last_provider = provider
|
|
for attempt in range(self.max_retries):
|
|
try:
|
|
debug.log(f"Using {provider.__name__} provider (attempt {attempt + 1})")
|
|
response = provider.get_async_create_function()(model, messages, stream=stream, **kwargs)
|
|
if hasattr(response, "__aiter__"):
|
|
async for chunk in response:
|
|
yield chunk
|
|
if is_content(chunk):
|
|
started = True
|
|
else:
|
|
response = await response
|
|
if response:
|
|
yield response
|
|
started = True
|
|
if started:
|
|
return
|
|
except Exception as e:
|
|
exceptions[provider.__name__] = e
|
|
if debug.logging:
|
|
print(f"{provider.__name__}: {e.__class__.__name__}: {e}")
|
|
raise_exceptions(exceptions)
|
|
else:
|
|
async for chunk in super().create_async_generator(model, messages, stream, **kwargs):
|
|
yield chunk
|
|
|
|
def raise_exceptions(exceptions: dict) -> None:
|
|
"""
|
|
Raise a combined exception if any occurred during retries.
|
|
|
|
Raises:
|
|
RetryProviderError: If any provider encountered an exception.
|
|
RetryNoProviderError: If no provider is found.
|
|
"""
|
|
if exceptions:
|
|
raise RetryProviderError("RetryProvider failed:\n" + "\n".join([
|
|
f"{p}: {type(exception).__name__}: {exception}" for p, exception in exceptions.items()
|
|
])) from list(exceptions.values())[0]
|
|
|
|
raise RetryNoProviderError("No provider found") |