mirror of
https://github.com/xtekky/gpt4free.git
synced 2025-12-06 02:30:41 -08:00
Fix unittests
This commit is contained in:
parent
5b65101a2a
commit
2aa40bb8a8
4 changed files with 9 additions and 6 deletions
|
|
@ -19,7 +19,7 @@ class TestProviderHasModel(unittest.TestCase):
|
||||||
if model.name in provider.model_aliases:
|
if model.name in provider.model_aliases:
|
||||||
model_name = provider.model_aliases[model.name]
|
model_name = provider.model_aliases[model.name]
|
||||||
else:
|
else:
|
||||||
model_name = model.name
|
model_name = model.get_long_name()
|
||||||
self.provider_has_model(provider, model_name)
|
self.provider_has_model(provider, model_name)
|
||||||
|
|
||||||
def provider_has_model(self, provider: Type[BaseProvider], model: str):
|
def provider_has_model(self, provider: Type[BaseProvider], model: str):
|
||||||
|
|
@ -29,6 +29,7 @@ class TestProviderHasModel(unittest.TestCase):
|
||||||
except (MissingRequirementsError, MissingAuthError):
|
except (MissingRequirementsError, MissingAuthError):
|
||||||
return
|
return
|
||||||
if self.cache[provider.__name__]:
|
if self.cache[provider.__name__]:
|
||||||
|
if not model in provider.model_aliases:
|
||||||
self.assertIn(model, self.cache[provider.__name__], provider.__name__)
|
self.assertIn(model, self.cache[provider.__name__], provider.__name__)
|
||||||
|
|
||||||
def test_all_providers_working(self):
|
def test_all_providers_working(self):
|
||||||
|
|
|
||||||
|
|
@ -16,6 +16,7 @@ class Kimi(AsyncAuthedProvider, ProviderModelMixin):
|
||||||
active_by_default = True
|
active_by_default = True
|
||||||
default_model = "kimi-k2"
|
default_model = "kimi-k2"
|
||||||
models = [default_model]
|
models = [default_model]
|
||||||
|
model_aliases = {"moonshotai/Kimi-K2-Instruct": default_model}
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
async def on_auth_async(cls, proxy: str = None, **kwargs) -> AsyncIterator:
|
async def on_auth_async(cls, proxy: str = None, **kwargs) -> AsyncIterator:
|
||||||
|
|
|
||||||
|
|
@ -88,7 +88,8 @@ class PollinationsAI(AsyncGeneratorProvider, ProviderModelMixin):
|
||||||
vision_models = [default_vision_model]
|
vision_models = [default_vision_model]
|
||||||
_models_loaded = False
|
_models_loaded = False
|
||||||
model_aliases = {
|
model_aliases = {
|
||||||
"gpt-4.1-nano": "openai",
|
"gpt-4.1-mini": "openai",
|
||||||
|
"gpt-4.1-nano": "openai-fast",
|
||||||
"gpt-4.1": "openai-large",
|
"gpt-4.1": "openai-large",
|
||||||
"o4-mini": "openai-reasoning",
|
"o4-mini": "openai-reasoning",
|
||||||
"qwen-2.5-coder-32b": "qwen-coder",
|
"qwen-2.5-coder-32b": "qwen-coder",
|
||||||
|
|
|
||||||
|
|
@ -197,20 +197,20 @@ default_vision = VisionModel(
|
||||||
gpt_4 = Model(
|
gpt_4 = Model(
|
||||||
name = 'gpt-4',
|
name = 'gpt-4',
|
||||||
base_provider = 'OpenAI',
|
base_provider = 'OpenAI',
|
||||||
best_provider = IterListProvider([Blackbox, PollinationsAI, Copilot, Yqcloud, WeWordle, OpenaiChat])
|
best_provider = IterListProvider([Blackbox, Copilot, Yqcloud, WeWordle, OpenaiChat])
|
||||||
)
|
)
|
||||||
|
|
||||||
# gpt-4o
|
# gpt-4o
|
||||||
gpt_4o = VisionModel(
|
gpt_4o = VisionModel(
|
||||||
name = 'gpt-4o',
|
name = 'gpt-4o',
|
||||||
base_provider = 'OpenAI',
|
base_provider = 'OpenAI',
|
||||||
best_provider = IterListProvider([Blackbox, PollinationsAI, OpenaiChat])
|
best_provider = IterListProvider([Blackbox, OpenaiChat])
|
||||||
)
|
)
|
||||||
|
|
||||||
gpt_4o_mini = Model(
|
gpt_4o_mini = Model(
|
||||||
name = 'gpt-4o-mini',
|
name = 'gpt-4o-mini',
|
||||||
base_provider = 'OpenAI',
|
base_provider = 'OpenAI',
|
||||||
best_provider = IterListProvider([Blackbox, PollinationsAI, Chatai, OIVSCodeSer2, Startnest, OpenaiChat])
|
best_provider = IterListProvider([Blackbox, Chatai, OIVSCodeSer2, Startnest, OpenaiChat])
|
||||||
)
|
)
|
||||||
|
|
||||||
gpt_4o_mini_audio = AudioModel(
|
gpt_4o_mini_audio = AudioModel(
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue