mirror of
https://github.com/xtekky/gpt4free.git
synced 2025-12-06 02:30:41 -08:00
Add g4f cli client
This commit is contained in:
parent
634f7d9a39
commit
eff8624ed0
4 changed files with 272 additions and 25 deletions
|
|
@ -1,6 +1,6 @@
|
|||
from __future__ import annotations
|
||||
|
||||
from .cli import get_api_parser, run_api_args
|
||||
from .cli.__main__ import get_api_parser, run_api_args
|
||||
|
||||
parser = get_api_parser()
|
||||
args = parser.parse_args()
|
||||
|
|
|
|||
|
|
@ -2,6 +2,7 @@ from __future__ import annotations
|
|||
|
||||
import argparse
|
||||
from argparse import ArgumentParser
|
||||
from .client import get_parser, run_client_args
|
||||
|
||||
from g4f import Provider
|
||||
from g4f.gui.run import gui_parser, run_gui_args
|
||||
|
|
@ -41,12 +42,15 @@ def main():
|
|||
subparsers = parser.add_subparsers(dest="mode", help="Mode to run the g4f in.")
|
||||
subparsers.add_parser("api", parents=[get_api_parser()], add_help=False)
|
||||
subparsers.add_parser("gui", parents=[gui_parser()], add_help=False)
|
||||
subparsers.add_parser("client", parents=[get_parser()], add_help=False)
|
||||
|
||||
args = parser.parse_args()
|
||||
if args.mode == "api":
|
||||
run_api_args(args)
|
||||
elif args.mode == "gui":
|
||||
run_gui_args(args)
|
||||
elif args.mode == "client":
|
||||
run_client_args(args)
|
||||
else:
|
||||
parser.print_help()
|
||||
exit(1)
|
||||
222
g4f/cli/client.py
Normal file
222
g4f/cli/client.py
Normal file
|
|
@ -0,0 +1,222 @@
|
|||
#!/usr/bin/env python3
|
||||
|
||||
import os
|
||||
import sys
|
||||
import asyncio
|
||||
import json
|
||||
import argparse
|
||||
from pathlib import Path
|
||||
from typing import Optional, List, Dict
|
||||
from g4f.client import AsyncClient
|
||||
from g4f.providers.response import JsonConversation
|
||||
from g4f.cookies import set_cookies_dir, read_cookie_files
|
||||
from g4f.Provider import ProviderUtils
|
||||
from g4f import debug
|
||||
|
||||
# Platform-appropriate directories
|
||||
def get_config_dir() -> Path:
|
||||
"""Get platform-appropriate config directory."""
|
||||
if sys.platform == "win32":
|
||||
return Path(os.environ.get("APPDATA", Path.home() / "AppData" / "Roaming"))
|
||||
elif sys.platform == "darwin":
|
||||
return Path.home() / "Library" / "Application Support"
|
||||
else: # Linux and other UNIX-like
|
||||
return Path.home() / ".config"
|
||||
|
||||
CONFIG_DIR = get_config_dir() / "g4f-cli"
|
||||
COOKIES_DIR = CONFIG_DIR / "cookies"
|
||||
CONVERSATION_FILE = CONFIG_DIR / "conversation.json"
|
||||
|
||||
class ConversationManager:
|
||||
"""Manages conversation history and state."""
|
||||
|
||||
def __init__(self, file_path: Path, model: Optional[str] = None, provider: Optional[str] = None) -> None:
|
||||
self.file_path = file_path
|
||||
self.model: Optional[str] = model
|
||||
self.provider: Optional[str] = provider
|
||||
self.conversation = None
|
||||
self.history: List[Dict[str, str]] = []
|
||||
self._load()
|
||||
|
||||
def _load(self) -> None:
|
||||
"""Load conversation from file."""
|
||||
if not self.file_path.exists():
|
||||
return
|
||||
|
||||
try:
|
||||
with open(self.file_path, 'r', encoding='utf-8') as f:
|
||||
data = json.load(f)
|
||||
self.model = data.get("model")
|
||||
self.provider = data.get("provider") if self.provider is None else self.provider
|
||||
self.data = data.get("data", {})
|
||||
if self.provider and data.get(self.provider):
|
||||
self.conversation = JsonConversation(**data.get(self.provider))
|
||||
self.history = data.get("history", [])
|
||||
except (json.JSONDecodeError, KeyError) as e:
|
||||
print(f"Error loading conversation: {e}", file=sys.stderr)
|
||||
except Exception as e:
|
||||
print(f"Unexpected error loading conversation: {e}", file=sys.stderr)
|
||||
|
||||
def save(self) -> None:
|
||||
"""Save conversation to file."""
|
||||
try:
|
||||
with open(self.file_path, 'w', encoding='utf-8') as f:
|
||||
if self.conversation and self.provider:
|
||||
self.data[self.provider] = self.conversation.get_dict()
|
||||
else:
|
||||
self.data = self.conversation.get_dict() if self.conversation else {}
|
||||
json.dump({
|
||||
"model": self.model,
|
||||
"provider": self.provider,
|
||||
"data": self.data,
|
||||
"history": self.history
|
||||
}, f, indent=2, ensure_ascii=False)
|
||||
except Exception as e:
|
||||
print(f"Error saving conversation: {e}", file=sys.stderr)
|
||||
|
||||
def add_message(self, role: str, content: str) -> None:
|
||||
"""Add a message to the conversation."""
|
||||
self.history.append({"role": role, "content": content})
|
||||
|
||||
def get_messages(self) -> List[Dict[str, str]]:
|
||||
"""Get all messages in the conversation."""
|
||||
return self.history
|
||||
|
||||
async def stream_response(
|
||||
client: AsyncClient,
|
||||
input_text: str,
|
||||
conversation: ConversationManager
|
||||
) -> None:
|
||||
"""Stream the response from the API and update conversation."""
|
||||
input_text = input_text.strip()
|
||||
if not input_text:
|
||||
raise ValueError("Input text cannot be empty")
|
||||
|
||||
# Add user message to conversation
|
||||
conversation.add_message("user", input_text)
|
||||
|
||||
create_args = {
|
||||
"messages": conversation.get_messages(),
|
||||
"stream": True
|
||||
}
|
||||
|
||||
if conversation.model:
|
||||
create_args["model"] = conversation.model
|
||||
|
||||
response_content = []
|
||||
last_chunk = None
|
||||
async for chunk in client.chat.completions.create(**create_args):
|
||||
last_chunk = chunk
|
||||
token = chunk.choices[0].delta.content
|
||||
if not token:
|
||||
continue
|
||||
token = str(token)
|
||||
response_content.append(token)
|
||||
try:
|
||||
for byte in token.encode('utf-8'):
|
||||
sys.stdout.buffer.write(bytes([byte]))
|
||||
sys.stdout.buffer.flush()
|
||||
except (IOError, BrokenPipeError) as e:
|
||||
print(f"\nError writing to stdout: {e}", file=sys.stderr)
|
||||
break
|
||||
|
||||
conversation.conversation = getattr(last_chunk, 'conversation', None)
|
||||
|
||||
if response_content:
|
||||
# Add assistant message to conversation
|
||||
full_response = "".join(response_content).strip()
|
||||
conversation.add_message("assistant", full_response)
|
||||
else:
|
||||
raise RuntimeError("No response received from the API")
|
||||
|
||||
def get_parser():
|
||||
"""Parse command line arguments."""
|
||||
parser = argparse.ArgumentParser(
|
||||
description="GPT CLI client with conversation history",
|
||||
formatter_class=argparse.ArgumentDefaultsHelpFormatter
|
||||
)
|
||||
parser.add_argument("--debug", "-d", action="store_true", help="Enable verbose logging.")
|
||||
parser.add_argument(
|
||||
'-p', '--provider',
|
||||
default=None,
|
||||
help=f"Provider to use. Available: {', '.join([key for key, provider in ProviderUtils.convert.items() if provider.working])}."
|
||||
)
|
||||
parser.add_argument(
|
||||
'-m', '--model',
|
||||
help="Model to use (provider-specific)"
|
||||
)
|
||||
parser.add_argument(
|
||||
'-c', '--cookies-dir',
|
||||
type=Path,
|
||||
default=COOKIES_DIR,
|
||||
help="Directory containing cookies for authenticated providers"
|
||||
)
|
||||
parser.add_argument(
|
||||
'--conversation-file',
|
||||
type=Path,
|
||||
default=CONVERSATION_FILE,
|
||||
help="File to store/load conversation state"
|
||||
)
|
||||
parser.add_argument(
|
||||
'--clear-history',
|
||||
action='store_true',
|
||||
help="Clear conversation history before starting"
|
||||
)
|
||||
parser.add_argument(
|
||||
'input',
|
||||
nargs='?',
|
||||
help="Input text (or read from stdin)"
|
||||
)
|
||||
|
||||
return parser
|
||||
|
||||
async def run_args(input_text: str, args):
|
||||
try:
|
||||
# Ensure directories exist
|
||||
CONFIG_DIR.mkdir(parents=True, exist_ok=True)
|
||||
COOKIES_DIR.mkdir(parents=True, exist_ok=True)
|
||||
|
||||
if args.debug:
|
||||
debug.logging = True
|
||||
|
||||
# Initialize conversation manager
|
||||
conversation = ConversationManager(args.conversation_file, args.model, args.provider)
|
||||
if args.clear_history:
|
||||
conversation.history = []
|
||||
conversation.conversation = None
|
||||
|
||||
# Set cookies directory if specified
|
||||
set_cookies_dir(str(args.cookies_dir))
|
||||
read_cookie_files()
|
||||
|
||||
# Initialize client with selected provider
|
||||
client = AsyncClient(provider=conversation.provider)
|
||||
|
||||
# Stream response and update conversation
|
||||
await stream_response(client, input_text, conversation)
|
||||
|
||||
# Save conversation state
|
||||
conversation.save()
|
||||
|
||||
print() # Ensure final newline
|
||||
|
||||
except Exception as e:
|
||||
print(f"\nError: {e}", file=sys.stderr)
|
||||
sys.exit(1)
|
||||
|
||||
def run_client_args(args):
|
||||
input_text = args.input
|
||||
if not input_text:
|
||||
input_text = sys.stdin.read().strip()
|
||||
if not input_text:
|
||||
print("No input provided. Use -h for help.", file=sys.stderr)
|
||||
sys.exit(1)
|
||||
if os.path.isfile(input_text):
|
||||
with open(input_text, 'r', encoding='utf-8') as f:
|
||||
input_text = f.read().strip()
|
||||
# Run the client with provided arguments
|
||||
asyncio.run(run_args(input_text, args))
|
||||
|
||||
if __name__ == "__main__":
|
||||
# Run the client with command line arguments
|
||||
run_client_args(get_parser().parse_args())
|
||||
69
setup.py
69
setup.py
|
|
@ -104,7 +104,7 @@ setup(
|
|||
long_description=long_description,
|
||||
packages=find_packages(),
|
||||
package_data={
|
||||
'g4f': ['g4f/inference/*', 'g4f/gui/client/*', 'g4f/gui/server/*', 'g4f/Provider/npm/*', 'g4f/local/models/*']
|
||||
'g4f': []
|
||||
},
|
||||
include_package_data=True,
|
||||
install_requires=INSTALL_REQUIRE,
|
||||
|
|
@ -118,29 +118,50 @@ setup(
|
|||
'Bug Tracker': 'https://github.com/xtekky/gpt4free/issues', # Link to issue tracker
|
||||
},
|
||||
keywords=[
|
||||
'python',
|
||||
'chatbot',
|
||||
'reverse-engineering',
|
||||
'openai',
|
||||
'chatbots',
|
||||
'gpt',
|
||||
'language-model',
|
||||
'gpt-3',
|
||||
'gpt3',
|
||||
'openai-api',
|
||||
'gpt-4',
|
||||
'gpt4',
|
||||
'chatgpt',
|
||||
'chatgpt-api',
|
||||
'openai-chatgpt',
|
||||
'chatgpt-free',
|
||||
'chatgpt-4',
|
||||
'chatgpt4',
|
||||
'chatgpt4-api',
|
||||
'free',
|
||||
'free-gpt',
|
||||
'gpt4free',
|
||||
'g4f',
|
||||
"gpt4free",
|
||||
"gpt4free.js",
|
||||
"g4f",
|
||||
"g4f.dev",
|
||||
"javascript",
|
||||
"npm",
|
||||
"browser",
|
||||
"gpt",
|
||||
"chatgpt",
|
||||
"deepseek",
|
||||
"openai",
|
||||
"ai",
|
||||
"client",
|
||||
"sdk",
|
||||
"free",
|
||||
"ai",
|
||||
"gpt-4",
|
||||
"gpt-4o",
|
||||
"chat",
|
||||
"api",
|
||||
"browser",
|
||||
"ai",
|
||||
"ai",
|
||||
"js",
|
||||
"client",
|
||||
"text",
|
||||
"generation",
|
||||
"image",
|
||||
"generation",
|
||||
"in-browser",
|
||||
"ai",
|
||||
"frontend",
|
||||
"ai",
|
||||
"openai",
|
||||
"alternative",
|
||||
"javascript",
|
||||
"ai",
|
||||
"library",
|
||||
"nodejs",
|
||||
"prompt",
|
||||
"engineering",
|
||||
"chatbot",
|
||||
"ai",
|
||||
"integration"
|
||||
],
|
||||
classifiers=[
|
||||
'Development Status :: 2 - Pre-Alpha',
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue