gpt4free/g4f/Provider/needs_auth/GithubCopilot.py

191 lines
6.4 KiB
Python

from __future__ import annotations
import json
from aiohttp import ClientSession
from ..base_provider import AsyncGeneratorProvider, ProviderModelMixin, BaseConversation
from ...typing import AsyncResult, Messages, Cookies
from ...requests.raise_for_status import raise_for_status
from ...requests.aiohttp import get_connector
from ...providers.helper import format_prompt, get_last_user_message
from ...cookies import get_cookies
class Conversation(BaseConversation):
conversation_id: str
def __init__(self, conversation_id: str):
self.conversation_id = conversation_id
class GithubCopilot(AsyncGeneratorProvider, ProviderModelMixin):
label = "GitHub Copilot"
url = "https://github.com/copilot"
working = True
needs_auth = True
supports_stream = True
default_model = "gpt-4.1"
models = [
# GPT-5 Series
"gpt-5",
"gpt-5-mini",
"gpt-5.1",
"gpt-5.2",
# GPT-5 Codex (optimized for code)
"gpt-5-codex",
"gpt-5.1-codex",
"gpt-5.1-codex-mini",
"gpt-5.1-codex-max",
"gpt-5.2-codex",
"gpt-5.3-codex",
# GPT-4 Series
"gpt-4.1",
"gpt-4.1-2025-04-14",
"gpt-4o",
"gpt-4o-mini",
"gpt-4o-2024-11-20",
"gpt-4o-2024-08-06",
"gpt-4o-2024-05-13",
"gpt-4o-mini-2024-07-18",
"gpt-4",
"gpt-4-0613",
"gpt-4-0125-preview",
"gpt-4-o-preview",
# Claude 4 Series
"claude-opus-4.6",
"claude-opus-4.6-fast",
"claude-opus-4.5",
"claude-sonnet-4.5",
"claude-sonnet-4",
"claude-haiku-4.5",
# Gemini Series
"gemini-3-pro-preview",
"gemini-3-flash-preview",
"gemini-2.5-pro",
# Grok
"grok-code-fast-1",
# Legacy GPT-3.5
"gpt-3.5-turbo",
"gpt-3.5-turbo-0613",
# Embeddings
"text-embedding-3-small",
"text-embedding-ada-002",
]
@classmethod
async def create_async_generator(
cls,
model: str,
messages: Messages,
stream: bool = True,
api_key: str = None,
proxy: str = None,
cookies: Cookies = None,
conversation_id: str = None,
conversation: Conversation = None,
return_conversation: bool = True,
**kwargs
) -> AsyncResult:
if not model:
model = cls.default_model
if cookies is None:
cookies = get_cookies("github.com")
async with ClientSession(
connector=get_connector(proxy=proxy),
cookies=cookies,
headers={
'User-Agent': 'Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:133.0) Gecko/20100101 Firefox/133.0',
'Accept': 'application/json',
'Accept-Language': 'en-US,en;q=0.5',
'Referer': 'https://github.com/copilot',
'Content-Type': 'application/json',
'GitHub-Verified-Fetch': 'true',
'X-Requested-With': 'XMLHttpRequest',
'Origin': 'https://github.com',
'Connection': 'keep-alive',
'Sec-Fetch-Dest': 'empty',
'Sec-Fetch-Mode': 'cors',
'Sec-Fetch-Site': 'same-origin',
'Priority': 'u=1'
}
) as session:
headers = {}
if api_key is None:
async with session.post("https://github.com/github-copilot/chat/token") as response:
await raise_for_status(response, "Get token")
api_key = (await response.json()).get("token")
headers = {
"Authorization": f"GitHub-Bearer {api_key}",
}
if conversation is not None:
conversation_id = conversation.conversation_id
if conversation_id is None:
async with session.post(
"https://api.individual.githubcopilot.com/github/chat/threads",
headers=headers
) as response:
await raise_for_status(response)
conversation_id = (await response.json()).get("thread_id")
if return_conversation:
yield Conversation(conversation_id)
content = get_last_user_message(messages)
else:
content = format_prompt(messages)
json_data = {
"content": content,
"intent": "conversation",
"references": [],
"context": [],
"currentURL": f"https://github.com/copilot/c/{conversation_id}",
"streaming": stream,
"confirmations": [],
"customInstructions": [],
"model": model,
"mode": "immersive"
}
async with session.post(
f"https://api.individual.githubcopilot.com/github/chat/threads/{conversation_id}/messages",
json=json_data,
headers=headers
) as response:
await raise_for_status(response, f"Send message with model {model}")
if stream:
async for line in response.content:
if line.startswith(b"data: "):
try:
data = json.loads(line[6:])
if data.get("type") == "content":
content = data.get("body", "")
if content:
yield content
except json.JSONDecodeError:
continue
else:
full_content = ""
async for line in response.content:
if line.startswith(b"data: "):
try:
data = json.loads(line[6:])
if data.get("type") == "content":
full_content += data.get("body", "")
except json.JSONDecodeError:
continue
if full_content:
yield full_content