Merge pull request #2540 from hlohaus/sun

Add Edge as Browser for nodriver
This commit is contained in:
H Lohaus 2025-01-05 11:44:11 -08:00 committed by GitHub
commit b35240c29c
No known key found for this signature in database
GPG key ID: B5690EEEBB952194
21 changed files with 386 additions and 159 deletions

View file

@ -78,7 +78,7 @@ RUN pip install --break-system-packages --upgrade pip \
&& pip install --break-system-packages \
undetected-chromedriver selenium-wire \
&& pip uninstall -y --break-system-packages \
pywebview plyer
pywebview
# Copy the entire package into the container.
ADD --chown=$G4F_USER:$G4F_USER g4f $G4F_DIR/g4f

View file

@ -2,12 +2,14 @@ from __future__ import annotations
import asyncio
import json
from pathlib import Path
from ..typing import AsyncResult, Messages, Cookies
from .base_provider import AsyncGeneratorProvider, ProviderModelMixin, get_running_loop
from ..requests import Session, StreamSession, get_args_from_nodriver, raise_for_status, merge_cookies
from ..requests import DEFAULT_HEADERS, has_nodriver, has_curl_cffi
from ..providers.response import FinishReason
from ..cookies import get_cookies_dir
from ..errors import ResponseStatusError, ModelNotFoundError
class Cloudflare(AsyncGeneratorProvider, ProviderModelMixin):
@ -19,7 +21,7 @@ class Cloudflare(AsyncGeneratorProvider, ProviderModelMixin):
supports_stream = True
supports_system_message = True
supports_message_history = True
default_model = "@cf/meta/llama-3.1-8b-instruct"
default_model = "@cf/meta/llama-3.3-70b-instruct-fp8-fast"
model_aliases = {
"llama-2-7b": "@cf/meta/llama-2-7b-chat-fp16",
"llama-2-7b": "@cf/meta/llama-2-7b-chat-int8",
@ -33,6 +35,10 @@ class Cloudflare(AsyncGeneratorProvider, ProviderModelMixin):
}
_args: dict = None
@classmethod
def get_cache_file(cls) -> Path:
return Path(get_cookies_dir()) / f"auth_{cls.parent if hasattr(cls, 'parent') else cls.__name__}.json"
@classmethod
def get_models(cls) -> str:
if not cls.models:
@ -67,7 +73,11 @@ class Cloudflare(AsyncGeneratorProvider, ProviderModelMixin):
timeout: int = 300,
**kwargs
) -> AsyncResult:
cache_file = cls.get_cache_file()
if cls._args is None:
if cache_file.exists():
with cache_file.open("r") as f:
cls._args = json.load(f)
if has_nodriver:
cls._args = await get_args_from_nodriver(cls.url, proxy, timeout, cookies)
else:
@ -93,6 +103,8 @@ class Cloudflare(AsyncGeneratorProvider, ProviderModelMixin):
await raise_for_status(response)
except ResponseStatusError:
cls._args = None
if cache_file.exists():
cache_file.unlink()
raise
reason = None
async for line in response.iter_lines():
@ -110,3 +122,6 @@ class Cloudflare(AsyncGeneratorProvider, ProviderModelMixin):
continue
if reason is not None:
yield FinishReason(reason)
with cache_file.open("w") as f:
json.dump(cls._args, f)

View file

@ -1,14 +1,18 @@
from __future__ import annotations
from aiohttp import ClientSession, ClientTimeout, ClientError
import asyncio
from aiohttp import ClientSession, ClientTimeout, ClientError, ClientResponseError
import json
from ..typing import AsyncResult, Messages
from .base_provider import AsyncGeneratorProvider, ProviderModelMixin, BaseConversation
from .helper import format_prompt
from ..providers.response import FinishReason
from .. import debug
class Conversation(BaseConversation):
vqd: str = None
message_history: Messages = []
cookies: dict = {}
def __init__(self, model: str):
self.model = model
@ -65,19 +69,23 @@ class DDG(AsyncGeneratorProvider, ProviderModelMixin):
conversation: Conversation = None,
return_conversation: bool = False,
proxy: str = None,
headers: dict = {
"Content-Type": "application/json",
},
cookies: dict = None,
max_retries: int = 3,
**kwargs
) -> AsyncResult:
headers = {
"Content-Type": "application/json",
}
async with ClientSession(headers=headers, timeout=ClientTimeout(total=30)) as session:
if cookies is None and conversation is not None:
cookies = conversation.cookies
async with ClientSession(headers=headers, cookies=cookies, timeout=ClientTimeout(total=30)) as session:
# Fetch VQD token
if conversation is None:
conversation = Conversation(model)
if conversation.vqd is None:
conversation.cookies = session.cookie_jar
conversation.vqd = await cls.fetch_vqd(session)
if conversation.vqd is not None:
headers["x-vqd-4"] = conversation.vqd
if return_conversation:
@ -97,15 +105,33 @@ class DDG(AsyncGeneratorProvider, ProviderModelMixin):
async with session.post(cls.api_endpoint, headers=headers, json=payload, proxy=proxy) as response:
conversation.vqd = response.headers.get("x-vqd-4")
response.raise_for_status()
reason = None
async for line in response.content:
line = line.decode("utf-8").strip()
if line.startswith("data:"):
try:
message = json.loads(line[5:].strip())
if "message" in message:
if "message" in message and message["message"]:
yield message["message"]
reason = "max_tokens"
elif message.get("message") == '':
reason = "stop"
except json.JSONDecodeError:
continue
if reason is not None:
yield FinishReason(reason)
except ClientResponseError as e:
if e.code in (400, 429) and max_retries > 0:
debug.log(f"Retry: max_retries={max_retries}, wait={512 - max_retries * 48}: {e}")
await asyncio.sleep(512 - max_retries * 48)
is_started = False
async for chunk in cls.create_async_generator(model, messages, conversation, return_conversation, max_retries=max_retries-1, **kwargs):
if chunk:
yield chunk
is_started = True
if is_started:
return
raise e
except ClientError as e:
raise Exception(f"HTTP ClientError occurred: {e}")
except asyncio.TimeoutError:

View file

@ -137,7 +137,7 @@ class HuggingFace(AsyncGeneratorProvider, ProviderModelMixin):
else:
is_special = True
debug.log(f"Special token: {is_special}")
yield FinishReason("stop" if is_special else "max_tokens", actions=["variant"] if is_special else ["continue", "variant"])
yield FinishReason("stop" if is_special else "length", actions=["variant"] if is_special else ["continue", "variant"])
else:
if response.headers["content-type"].startswith("image/"):
base64_data = base64.b64encode(b"".join([chunk async for chunk in response.iter_content()]))

View file

@ -105,11 +105,11 @@ class OpenaiChat(AsyncAuthedProvider, ProviderModelMixin):
_expires: int = None
@classmethod
async def on_auth_async(cls, **kwargs) -> AuthResult:
async def on_auth_async(cls, **kwargs) -> AsyncIterator:
if cls.needs_auth:
async for _ in cls.login():
pass
return AuthResult(
async for chunk in cls.login():
yield chunk
yield AuthResult(
api_key=cls._api_key,
cookies=cls._cookies or RequestConfig.cookies or {},
headers=cls._headers or RequestConfig.headers or cls.get_default_headers(),
@ -174,7 +174,8 @@ class OpenaiChat(AsyncAuthedProvider, ProviderModelMixin):
"use_case": "multimodal"
}
# Post the image data to the service and get the image data
async with session.post(f"{cls.url}/backend-api/files", json=data, headers=auth_result.headers) as response:
headers = auth_result.headers if hasattr(auth_result, "headers") else None
async with session.post(f"{cls.url}/backend-api/files", json=data, headers=headers) as response:
cls._update_request_args(auth_result, session)
await raise_for_status(response, "Create file failed")
image_data = {
@ -360,7 +361,7 @@ class OpenaiChat(AsyncAuthedProvider, ProviderModelMixin):
f"{cls.url}/backend-anon/sentinel/chat-requirements"
if cls._api_key is None else
f"{cls.url}/backend-api/sentinel/chat-requirements",
json={"p": None if auth_result.proof_token is None else get_requirements_token(auth_result.proof_token)},
json={"p": None if not getattr(auth_result, "proof_token") else get_requirements_token(auth_result.proof_token)},
headers=cls._headers
) as response:
if response.status == 401:
@ -386,7 +387,7 @@ class OpenaiChat(AsyncAuthedProvider, ProviderModelMixin):
proofofwork = generate_proof_token(
**chat_requirements["proofofwork"],
user_agent=auth_result.headers.get("user-agent"),
proof_token=auth_result.proof_token
proof_token=getattr(auth_result, "proof_token")
)
[debug.log(text) for text in (
#f"Arkose: {'False' if not need_arkose else auth_result.arkose_token[:12]+'...'}",

View file

@ -41,7 +41,7 @@ from g4f.errors import ProviderNotFoundError, ModelNotFoundError, MissingAuthErr
from g4f.cookies import read_cookie_files, get_cookies_dir
from g4f.Provider import ProviderType, ProviderUtils, __providers__
from g4f.gui import get_gui_app
from g4f.tools.files import supports_filename, get_streaming
from g4f.tools.files import supports_filename, get_async_streaming
from .stubs import (
ChatCompletionsConfig, ImageGenerationConfig,
ProviderResponseModel, ModelResponseModel,
@ -436,7 +436,8 @@ class Api:
event_stream = "text/event-stream" in request.headers.get("accept", "")
if not os.path.isdir(bucket_dir):
return ErrorResponse.from_message("Bucket dir not found", 404)
return StreamingResponse(get_streaming(bucket_dir, delete_files, refine_chunks_with_spacy, event_stream), media_type="text/plain")
return StreamingResponse(get_async_streaming(bucket_dir, delete_files, refine_chunks_with_spacy, event_stream),
media_type="text/event-stream" if event_stream else "text/plain")
@self.app.post("/v1/files/{bucket_id}", responses={
HTTP_200_OK: {"model": UploadResponseModel}

View file

@ -103,17 +103,29 @@
z-index: -1;
}
iframe.stream {
.stream-widget {
max-height: 0;
transition: max-height 0.15s ease-out;
color: var(--colour-5);
overflow: scroll;
text-align: left;
}
iframe.stream.show {
.stream-widget.show {
max-height: 1000px;
height: 1000px;
transition: max-height 0.25s ease-in;
background: rgba(255,255,255,0.7);
border-top: 2px solid rgba(255,255,255,0.5);
padding: 20px;
}
.stream-widget img {
max-width: 320px;
}
#stream-container {
width: 100%;
}
.description {
@ -207,32 +219,87 @@
<p>Powered by the G4F framework</p>
</div>
<iframe id="stream-widget" class="stream" frameborder="0"></iframe>
<iframe class="stream-widget" frameborder="0"></iframe>
</div>
<script>
const iframe = document.getElementById('stream-widget');""
let search = (navigator.language == "de" ? "news in deutschland" : navigator.language == "en" ? "world news" : navigator.language);
if (Math.floor(Math.random() * 6) % 2 == 0) {
const iframe = document.querySelector('.stream-widget');
const rand_idx = Math.floor(Math.random() * 9)
if (rand_idx < 3) {
search = "xtekky/gpt4free releases";
} else if (rand_idx < 5) {
search = "developer news";
} else {
search = (navigator.language == "de" ? "news in deutsch" : navigator.language == "en" ? "world news" : `news in ${navigator.language}`);
}
const url = "/backend-api/v2/create?prompt=Create of overview of the news in plain text&stream=1&web_search=" + search;
const summary_prompt = "Give a summary of the provided text in ```markdown``` format. Add maybe one or more images.";
const url = `/backend-api/v2/create?prompt=${summary_prompt}&stream=1&web_search=${search}`;
iframe.src = url;
setTimeout(()=>iframe.classList.add('show'), 3000);
const message = "Loading...";
setTimeout(()=>{
iframe.classList.add('show');
const iframeDocument = iframe.contentDocument || iframe.contentWindow?.document;
if (iframeDocument) {
const iframeBody = iframeDocument.querySelector("body");
if (iframeBody) {
iframeBody.innerHTML = message + iframeBody.innerHTML;
}
} else {
iframe.parentElement.removeChild(iframe);
}
}, 1000);
function filterMarkdown(text, allowedTypes = null, defaultValue = null) {
const match = text.match(/```(.+)\n(?<code>[\s\S]+?)(\n```|$)/);
if (match) {
const [, type, code] = match;
if (!allowedTypes || allowedTypes.includes(type)) {
return code;
}
}
return defaultValue;
}
let scroll_to_bottom_callback = () => {
const i = document.querySelector(".stream-widget");
if (!i.contentWindow || !i.contentDocument) {
return;
}
clientHeight = i.contentDocument.body.scrollHeight;
i.contentWindow.scrollTo(0, clientHeight);
if (clientHeight - i.contentWindow.scrollY < 2 * clientHeight) {
setTimeout(scroll_to_bottom_callback, 1000);
}
};
setTimeout(scroll_to_bottom_callback, 1000);
iframe.onload = () => {
const iframeDocument = iframe.contentDocument || iframe.contentWindow.document;
const iframeBody = iframeDocument.querySelector("body");
const iframeContent = iframeDocument.querySelector("pre");
let iframeText = iframeContent.innerHTML;
const markdown = window.markdownit();
iframeBody.innerHTML = markdown.render(iframeContent.innerHTML);
const iframeContainer = document.querySelector(".container");
iframe.remove()
if (iframeText.indexOf('"error"') < 0) {
iframeContainer.innerHTML += `<div class="stream-widget show">${markdown.render(filterMarkdown(iframeText, "markdown", iframeText))}</div>`;
}
scroll_to_bottom_callback = () => null;
}
(async () => {
const prompt = `
const today = new Date().toJSON().slice(0, 10);
const max = 100;
const cache_id = Math.floor(Math.random() * max);
let prompt;
if (cache_id % 2 == 0) {
prompt = `
Today is ${new Date().toJSON().slice(0, 10)}.
Create a single-page HTML screensaver reflecting the current season (based on the date).
For example, if it's Spring, it might use floral patterns or pastel colors.
Avoid using any text. Consider a subtle animation or transition effect.`;
const response = await fetch(`/backend-api/v2/create?prompt=${prompt}&filter_markdown=html`)
Avoid using any text.`;
} else {
prompt = `Create a single-page HTML screensaver. Avoid using any text.`;
const response = await fetch(`/backend-api/v2/create?prompt=${prompt}&filter_markdown=html&cache=${cache_id}`);
}
const response = await fetch(`/backend-api/v2/create?prompt=${prompt}&filter_markdown=html&cache=${cache_id}`);
const text = await response.text()
background.src = `data:text/html;charset=utf-8,${encodeURIComponent(text)}`;
const gradient = document.querySelector('.gradient');

View file

@ -239,7 +239,8 @@
<button class="hide-input">
<i class="fa-solid fa-angles-down"></i>
</button>
<span class="text"></span>
<input type="checkbox" id="agree" name="agree" value="yes" checked>
<label for="agree" class="text" onclick="this.innerText='';">Scroll to bottom</label>
</div>
<div class="stop_generating stop_generating-hidden">
<button id="cancelButton">

View file

@ -516,7 +516,11 @@ body:not(.white) a:visited{
padding: 6px 6px;
}
#input-count .text {
input-count .text {
min-width: 12px
}
#input-count .text, #input-count input {
padding: 0 4px;
}
@ -793,7 +797,7 @@ select {
appearance: none;
width: 100%;
height: 20px;
background: var(--accent);
background: var(--colour-2);
outline: none;
transition: opacity .2s;
border-radius: 10px;
@ -859,11 +863,18 @@ select:hover,
font-size: 15px;
width: 100%;
color: var(--colour-3);
min-height: 49px;
height: 59px;
outline: none;
padding: var(--inner-gap) var(--section-gap);
resize: vertical;
min-height: 59px;
transition: max-height 0.15s ease-out;
}
#systemPrompt:focus {
min-height: 200px;
max-height: 1000px;
transition: max-height 0.25s ease-in;
}
.pswp {
@ -929,6 +940,9 @@ select:hover,
body:not(.white) .gradient{
display: block;
}
.settings .label, form .label, .settings label, form label {
min-width: 200px;
}
}
.input-box {
@ -1112,8 +1126,12 @@ ul {
display: flex;
}
#systemPrompt {
padding-left: 48px;
#systemPrompt::placeholder {
text-align: center;
}
.settings h3 {
text-align: center;
}
}
@ -1354,7 +1372,6 @@ form .field.saved .fa-xmark {
.settings .label, form .label, .settings label, form label {
font-size: 15px;
margin-left: var(--inner-gap);
min-width: 200px;
}
.settings .label, form .label {

View file

@ -511,7 +511,9 @@ const prepare_messages = (messages, message_index = -1, do_continue = false) =>
// Include only not regenerated messages
if (new_message && !new_message.regenerate) {
// Remove generated images from history
if (new_message.content) {
new_message.content = filter_message(new_message.content);
}
// Remove internal fields
delete new_message.provider;
delete new_message.synthesize;
@ -658,7 +660,7 @@ async function load_provider_parameters(provider) {
}
}
async function add_message_chunk(message, message_id, provider) {
async function add_message_chunk(message, message_id, provider, scroll) {
content_map = content_storage[message_id];
if (message.type == "conversation") {
const conversation = await get_conversation(window.conversation_id);
@ -698,7 +700,7 @@ async function add_message_chunk(message, message_id, provider) {
content_map.inner.innerHTML = markdown_render(message.preview);
} else if (message.type == "content") {
message_storage[message_id] += message.content;
update_message(content_map, message_id);
update_message(content_map, message_id, null, scroll);
content_map.inner.style.height = "";
} else if (message.type == "log") {
let p = document.createElement("p");
@ -709,9 +711,7 @@ async function add_message_chunk(message, message_id, provider) {
} else if (message.type == "title") {
title_storage[message_id] = message.title;
} else if (message.type == "login") {
update_message(content_map, message_id, message.login);
} else if (message.type == "login") {
update_message(content_map, message_id, message.login);
update_message(content_map, message_id, message.login, scroll);
} else if (message.type == "finish") {
finish_storage[message_id] = message.finish;
} else if (message.type == "parameters") {
@ -734,8 +734,12 @@ const ask_gpt = async (message_id, message_index = -1, regenerate = false, provi
messages = prepare_messages(conversation.items, message_index, action=="continue");
message_storage[message_id] = "";
stop_generating.classList.remove("stop_generating-hidden");
let scroll = true;
if (message_index > 0 && parseInt(message_index, 10) + 1 < conversation.items.length) {
scroll = false;
}
if (message_index == -1) {
if (scroll) {
await lazy_scroll_to_bottom();
}
@ -780,7 +784,7 @@ const ask_gpt = async (message_id, message_index = -1, regenerate = false, provi
update_timeouts: [],
message_index: message_index,
}
if (message_index == -1) {
if (scroll) {
await lazy_scroll_to_bottom();
}
try {
@ -801,7 +805,7 @@ const ask_gpt = async (message_id, message_index = -1, regenerate = false, provi
download_images: download_images,
api_key: api_key,
ignored: ignored,
}, files, message_id);
}, files, message_id, scroll);
content_map.update_timeouts.forEach((timeoutId)=>clearTimeout(timeoutId));
content_map.update_timeouts = [];
if (!error_storage[message_id]) {
@ -836,12 +840,12 @@ const ask_gpt = async (message_id, message_index = -1, regenerate = false, provi
);
delete message_storage[message_id];
if (!error_storage[message_id]) {
await safe_load_conversation(window.conversation_id, message_index == -1);
await safe_load_conversation(window.conversation_id, scroll);
}
}
let cursorDiv = message_el.querySelector(".cursor");
if (cursorDiv) cursorDiv.parentNode.removeChild(cursorDiv);
if (message_index == -1) {
if (scroll) {
await lazy_scroll_to_bottom();
}
await safe_remove_cancel_button();
@ -856,7 +860,7 @@ async function scroll_to_bottom() {
}
async function lazy_scroll_to_bottom() {
if (message_box.scrollHeight - message_box.scrollTop < 2 * message_box.clientHeight) {
if (document.querySelector("#input-count input").checked) {
await scroll_to_bottom();
}
}
@ -1013,6 +1017,8 @@ const load_conversation = async (conversation_id, scroll=true) => {
if (newContent.startsWith("```")) {
const index = str.indexOf("\n");
newContent = newContent.substring(index);
} else if (newContent.startsWith("...")) {
newContent = " " + newContent.substring(3);
}
if (newContent.startsWith(lastLine)) {
newContent = newContent.substring(lastLine.length);
@ -1054,7 +1060,7 @@ const load_conversation = async (conversation_id, scroll=true) => {
if (item.finish && item.finish.actions) {
actions = item.finish.actions
}
if (!("continue" in actions)) {
if (item.role == "assistant" && !actions.includes("continue")) {
let reason = "stop";
// Read finish reason from conversation
if (item.finish && item.finish.reason) {
@ -1067,7 +1073,7 @@ const load_conversation = async (conversation_id, scroll=true) => {
reason = "error";
// Has an even number of start or end code tags
} else if (buffer.split("```").length - 1 % 2 === 1) {
reason = "error";
reason = "length";
// Has a end token at the end
} else if (lastLine.endsWith("```") || lastLine.endsWith(".") || lastLine.endsWith("?") || lastLine.endsWith("!")
|| lastLine.endsWith('"') || lastLine.endsWith("'") || lastLine.endsWith(")")
@ -1152,7 +1158,7 @@ const load_conversation = async (conversation_id, scroll=true) => {
highlight(message_box);
regenerate_button.classList.remove("regenerate-hidden");
if (scroll) {
if (scroll && document.querySelector("#input-count input").checked) {
message_box.scrollTo({ top: message_box.scrollHeight, behavior: "smooth" });
setTimeout(() => {
@ -1517,7 +1523,7 @@ function count_words_and_tokens(text, model) {
return `(${count_words(text)} words, ${count_chars(text)} chars, ${count_tokens(model, text)} tokens)`;
}
function update_message(content_map, message_id, content = null) {
function update_message(content_map, message_id, content = null, scroll = true) {
content_map.update_timeouts.push(setTimeout(() => {
if (!content) content = message_storage[message_id];
html = markdown_render(content);
@ -1538,7 +1544,7 @@ function update_message(content_map, message_id, content = null) {
content_map.inner.innerHTML = html;
content_map.count.innerText = count_words_and_tokens(message_storage[message_id], provider_storage[message_id]?.model);
highlight(content_map.inner);
if (content_map.message_index == -1) {
if (scroll) {
lazy_scroll_to_bottom();
}
content_map.update_timeouts.forEach((timeoutId)=>clearTimeout(timeoutId));
@ -1890,7 +1896,7 @@ fileInput.addEventListener('change', async (event) => {
fileInput.value = "";
inputCount.innerText = `${count} Conversations were imported successfully`;
} else {
is_cookie_file = false;
is_cookie_file = data.api_key;
if (Array.isArray(data)) {
data.forEach((item) => {
if (item.domain && item.name && item.value) {
@ -1927,7 +1933,7 @@ function get_selected_model() {
}
}
async function api(ressource, args=null, files=null, message_id=null) {
async function api(ressource, args=null, files=null, message_id=null, scroll=true) {
let api_key;
if (ressource == "models" && args) {
api_key = get_api_key_by_provider(args);
@ -1957,7 +1963,7 @@ async function api(ressource, args=null, files=null, message_id=null) {
headers: headers,
body: body,
});
return read_response(response, message_id, args.provider || null);
return read_response(response, message_id, args.provider || null, scroll);
}
response = await fetch(url, {headers: headers});
if (response.status == 200) {
@ -1966,7 +1972,7 @@ async function api(ressource, args=null, files=null, message_id=null) {
console.error(response);
}
async function read_response(response, message_id, provider) {
async function read_response(response, message_id, provider, scroll) {
const reader = response.body.pipeThrough(new TextDecoderStream()).getReader();
let buffer = ""
while (true) {
@ -1979,7 +1985,7 @@ async function read_response(response, message_id, provider) {
continue;
}
try {
add_message_chunk(JSON.parse(buffer + line), message_id, provider);
add_message_chunk(JSON.parse(buffer + line), message_id, provider, scroll);
buffer = "";
} catch {
buffer += line
@ -2106,6 +2112,7 @@ if (SpeechRecognition) {
recognition.maxAlternatives = 1;
let startValue;
let buffer;
let lastDebounceTranscript;
recognition.onstart = function() {
microLabel.classList.add("recognition");
@ -2114,6 +2121,7 @@ if (SpeechRecognition) {
messageInput.readOnly = true;
};
recognition.onend = function() {
messageInput.value = `${startValue ? startValue + "\n" : ""}${buffer}`;
messageInput.readOnly = false;
messageInput.focus();
};
@ -2131,18 +2139,17 @@ if (SpeechRecognition) {
lastDebounceTranscript = transcript;
}
if (transcript) {
messageInput.value = `${startValue ? startValue+"\n" : ""}${transcript.trim()}`;
inputCount.innerText = transcript;
if (isFinal) {
startValue = messageInput.value;
buffer = `${buffer ? buffer + "\n" : ""}${transcript.trim()}`;
}
messageInput.style.height = messageInput.scrollHeight + "px";
messageInput.scrollTop = messageInput.scrollHeight;
}
};
microLabel.addEventListener("click", (e) => {
if (microLabel.classList.contains("recognition")) {
recognition.stop();
messageInput.value = `${startValue ? startValue + "\n" : ""}${buffer}`;
microLabel.classList.remove("recognition");
} else {
const lang = document.getElementById("recognition-language")?.value;

View file

@ -9,6 +9,8 @@ import shutil
from flask import Flask, Response, request, jsonify
from typing import Generator
from pathlib import Path
from urllib.parse import quote_plus
from hashlib import sha256
from werkzeug.utils import secure_filename
from ...image import is_allowed_extension, to_image
@ -123,15 +125,30 @@ class Backend_Api(Api):
"type": "function"
})
do_filter_markdown = request.args.get("filter_markdown")
response = iter_run_tools(
ChatCompletion.create,
model=request.args.get("model"),
messages=[{"role": "user", "content": request.args.get("prompt")}],
provider=request.args.get("provider", None),
stream=not do_filter_markdown,
ignore_stream=not request.args.get("stream"),
tool_calls=tool_calls,
)
cache_id = request.args.get('cache')
parameters = {
"model": request.args.get("model"),
"messages": [{"role": "user", "content": request.args.get("prompt")}],
"provider": request.args.get("provider", None),
"stream": not do_filter_markdown and not cache_id,
"ignore_stream": not request.args.get("stream"),
"tool_calls": tool_calls,
}
if cache_id:
cache_id = sha256(cache_id.encode() + json.dumps(parameters, sort_keys=True).encode()).hexdigest()
cache_dir = Path(get_cookies_dir()) / ".scrape_cache" / "create"
cache_file = cache_dir / f"{quote_plus(request.args.get('prompt').strip()[:20])}.{cache_id}.txt"
if cache_file.exists():
with cache_file.open("r") as f:
response = f.read()
else:
response = iter_run_tools(ChatCompletion.create, **parameters)
cache_dir.mkdir(parents=True, exist_ok=True)
with cache_file.open("w") as f:
f.write(response)
else:
response = iter_run_tools(ChatCompletion.create, **parameters)
if do_filter_markdown:
return Response(filter_markdown(response, do_filter_markdown), mimetype='text/plain')
def cast_str():

View file

@ -38,8 +38,7 @@ def run_webview(
storage_path=storage_path,
debug=debug,
http_port=http_port,
ssl=ssl,
gui=gui
ssl=ssl
)
if __name__ == "__main__":
@ -47,4 +46,4 @@ if __name__ == "__main__":
args = parser.parse_args()
if args.debug:
g4f.debug.logging = True
run_webview(args.debug, args.port)
run_webview(args.debug, args.port, not args.debug)

View file

@ -213,8 +213,7 @@ mistral_nemo = Model(
mistral_large = Model(
name = "mistral-large",
base_provider = "Mistral",
best_provider = PollinationsAI
base_provider = "Mistral"
)
### NousResearch ###
@ -348,7 +347,6 @@ command_r_plus = Model(
command_r = Model(
name = 'command-r',
base_provider = 'CohereForAI',
best_provider = PollinationsAI
)
### Qwen ###

View file

@ -269,7 +269,7 @@ class AsyncProvider(AbstractProvider):
def get_async_create_function(cls) -> callable:
return cls.create_async
class AsyncGeneratorProvider(AsyncProvider):
class AsyncGeneratorProvider(AbstractProvider):
"""
Provides asynchronous generator functionality for streaming results.
"""
@ -395,6 +395,10 @@ class AsyncAuthedProvider(AsyncGeneratorProvider):
def get_async_create_function(cls) -> callable:
return cls.create_async_generator
@classmethod
def get_cache_file(cls) -> Path:
return Path(get_cookies_dir()) / f"auth_{cls.parent if hasattr(cls, 'parent') else cls.__name__}.json"
@classmethod
def create_completion(
cls,
@ -404,18 +408,24 @@ class AsyncAuthedProvider(AsyncGeneratorProvider):
) -> CreateResult:
try:
auth_result = AuthResult()
cache_file = Path(get_cookies_dir()) / f"auth_{cls.parent if hasattr(cls, 'parent') else cls.__name__}.json"
cache_file = cls.get_cache_file()
if cache_file.exists():
with cache_file.open("r") as f:
auth_result = AuthResult(**json.load(f))
else:
auth_result = cls.on_auth(**kwargs)
return to_sync_generator(cls.create_authed(model, messages, auth_result, **kwargs))
if hasattr(auth_result, "_iter__"):
for chunk in auth_result:
if isinstance(chunk, AsyncResult):
auth_result = chunk
else:
yield chunk
yield from to_sync_generator(cls.create_authed(model, messages, auth_result, **kwargs))
except (MissingAuthError, NoValidHarFileError):
if cache_file.exists():
cache_file.unlink()
auth_result = cls.on_auth(**kwargs)
return to_sync_generator(cls.create_authed(model, messages, auth_result, **kwargs))
yield from to_sync_generator(cls.create_authed(model, messages, auth_result, **kwargs))
finally:
cache_file.parent.mkdir(parents=True, exist_ok=True)
cache_file.write_text(json.dumps(auth_result.get_dict()))
@ -434,6 +444,12 @@ class AsyncAuthedProvider(AsyncGeneratorProvider):
auth_result = AuthResult(**json.load(f))
else:
auth_result = await cls.on_auth_async(**kwargs)
if hasattr(auth_result, "_aiter__"):
async for chunk in auth_result:
if isinstance(chunk, AsyncResult):
auth_result = chunk
else:
yield chunk
response = to_async_iterator(cls.create_authed(model, messages, **kwargs, auth_result=auth_result))
async for chunk in response:
yield chunk

View file

@ -19,7 +19,9 @@ def quote_url(url: str) -> str:
def quote_title(title: str) -> str:
if title:
return title.replace("\n", "").replace('"', '')
title = title.strip()
title = " ".join(title.split())
return title.replace('[', '').replace(']', '')
return ""
def format_link(url: str, title: str = None) -> str:

View file

@ -58,10 +58,11 @@ class IterListProvider(BaseRetryProvider):
for chunk in response:
if chunk:
yield chunk
if isinstance(chunk, str):
started = True
elif response:
yield response
started = True
return
if started:
return
except Exception as e:
@ -93,6 +94,7 @@ class IterListProvider(BaseRetryProvider):
async for chunk in response:
if chunk:
yield chunk
if isinstance(chunk, str):
started = True
elif response:
response = await response

View file

@ -1,5 +1,6 @@
from __future__ import annotations
import os
from urllib.parse import urlparse
from typing import Iterator
from http.cookies import Morsel
@ -20,6 +21,7 @@ except ImportError:
try:
import nodriver
from nodriver.cdp.network import CookieParam
from nodriver.core.config import find_chrome_executable
from nodriver import Browser
has_nodriver = True
except ImportError:
@ -95,6 +97,8 @@ async def get_args_from_nodriver(
cookies[c.name] = c.value
user_agent = await page.evaluate("window.navigator.userAgent")
await page.wait_for("body:not(.no-js)", timeout=timeout)
for c in await page.send(nodriver.cdp.network.get_cookies([url])):
cookies[c.name] = c.value
await page.close()
browser.stop()
return {
@ -114,13 +118,21 @@ def merge_cookies(cookies: Iterator[Morsel], response: Response) -> Cookies:
for cookie in response.cookies.jar:
cookies[cookie.name] = cookie.value
async def get_nodriver(proxy: str = None, user_data_dir = "nodriver", **kwargs)-> Browser:
async def get_nodriver(proxy: str = None, user_data_dir = "nodriver", browser_executable_path=None, **kwargs)-> Browser:
if not has_nodriver:
raise MissingRequirementsError('Install "nodriver" package | pip install -U nodriver')
user_data_dir = user_config_dir(f"g4f-{user_data_dir}") if has_platformdirs else None
if browser_executable_path is None:
try:
browser_executable_path = find_chrome_executable()
except FileNotFoundError:
# Default to Edge if Chrome is not found
if os.path.exists("C:\Program Files (x86)\Microsoft\Edge\Application\msedge.exe"):
browser_executable_path = "C:\Program Files (x86)\Microsoft\Edge\Application\msedge.exe"
debug.log(f"Open nodriver with user_dir: {user_data_dir}")
return await nodriver.start(
user_data_dir=user_data_dir,
browser_args=None if proxy is None else [f"--proxy-server={proxy}"],
browser_executable_path=browser_executable_path,
**kwargs
)

View file

@ -471,20 +471,20 @@ async def download_urls(
await asyncio.sleep(delay)
new_urls = next_urls
def get_streaming(bucket_dir: str, delete_files = False, refine_chunks_with_spacy = False, event_stream: bool = False) -> Iterator[str]:
bucket_dir = Path(bucket_dir)
bucket_dir.mkdir(parents=True, exist_ok=True)
try:
def get_downloads_urls(bucket_dir: Path, delete_files: bool = False) -> Iterator[str]:
download_file = bucket_dir / DOWNLOADS_FILE
if download_file.exists():
urls = []
with download_file.open('r') as f:
data = json.load(f)
if delete_files:
download_file.unlink()
if isinstance(data, list):
for item in data:
if "url" in item:
urls.append(item["url"])
yield item["url"]
def read_and_download_urls(bucket_dir: Path, event_stream: bool = False) -> Iterator[str]:
urls = get_downloads_urls(bucket_dir)
if urls:
count = 0
with open(os.path.join(bucket_dir, FILE_LIST), 'w') as f:
@ -494,8 +494,20 @@ def get_streaming(bucket_dir: str, delete_files = False, refine_chunks_with_spac
count += 1
yield f'data: {json.dumps({"action": "download", "count": count})}\n\n'
if refine_chunks_with_spacy:
async def async_read_and_download_urls(bucket_dir: Path, event_stream: bool = False) -> Iterator[str]:
urls = get_downloads_urls(bucket_dir)
if urls:
count = 0
with open(os.path.join(bucket_dir, FILE_LIST), 'w') as f:
async for filename in download_urls(bucket_dir, urls):
f.write(f"{filename}\n")
if event_stream:
count += 1
yield f'data: {json.dumps({"action": "download", "count": count})}\n\n'
def stream_chunks(bucket_dir: Path, delete_files: bool = False, refine_chunks_with_spacy: bool = False, event_stream: bool = False) -> Iterator[str]:
size = 0
if refine_chunks_with_spacy:
for chunk in stream_read_parts_and_refine(bucket_dir, delete_files):
if event_stream:
size += len(chunk)
@ -505,7 +517,6 @@ def get_streaming(bucket_dir: str, delete_files = False, refine_chunks_with_spac
else:
streaming = stream_read_files(bucket_dir, get_filenames(bucket_dir), delete_files)
streaming = cache_stream(streaming, bucket_dir)
size = 0
for chunk in streaming:
if event_stream:
size += len(chunk)
@ -522,6 +533,26 @@ def get_streaming(bucket_dir: str, delete_files = False, refine_chunks_with_spac
yield f'data: {json.dumps({"action": "delete_files"})}\n\n'
if event_stream:
yield f'data: {json.dumps({"action": "done", "size": size})}\n\n'
def get_streaming(bucket_dir: str, delete_files = False, refine_chunks_with_spacy = False, event_stream: bool = False) -> Iterator[str]:
bucket_dir = Path(bucket_dir)
bucket_dir.mkdir(parents=True, exist_ok=True)
try:
yield from read_and_download_urls(bucket_dir, event_stream)
yield from stream_chunks(bucket_dir, delete_files, refine_chunks_with_spacy, event_stream)
except Exception as e:
if event_stream:
yield f'data: {json.dumps({"error": {"message": str(e)}})}\n\n'
raise e
async def get_async_streaming(bucket_dir: str, delete_files = False, refine_chunks_with_spacy = False, event_stream: bool = False) -> Iterator[str]:
bucket_dir = Path(bucket_dir)
bucket_dir.mkdir(parents=True, exist_ok=True)
try:
async for chunk in async_read_and_download_urls(bucket_dir, event_stream):
yield chunk
for chunk in stream_chunks(bucket_dir, delete_files, refine_chunks_with_spacy, event_stream):
yield chunk
except Exception as e:
if event_stream:
yield f'data: {json.dumps({"error": {"message": str(e)}})}\n\n'

View file

@ -24,6 +24,7 @@ except:
from typing import Iterator
from ..cookies import get_cookies_dir
from ..providers.response import format_link
from ..errors import MissingRequirementsError
from .. import debug
@ -66,7 +67,7 @@ class SearchResultEntry():
def set_text(self, text: str):
self.text = text
def scrape_text(html: str, max_words: int = None, add_source=True) -> Iterator[str]:
def scrape_text(html: str, max_words: int = None, add_source=True, count_images: int = 2) -> Iterator[str]:
source = BeautifulSoup(html, "html.parser")
soup = source
for selector in [
@ -88,7 +89,20 @@ def scrape_text(html: str, max_words: int = None, add_source=True) -> Iterator[s
if select:
select.extract()
for paragraph in soup.select("p, table:not(:has(p)), ul:not(:has(p)), h1, h2, h3, h4, h5, h6"):
image_select = "img[alt][src^=http]:not([alt=''])"
image_link_select = f"a:has({image_select})"
for paragraph in soup.select(f"h1, h2, h3, h4, h5, h6, p, table:not(:has(p)), ul:not(:has(p)), {image_link_select}"):
image = paragraph.select_one(image_select)
if count_images > 0:
if image:
title = paragraph.get("title") or paragraph.text
if title:
yield f"!{format_link(image['src'], title)}\n"
if max_words is not None:
max_words -= 10
count_images -= 1
continue
for line in paragraph.text.splitlines():
words = [word for word in line.replace("\t", " ").split(" ") if word]
count = len(words)
@ -112,7 +126,7 @@ async def fetch_and_scrape(session: ClientSession, url: str, max_words: int = No
bucket_dir: Path = Path(get_cookies_dir()) / ".scrape_cache" / "fetch_and_scrape"
bucket_dir.mkdir(parents=True, exist_ok=True)
md5_hash = hashlib.md5(url.encode()).hexdigest()
cache_file = bucket_dir / f"{url.split('?')[0].split('//')[1].replace('/', '+')[:16]}.{datetime.date.today()}.{md5_hash}.txt"
cache_file = bucket_dir / f"{url.split('?')[0].split('//')[1].replace('/', '+')[:16]}.{datetime.date.today()}.{md5_hash}.cache"
if cache_file.exists():
return cache_file.read_text()
async with session.get(url) as response:
@ -179,12 +193,13 @@ async def do_search(prompt: str, query: str = None, instructions: str = DEFAULT_
md5_hash = hashlib.md5(json_bytes).hexdigest()
bucket_dir: Path = Path(get_cookies_dir()) / ".scrape_cache" / f"web_search" / f"{datetime.date.today()}"
bucket_dir.mkdir(parents=True, exist_ok=True)
cache_file = bucket_dir / f"{quote_plus(query[:20])}.{md5_hash}.txt"
cache_file = bucket_dir / f"{quote_plus(query[:20])}.{md5_hash}.cache"
if cache_file.exists():
with cache_file.open("r") as f:
search_results = f.read()
else:
search_results = await search(query, **kwargs)
if search_results.results:
with cache_file.open("w") as f:
f.write(str(search_results))

View file

@ -16,7 +16,8 @@ brotli
beautifulsoup4
aiohttp_socks
pywebview
plyer
cryptography
nodriver
python-multipart
pypdf2
docx

View file

@ -36,6 +36,11 @@ EXTRA_REQUIRE = {
"uvicorn", # api
"nodriver",
"python-multipart",
"pypdf2", # files
"docx",
"odfpy",
"ebooklib",
"openpyxl",
],
'slim': [
"curl_cffi>=0.6.2",
@ -49,22 +54,17 @@ EXTRA_REQUIRE = {
"fastapi", # api
"uvicorn", # api
"python-multipart",
"pypdf2", # files
"docx",
],
"image": [
"pillow",
"cairosvg",
"beautifulsoup4"
],
"webdriver": [
"platformdirs",
"undetected-chromedriver>=3.5.5",
"setuptools",
"selenium-wire"
],
"webview": [
"webview",
"pywebview",
"platformdirs",
"plyer",
"cryptography"
],
"api": [
@ -87,7 +87,6 @@ EXTRA_REQUIRE = {
],
"files": [
"spacy",
"filesplit",
"beautifulsoup4",
"pypdf2",
"docx",