|
from __future__ import annotations |
|
|
|
import json, random |
|
from aiohttp import ClientSession |
|
|
|
from ..typing import AsyncResult, Messages |
|
from .base_provider import AsyncGeneratorProvider |
|
|
|
models = { |
|
"claude-v2": "claude-2.0", |
|
"claude-v2.1":"claude-2.1", |
|
"gemini-pro": "google-gemini-pro" |
|
} |
|
urls = [ |
|
"https://free.chatgpt.org.uk", |
|
"https://ai.chatgpt.org.uk" |
|
] |
|
|
|
class FreeChatgpt(AsyncGeneratorProvider): |
|
url = "https://free.chatgpt.org.uk" |
|
working = True |
|
supports_gpt_35_turbo = True |
|
supports_gpt_4 = True |
|
supports_message_history = True |
|
|
|
@classmethod |
|
async def create_async_generator( |
|
cls, |
|
model: str, |
|
messages: Messages, |
|
proxy: str = None, |
|
**kwargs |
|
) -> AsyncResult: |
|
if model in models: |
|
model = models[model] |
|
elif not model: |
|
model = "gpt-3.5-turbo" |
|
url = random.choice(urls) |
|
headers = { |
|
"Accept": "application/json, text/event-stream", |
|
"Content-Type":"application/json", |
|
"Accept-Encoding": "gzip, deflate, br", |
|
"Accept-Language": "en-US,en;q=0.5", |
|
"Host":"free.chatgpt.org.uk", |
|
"Referer":f"{cls.url}/", |
|
"Origin":f"{cls.url}", |
|
"Sec-Fetch-Dest": "empty", |
|
"Sec-Fetch-Mode": "cors", |
|
"Sec-Fetch-Site": "same-origin", |
|
"User-Agent": "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/120.0.0.0 Safari/537.36", |
|
} |
|
async with ClientSession(headers=headers) as session: |
|
data = { |
|
"messages":messages, |
|
"stream":True, |
|
"model":model, |
|
"temperature":0.5, |
|
"presence_penalty":0, |
|
"frequency_penalty":0, |
|
"top_p":1, |
|
**kwargs |
|
} |
|
async with session.post(f'{url}/api/openai/v1/chat/completions', json=data, proxy=proxy) as response: |
|
response.raise_for_status() |
|
started = False |
|
async for line in response.content: |
|
if line.startswith(b"data: [DONE]"): |
|
break |
|
elif line.startswith(b"data: "): |
|
line = json.loads(line[6:]) |
|
if(line["choices"]==[]): |
|
continue |
|
chunk = line["choices"][0]["delta"].get("content") |
|
if chunk: |
|
started = True |
|
yield chunk |
|
if not started: |
|
raise RuntimeError("Empty response") |