Spaces:
Paused
Paused
Some adjustments to the API providers
Browse files
g4f/Provider/Providers/Bing.py
CHANGED
|
@@ -24,7 +24,6 @@ def _create_completion(model: str, messages: list, stream: bool, **kwargs):
|
|
| 24 |
p = subprocess.Popen(cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
|
| 25 |
|
| 26 |
for line in iter(p.stdout.readline, b''):
|
| 27 |
-
#print(line)
|
| 28 |
yield line.decode('utf-8', errors='ignore') #[:-1]
|
| 29 |
|
| 30 |
|
|
|
|
| 24 |
p = subprocess.Popen(cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
|
| 25 |
|
| 26 |
for line in iter(p.stdout.readline, b''):
|
|
|
|
| 27 |
yield line.decode('utf-8', errors='ignore') #[:-1]
|
| 28 |
|
| 29 |
|
g4f/Provider/Providers/Forefront.py
CHANGED
|
@@ -7,6 +7,7 @@ url = 'forefront.com'
|
|
| 7 |
model = ['gpt-3.5-turbo']
|
| 8 |
supports_stream = True
|
| 9 |
|
|
|
|
| 10 |
def _create_completion(model: str, messages: list, stream: bool, **kwargs):
|
| 11 |
json_data = {
|
| 12 |
'text': messages[-1]['content'],
|
|
@@ -19,18 +20,17 @@ def _create_completion(model: str, messages: list, stream: bool, **kwargs):
|
|
| 19 |
'messages': messages[:-1] if len(messages) > 1 else [],
|
| 20 |
'internetMode': 'auto'
|
| 21 |
}
|
| 22 |
-
print(json_data)
|
| 23 |
-
response = requests.post( 'https://streaming.tenant-forefront-default.knative.chi.coreweave.com/free-chat',
|
| 24 |
-
json=json_data, stream=True)
|
| 25 |
|
| 26 |
-
|
|
|
|
|
|
|
|
|
|
| 27 |
if b'delta' in token:
|
| 28 |
token = json.loads(token.decode().split('data: ')[1])['delta']
|
| 29 |
yield (token)
|
| 30 |
|
| 31 |
-
params = f'g4f.Providers.{os.path.basename(__file__)[:-3]} supports: ' + \
|
| 32 |
-
'(%s)' % ', '.join([f"{name}: {get_type_hints(_create_completion)[name].__name__}" for name in _create_completion.__code__.co_varnames[:_create_completion.__code__.co_argcount]])
|
| 33 |
-
|
| 34 |
-
|
| 35 |
-
|
| 36 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 7 |
model = ['gpt-3.5-turbo']
|
| 8 |
supports_stream = True
|
| 9 |
|
| 10 |
+
|
| 11 |
def _create_completion(model: str, messages: list, stream: bool, **kwargs):
|
| 12 |
json_data = {
|
| 13 |
'text': messages[-1]['content'],
|
|
|
|
| 20 |
'messages': messages[:-1] if len(messages) > 1 else [],
|
| 21 |
'internetMode': 'auto'
|
| 22 |
}
|
|
|
|
|
|
|
|
|
|
| 23 |
|
| 24 |
+
response = requests.post('https://streaming.tenant-forefront-default.knative.chi.coreweave.com/free-chat',
|
| 25 |
+
json=json_data, stream=True)
|
| 26 |
+
|
| 27 |
+
for token in response.iter_lines():
|
| 28 |
if b'delta' in token:
|
| 29 |
token = json.loads(token.decode().split('data: ')[1])['delta']
|
| 30 |
yield (token)
|
| 31 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 32 |
|
| 33 |
+
params = f'g4f.Providers.{os.path.basename(__file__)[:-3]} supports: ' + \
|
| 34 |
+
'(%s)' % ', '.join(
|
| 35 |
+
[f"{name}: {get_type_hints(_create_completion)[name].__name__}" for name in
|
| 36 |
+
_create_completion.__code__.co_varnames[:_create_completion.__code__.co_argcount]])
|
g4f/Provider/Providers/Openai.py
CHANGED
|
@@ -14,19 +14,19 @@ def _create_completion(model: str, messages: list, stream: bool, **kwargs):
|
|
| 14 |
|
| 15 |
path = os.path.dirname(os.path.realpath(__file__))
|
| 16 |
config = json.dumps({
|
| 17 |
-
'
|
| 18 |
-
'
|
|
|
|
| 19 |
|
| 20 |
try:
|
| 21 |
subprocess.run(["python3", "--version"], capture_output=True, check=True)
|
| 22 |
-
cmd = ["python3", f"{path}/helpers/
|
| 23 |
except subprocess.CalledProcessError:
|
| 24 |
-
cmd = ["python", f"{path}/helpers/
|
| 25 |
|
| 26 |
p = subprocess.Popen(cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
|
| 27 |
|
| 28 |
for line in iter(p.stdout.readline, b''):
|
| 29 |
-
#print(line)
|
| 30 |
yield line.decode('utf-8', errors='ignore') #[:-1]
|
| 31 |
|
| 32 |
params = f'g4f.Providers.{os.path.basename(__file__)[:-3]} supports: ' + \
|
|
|
|
| 14 |
|
| 15 |
path = os.path.dirname(os.path.realpath(__file__))
|
| 16 |
config = json.dumps({
|
| 17 |
+
'model': model,
|
| 18 |
+
'messages': messages[:-1] if len(messages) > 1 else [],
|
| 19 |
+
})
|
| 20 |
|
| 21 |
try:
|
| 22 |
subprocess.run(["python3", "--version"], capture_output=True, check=True)
|
| 23 |
+
cmd = ["python3", f"{path}/helpers/openai.py", config]
|
| 24 |
except subprocess.CalledProcessError:
|
| 25 |
+
cmd = ["python", f"{path}/helpers/openai.py", config]
|
| 26 |
|
| 27 |
p = subprocess.Popen(cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
|
| 28 |
|
| 29 |
for line in iter(p.stdout.readline, b''):
|
|
|
|
| 30 |
yield line.decode('utf-8', errors='ignore') #[:-1]
|
| 31 |
|
| 32 |
params = f'g4f.Providers.{os.path.basename(__file__)[:-3]} supports: ' + \
|
g4f/Provider/Providers/Phind.py
CHANGED
|
@@ -9,14 +9,21 @@ url = 'https://phind.com'
|
|
| 9 |
model = ['gpt-3.5-turbo', 'gpt-4']
|
| 10 |
supports_stream = True
|
| 11 |
|
|
|
|
| 12 |
def _create_completion(model: str, messages: list, stream: bool, **kwargs):
|
| 13 |
|
| 14 |
path = os.path.dirname(os.path.realpath(__file__))
|
| 15 |
config = json.dumps({
|
| 16 |
'model': model,
|
| 17 |
-
'messages': messages
|
|
|
|
| 18 |
|
| 19 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 20 |
|
| 21 |
p = subprocess.Popen(cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
|
| 22 |
|
|
@@ -25,13 +32,13 @@ def _create_completion(model: str, messages: list, stream: bool, **kwargs):
|
|
| 25 |
os.system('clear' if os.name == 'posix' else 'cls')
|
| 26 |
yield 'Clouflare error, please try again...'
|
| 27 |
os._exit(0)
|
| 28 |
-
|
| 29 |
else:
|
| 30 |
if b'ping - 2023-' in line:
|
| 31 |
continue
|
| 32 |
-
|
| 33 |
-
|
| 34 |
|
| 35 |
-
|
| 36 |
params = f'g4f.Providers.{os.path.basename(__file__)[:-3]} supports: ' + \
|
| 37 |
-
'(%s)' % ', '.join(
|
|
|
|
|
|
| 9 |
model = ['gpt-3.5-turbo', 'gpt-4']
|
| 10 |
supports_stream = True
|
| 11 |
|
| 12 |
+
|
| 13 |
def _create_completion(model: str, messages: list, stream: bool, **kwargs):
|
| 14 |
|
| 15 |
path = os.path.dirname(os.path.realpath(__file__))
|
| 16 |
config = json.dumps({
|
| 17 |
'model': model,
|
| 18 |
+
'messages': messages[:-1] if len(messages) > 1 else [],
|
| 19 |
+
})
|
| 20 |
|
| 21 |
+
try:
|
| 22 |
+
subprocess.run(["python3", "--version"],
|
| 23 |
+
capture_output=True, check=True)
|
| 24 |
+
cmd = ["python3", f"{path}/helpers/phind.py", config]
|
| 25 |
+
except subprocess.CalledProcessError:
|
| 26 |
+
cmd = ["python", f"{path}/helpers/phind.py", config]
|
| 27 |
|
| 28 |
p = subprocess.Popen(cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
|
| 29 |
|
|
|
|
| 32 |
os.system('clear' if os.name == 'posix' else 'cls')
|
| 33 |
yield 'Clouflare error, please try again...'
|
| 34 |
os._exit(0)
|
| 35 |
+
|
| 36 |
else:
|
| 37 |
if b'ping - 2023-' in line:
|
| 38 |
continue
|
| 39 |
+
yield line.decode('utf-8', errors='ignore') # [:-1]
|
| 40 |
+
|
| 41 |
|
|
|
|
| 42 |
params = f'g4f.Providers.{os.path.basename(__file__)[:-3]} supports: ' + \
|
| 43 |
+
'(%s)' % ', '.join(
|
| 44 |
+
[f"{name}: {get_type_hints(_create_completion)[name].__name__}" for name in _create_completion.__code__.co_varnames[:_create_completion.__code__.co_argcount]])
|
g4f/Provider/Providers/Theb.py
CHANGED
|
@@ -13,19 +13,19 @@ def _create_completion(model: str, messages: list, stream: bool, **kwargs):
|
|
| 13 |
|
| 14 |
path = os.path.dirname(os.path.realpath(__file__))
|
| 15 |
config = json.dumps({
|
| 16 |
-
'
|
| 17 |
-
'
|
|
|
|
| 18 |
|
| 19 |
try:
|
| 20 |
subprocess.run(["python3", "--version"], capture_output=True, check=True)
|
| 21 |
-
cmd = ["python3", f"{path}/helpers/
|
| 22 |
except subprocess.CalledProcessError:
|
| 23 |
-
cmd = ["python", f"{path}/helpers/
|
| 24 |
|
| 25 |
p = subprocess.Popen(cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
|
| 26 |
|
| 27 |
for line in iter(p.stdout.readline, b''):
|
| 28 |
-
#print(line)
|
| 29 |
yield line.decode('utf-8', errors='ignore') #[:-1]
|
| 30 |
|
| 31 |
|
|
|
|
| 13 |
|
| 14 |
path = os.path.dirname(os.path.realpath(__file__))
|
| 15 |
config = json.dumps({
|
| 16 |
+
'model': model,
|
| 17 |
+
'messages': messages[:-1] if len(messages) > 1 else [],
|
| 18 |
+
})
|
| 19 |
|
| 20 |
try:
|
| 21 |
subprocess.run(["python3", "--version"], capture_output=True, check=True)
|
| 22 |
+
cmd = ["python3", f"{path}/helpers/theb.py", config]
|
| 23 |
except subprocess.CalledProcessError:
|
| 24 |
+
cmd = ["python", f"{path}/helpers/theb.py", config]
|
| 25 |
|
| 26 |
p = subprocess.Popen(cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
|
| 27 |
|
| 28 |
for line in iter(p.stdout.readline, b''):
|
|
|
|
| 29 |
yield line.decode('utf-8', errors='ignore') #[:-1]
|
| 30 |
|
| 31 |
|
g4f/Provider/Providers/You.py
CHANGED
|
@@ -9,15 +9,22 @@ url = 'https://you.com'
|
|
| 9 |
model = 'gpt-3.5-turbo'
|
| 10 |
supports_stream = True
|
| 11 |
|
|
|
|
| 12 |
def _create_completion(model: str, messages: list, stream: bool, **kwargs):
|
| 13 |
|
| 14 |
path = os.path.dirname(os.path.realpath(__file__))
|
| 15 |
config = json.dumps({
|
| 16 |
-
'messages': messages
|
| 17 |
-
|
| 18 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 19 |
|
| 20 |
p = subprocess.Popen(cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
|
| 21 |
|
| 22 |
for line in iter(p.stdout.readline, b''):
|
| 23 |
-
yield line.decode('utf-8')
|
|
|
|
| 9 |
model = 'gpt-3.5-turbo'
|
| 10 |
supports_stream = True
|
| 11 |
|
| 12 |
+
|
| 13 |
def _create_completion(model: str, messages: list, stream: bool, **kwargs):
|
| 14 |
|
| 15 |
path = os.path.dirname(os.path.realpath(__file__))
|
| 16 |
config = json.dumps({
|
| 17 |
+
'messages': messages[:-1] if len(messages) > 1 else [],
|
| 18 |
+
})
|
| 19 |
+
|
| 20 |
+
try:
|
| 21 |
+
subprocess.run(["python3", "--version"],
|
| 22 |
+
capture_output=True, check=True)
|
| 23 |
+
cmd = ["python3", f"{path}/helpers/you.py", config]
|
| 24 |
+
except subprocess.CalledProcessError:
|
| 25 |
+
cmd = ["python", f"{path}/helpers/you.py", config]
|
| 26 |
|
| 27 |
p = subprocess.Popen(cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
|
| 28 |
|
| 29 |
for line in iter(p.stdout.readline, b''):
|
| 30 |
+
yield line.decode('utf-8', errors='ignore') # [:-1]
|
g4f/Provider/Providers/helpers/bing.py
CHANGED
|
@@ -198,7 +198,6 @@ class AsyncCompletion:
|
|
| 198 |
continue
|
| 199 |
|
| 200 |
response = json.loads(obj)
|
| 201 |
-
#print(response, flush=True, end='')
|
| 202 |
if response.get('type') == 1 and response['arguments'][0].get('messages',):
|
| 203 |
response_text = response['arguments'][0]['messages'][0]['adaptiveCards'][0]['body'][0].get('text')
|
| 204 |
|
|
|
|
| 198 |
continue
|
| 199 |
|
| 200 |
response = json.loads(obj)
|
|
|
|
| 201 |
if response.get('type') == 1 and response['arguments'][0].get('messages',):
|
| 202 |
response_text = response['arguments'][0]['messages'][0]['adaptiveCards'][0]['body'][0].get('text')
|
| 203 |
|
server/backend.py
CHANGED
|
@@ -9,8 +9,6 @@ from requests import get
|
|
| 9 |
from server.auto_proxy import get_random_proxy, update_working_proxies
|
| 10 |
from server.config import special_instructions
|
| 11 |
|
| 12 |
-
# Backend_Api class definition
|
| 13 |
-
|
| 14 |
|
| 15 |
class Backend_Api:
|
| 16 |
def __init__(self, app, config: dict) -> None:
|
|
@@ -47,7 +45,7 @@ class Backend_Api:
|
|
| 47 |
|
| 48 |
# Generate response
|
| 49 |
response = ChatCompletion.create(model=model, stream=True,
|
| 50 |
-
messages=messages, provider=g4f.Provider.
|
| 51 |
|
| 52 |
return self.app.response_class(generate_stream(response, jailbreak), mimetype='text/event-stream')
|
| 53 |
|
|
|
|
| 9 |
from server.auto_proxy import get_random_proxy, update_working_proxies
|
| 10 |
from server.config import special_instructions
|
| 11 |
|
|
|
|
|
|
|
| 12 |
|
| 13 |
class Backend_Api:
|
| 14 |
def __init__(self, app, config: dict) -> None:
|
|
|
|
| 45 |
|
| 46 |
# Generate response
|
| 47 |
response = ChatCompletion.create(model=model, stream=True,
|
| 48 |
+
messages=messages, provider=g4f.Provider.Yqcloud)
|
| 49 |
|
| 50 |
return self.app.response_class(generate_stream(response, jailbreak), mimetype='text/event-stream')
|
| 51 |
|