Spaces:
Sleeping
Sleeping
Update app.py (#6)
Browse files- Update app.py (079b66891aec1c5e2f7f57ad1e3e7a83b33674d2)
app.py
CHANGED
|
@@ -5,13 +5,16 @@ import gradio as gr
|
|
| 5 |
|
| 6 |
from huggingface_hub import create_repo, HfApi
|
| 7 |
from huggingface_hub import snapshot_download
|
|
|
|
| 8 |
|
| 9 |
api = HfApi()
|
| 10 |
|
| 11 |
-
def process_model(model_id, q_method,
|
| 12 |
|
| 13 |
MODEL_NAME = model_id.split('/')[-1]
|
| 14 |
fp16 = f"{MODEL_NAME}/{MODEL_NAME.lower()}.fp16.bin"
|
|
|
|
|
|
|
| 15 |
|
| 16 |
snapshot_download(repo_id=model_id, local_dir = f"{MODEL_NAME}", local_dir_use_symlinks=False)
|
| 17 |
print("Model downloaded successully!")
|
|
@@ -35,11 +38,16 @@ def process_model(model_id, q_method, username, hf_token):
|
|
| 35 |
print("Empty repo created successfully!")
|
| 36 |
|
| 37 |
# Upload gguf files
|
| 38 |
-
api.upload_folder(
|
| 39 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 40 |
repo_id=f"{username}/{MODEL_NAME}-{q_method}-GGUF",
|
| 41 |
-
|
| 42 |
-
token=hf_token
|
| 43 |
)
|
| 44 |
print("Uploaded successfully!")
|
| 45 |
|
|
@@ -65,11 +73,6 @@ iface = gr.Interface(
|
|
| 65 |
label="Quantization Method",
|
| 66 |
info="GGML quantisation type"
|
| 67 |
),
|
| 68 |
-
gr.Textbox(
|
| 69 |
-
lines=1,
|
| 70 |
-
label="Username",
|
| 71 |
-
info="Your Hugging Face username"
|
| 72 |
-
),
|
| 73 |
gr.Textbox(
|
| 74 |
lines=1,
|
| 75 |
label="HF Write Token",
|
|
|
|
| 5 |
|
| 6 |
from huggingface_hub import create_repo, HfApi
|
| 7 |
from huggingface_hub import snapshot_download
|
| 8 |
+
from huggingface_hub import whoami
|
| 9 |
|
| 10 |
api = HfApi()
|
| 11 |
|
| 12 |
+
def process_model(model_id, q_method, hf_token):
|
| 13 |
|
| 14 |
MODEL_NAME = model_id.split('/')[-1]
|
| 15 |
fp16 = f"{MODEL_NAME}/{MODEL_NAME.lower()}.fp16.bin"
|
| 16 |
+
|
| 17 |
+
username = whoami(hf_token)["name"]
|
| 18 |
|
| 19 |
snapshot_download(repo_id=model_id, local_dir = f"{MODEL_NAME}", local_dir_use_symlinks=False)
|
| 20 |
print("Model downloaded successully!")
|
|
|
|
| 38 |
print("Empty repo created successfully!")
|
| 39 |
|
| 40 |
# Upload gguf files
|
| 41 |
+
# api.upload_folder(
|
| 42 |
+
# folder_path=MODEL_NAME,
|
| 43 |
+
# repo_id=f"{username}/{MODEL_NAME}-{q_method}-GGUF",
|
| 44 |
+
# allow_patterns=["*.gguf"],
|
| 45 |
+
# token=hf_token
|
| 46 |
+
# )
|
| 47 |
+
api.upload_file(
|
| 48 |
+
path_or_fileobj=qtype,
|
| 49 |
repo_id=f"{username}/{MODEL_NAME}-{q_method}-GGUF",
|
| 50 |
+
repo_type="model",
|
|
|
|
| 51 |
)
|
| 52 |
print("Uploaded successfully!")
|
| 53 |
|
|
|
|
| 73 |
label="Quantization Method",
|
| 74 |
info="GGML quantisation type"
|
| 75 |
),
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 76 |
gr.Textbox(
|
| 77 |
lines=1,
|
| 78 |
label="HF Write Token",
|