Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
| from huggingface_hub import ModelFilter, snapshot_download | |
| from huggingface_hub import ModelCard | |
| import json | |
| import os | |
| import time | |
| from src.submission.check_validity import is_model_on_hub, check_model_card, get_model_tags | |
| from src.envs import QUEUE_REPO, EVAL_REQUESTS_PATH, DYNAMIC_INFO_REPO, DYNAMIC_INFO_PATH, DYNAMIC_INFO_FILE_PATH, API, H4_TOKEN | |
| def update_one_model(model_id, data, models_on_the_hub): | |
| # Model no longer on the hub at all | |
| if model_id not in models_on_the_hub: | |
| data['still_on_hub'] = False | |
| data['likes'] = 0 | |
| data['downloads'] = 0 | |
| data['created_at'] = "" | |
| data["tags"] = [] | |
| return data | |
| # Grabbing model parameters | |
| model_cfg = models_on_the_hub[model_id] | |
| data['likes'] = model_cfg.likes | |
| data['downloads'] = model_cfg.downloads | |
| data['created_at'] = str(model_cfg.created_at) | |
| data['license'] = model_cfg.card_data.license if model_cfg.card_data is not None else "" | |
| # Grabbing model details | |
| model_name = model_id | |
| if model_cfg.card_data is not None and model_cfg.card_data.base_model is not None: | |
| if isinstance(model_cfg.card_data.base_model, str): | |
| model_name = model_cfg.card_data.base_model # for adapters, we look at the parent model | |
| still_on_hub, _, _ = is_model_on_hub( | |
| model_name=model_name, revision=data.get("revision"), trust_remote_code=True, test_tokenizer=False, token=H4_TOKEN | |
| ) | |
| # If the model doesn't have a model card or a license, we consider it's deleted | |
| if still_on_hub: | |
| try: | |
| status, _, model_card = check_model_card(model_id) | |
| if status is False: | |
| still_on_hub = False | |
| except Exception: | |
| model_card = None | |
| still_on_hub = False | |
| data['still_on_hub'] = still_on_hub | |
| tags = get_model_tags(model_card, model_id) if still_on_hub else [] | |
| data["tags"] = tags | |
| return data | |
| def update_models(file_path, models_on_the_hub): | |
| """ | |
| Search through all JSON files in the specified root folder and its subfolders, | |
| and update the likes key in JSON dict from value of input dict | |
| """ | |
| seen_models = [] | |
| with open(file_path, "r") as f: | |
| model_infos = json.load(f) | |
| for model_id in model_infos.keys(): | |
| seen_models.append(model_id) | |
| model_infos[model_id] = update_one_model( | |
| model_id = model_id, | |
| data=model_infos[model_id], | |
| models_on_the_hub=models_on_the_hub | |
| ) | |
| # If new requests files have been created since we started all this | |
| # we grab them | |
| all_models = [] | |
| try: | |
| for ix, (root, _, files) in enumerate(os.walk(EVAL_REQUESTS_PATH)): | |
| if ix == 0: continue | |
| for file in files: | |
| if "eval_request" in file: | |
| path = root.split("/")[-1] + "/" + file.split("_eval_request")[0] | |
| all_models.append(path) | |
| except Exception as e: | |
| print(e) | |
| pass | |
| for model_id in all_models: | |
| if model_id not in seen_models: | |
| model_infos[model_id] = update_one_model( | |
| model_id = model_id, | |
| data={}, | |
| models_on_the_hub=models_on_the_hub | |
| ) | |
| with open(file_path, 'w') as f: | |
| json.dump(model_infos, f, indent=2) | |
| def update_dynamic_files(): | |
| """ This will only update metadata for models already linked in the repo, not add missing ones. | |
| """ | |
| snapshot_download( | |
| repo_id=DYNAMIC_INFO_REPO, local_dir=DYNAMIC_INFO_PATH, repo_type="dataset", tqdm_class=None, etag_timeout=30 | |
| ) | |
| print("UPDATE_DYNAMIC: Loaded snapshot") | |
| # Get models | |
| start = time.time() | |
| models = list(API.list_models( | |
| #filter=ModelFilter(task="text-generation"), | |
| full=False, | |
| cardData=True, | |
| fetch_config=True, | |
| )) | |
| id_to_model = {model.id : model for model in models} | |
| print(f"UPDATE_DYNAMIC: Downloaded list of models in {time.time() - start:.2f} seconds") | |
| start = time.time() | |
| update_models(DYNAMIC_INFO_FILE_PATH, id_to_model) | |
| print(f"UPDATE_DYNAMIC: updated in {time.time() - start:.2f} seconds") | |
| API.upload_file( | |
| path_or_fileobj=DYNAMIC_INFO_FILE_PATH, | |
| path_in_repo=DYNAMIC_INFO_FILE_PATH.split("/")[-1], | |
| repo_id=DYNAMIC_INFO_REPO, | |
| repo_type="dataset", | |
| commit_message=f"Daily request file update.", | |
| ) | |
| print(f"UPDATE_DYNAMIC: pushed to hub") | |