|
from __future__ import annotations
|
|
|
|
import configparser
|
|
import os
|
|
import threading
|
|
import re
|
|
|
|
from modules import shared, errors, cache, scripts
|
|
from modules.gitpython_hack import Repo
|
|
from modules.paths_internal import extensions_dir, extensions_builtin_dir, script_path
|
|
from modules_forge.config import always_disabled_extensions
|
|
|
|
|
|
os.makedirs(extensions_dir, exist_ok=True)
|
|
extensions: list["Extension"] = []
|
|
|
|
|
|
def active():
|
|
if shared.cmd_opts.disable_all_extensions or shared.opts.disable_all_extensions == "all":
|
|
return []
|
|
elif shared.cmd_opts.disable_extra_extensions or shared.opts.disable_all_extensions == "extra":
|
|
return [x for x in extensions if x.enabled and x.is_builtin]
|
|
else:
|
|
return [x for x in extensions if x.enabled]
|
|
|
|
|
|
class ExtensionMetadata:
|
|
filename = "metadata.ini"
|
|
config: configparser.ConfigParser
|
|
canonical_name: str
|
|
requires: list[str]
|
|
|
|
def __init__(self, path, folder_name):
|
|
filepath = os.path.join(path, self.filename)
|
|
|
|
try:
|
|
self.config = configparser.ConfigParser()
|
|
self.config.read(filepath)
|
|
except Exception:
|
|
|
|
errors.report(f"Error reading {self.filename} for extension {folder_name}", exc_info=True)
|
|
|
|
canonical_name = self.config.get("Extension", "Name", fallback=folder_name)
|
|
self.canonical_name = canonical_name.lower().strip()
|
|
self.requires = self.get_script_requirements("Requires", "Extension")
|
|
|
|
def get_script_requirements(self, field, section, extra_section=None):
|
|
"""
|
|
reads a list of requirements from the config; `field` is the name of the field in the `.ini` file,
|
|
like **Requires** or **Before**, and `section` is the name of the `[section]` in the `.ini` file;
|
|
additionally, reads more requirements from `[extra_section]` if specified
|
|
"""
|
|
|
|
x = self.config.get(section, field, fallback="")
|
|
if extra_section:
|
|
x = ",".join([x, self.config.get(extra_section, field, fallback="")])
|
|
|
|
return self.parse_list(x.lower())
|
|
|
|
def parse_list(self, text):
|
|
"""
|
|
converts a line from config `"ext1 ext2, ext3 "` into a python list `["ext1", "ext2", "ext3"]`
|
|
"""
|
|
|
|
if not text:
|
|
return []
|
|
|
|
return [x.strip() for x in re.split(r"[,\s]+", text) if x.strip()]
|
|
|
|
|
|
class Extension:
|
|
lock = threading.Lock()
|
|
metadata: ExtensionMetadata
|
|
cached_fields = ("remote", "commit_date", "branch", "commit_hash", "version")
|
|
|
|
def __init__(self, name, path, enabled=True, is_builtin=False, metadata=None):
|
|
self.name = name
|
|
self.path = path
|
|
self.enabled = enabled
|
|
self.is_builtin = is_builtin
|
|
self.metadata = metadata
|
|
|
|
self.status = ""
|
|
self.can_update = False
|
|
self.commit_hash = ""
|
|
self.commit_date = None
|
|
self.version = ""
|
|
self.branch = None
|
|
self.remote = None
|
|
self.have_info_from_repo = False
|
|
self.canonical_name = metadata.canonical_name
|
|
|
|
def to_dict(self):
|
|
return {x: getattr(self, x) for x in self.cached_fields}
|
|
|
|
def from_dict(self, d):
|
|
for field in self.cached_fields:
|
|
setattr(self, field, d[field])
|
|
|
|
def read_info_from_repo(self):
|
|
if self.is_builtin or self.have_info_from_repo:
|
|
return
|
|
|
|
def read_from_repo():
|
|
with self.lock:
|
|
self.do_read_info_from_repo()
|
|
return self.to_dict()
|
|
|
|
try:
|
|
d = cache.cached_data_for_file("extensions-git", self.name, os.path.join(self.path, ".git"), read_from_repo)
|
|
self.from_dict(d)
|
|
except FileNotFoundError:
|
|
pass
|
|
|
|
self.status = "unknown" if self.status == "" else self.status
|
|
|
|
def do_read_info_from_repo(self):
|
|
repo = None
|
|
if os.path.exists(os.path.join(self.path, ".git")):
|
|
try:
|
|
repo = Repo(self.path)
|
|
except Exception:
|
|
errors.report(f"Error reading github repository info from {self.path}", exc_info=True)
|
|
repo = None
|
|
|
|
self.have_info_from_repo = True
|
|
|
|
if repo is None or repo.bare:
|
|
self.remote = None
|
|
return
|
|
|
|
try:
|
|
self.remote = next(repo.remote().urls, None)
|
|
commit = repo.head.commit
|
|
self.commit_date = commit.committed_date
|
|
if repo.active_branch:
|
|
self.branch = repo.active_branch.name
|
|
self.commit_hash = commit.hexsha
|
|
self.version = self.commit_hash[:8]
|
|
|
|
except Exception:
|
|
errors.report(f"Error reading extension data ({self.name})", exc_info=True)
|
|
self.remote = None
|
|
|
|
def list_files(self, subdir, extension):
|
|
dirpath = os.path.join(self.path, subdir)
|
|
if not os.path.isdir(dirpath):
|
|
return []
|
|
|
|
res = []
|
|
for filename in sorted(os.listdir(dirpath)):
|
|
res.append(scripts.ScriptFile(self.path, filename, os.path.join(dirpath, filename)))
|
|
|
|
return [x for x in res if os.path.splitext(x.path)[1].lower() == extension and os.path.isfile(x.path)]
|
|
|
|
def check_updates(self):
|
|
repo = Repo(self.path)
|
|
for fetch in repo.remote().fetch(dry_run=True):
|
|
if fetch.flags != fetch.HEAD_UPTODATE:
|
|
self.can_update = True
|
|
self.status = "new commits"
|
|
return
|
|
|
|
try:
|
|
origin = repo.rev_parse("origin")
|
|
if repo.head.commit != origin:
|
|
self.can_update = True
|
|
self.status = "behind HEAD"
|
|
return
|
|
except Exception:
|
|
self.can_update = False
|
|
self.status = "unknown (remote error)"
|
|
return
|
|
|
|
self.can_update = False
|
|
self.status = "latest"
|
|
|
|
def fetch_and_reset_hard(self, commit="origin"):
|
|
|
|
repo = Repo(self.path)
|
|
repo.git.fetch(all=True)
|
|
repo.git.reset(commit, hard=True)
|
|
self.have_info_from_repo = False
|
|
|
|
|
|
def list_extensions():
|
|
extensions.clear()
|
|
loaded_extensions = {}
|
|
|
|
if shared.cmd_opts.disable_all_extensions:
|
|
print('*** "--disable-all-extensions" arg was used, will not load any extensions ***')
|
|
return
|
|
if shared.opts.disable_all_extensions == "all":
|
|
print('*** "Disable all extensions" option was set, will not load any extensions ***')
|
|
return
|
|
|
|
if shared.cmd_opts.disable_extra_extensions:
|
|
print('*** "--disable-extra-extensions" arg was used, will only load built-in extensions ***')
|
|
elif shared.opts.disable_all_extensions == "extra":
|
|
print('*** "Disable all extensions" option was set, will only load built-in extensions ***')
|
|
|
|
disabled_extensions = shared.opts.disabled_extensions + always_disabled_extensions
|
|
|
|
for dirname in (extensions_builtin_dir, extensions_dir):
|
|
if not os.path.isdir(dirname):
|
|
continue
|
|
|
|
for extension_dirname in sorted(os.listdir(dirname)):
|
|
path = os.path.join(dirname, extension_dirname)
|
|
if not os.path.isdir(path):
|
|
continue
|
|
|
|
canonical_name = extension_dirname
|
|
metadata = ExtensionMetadata(path, canonical_name)
|
|
is_builtin = dirname == extensions_builtin_dir
|
|
|
|
already_loaded_extension = loaded_extensions.get(metadata.canonical_name)
|
|
if already_loaded_extension is not None:
|
|
errors.report(f'Duplicated name "{canonical_name}" found in extensions "{extension_dirname}" and "{already_loaded_extension.name}". Former will be discarded!', exc_info=False)
|
|
continue
|
|
|
|
extension = Extension(name=extension_dirname, path=path, enabled=extension_dirname not in disabled_extensions, is_builtin=is_builtin, metadata=metadata)
|
|
|
|
extensions.append(extension)
|
|
loaded_extensions[canonical_name] = extension
|
|
|
|
for extension in extensions:
|
|
if not extension.enabled:
|
|
continue
|
|
|
|
for req in extension.metadata.requires:
|
|
required_extension = loaded_extensions.get(req)
|
|
if required_extension is None:
|
|
errors.report(f'Extension "{extension.name}" requires "{req}" which is not installed.', exc_info=False)
|
|
continue
|
|
|
|
if not required_extension.enabled:
|
|
errors.report(f'Extension "{extension.name}" requires "{required_extension.name}" which is disabled.', exc_info=False)
|
|
continue
|
|
|