Plex-Meta-Manager/plex_meta_manager.py

1124 lines
58 KiB
Python
Raw Normal View History

2024-01-04 20:24:53 +00:00
import argparse, os, platform, re, sys, time, uuid
2023-01-09 21:22:39 +00:00
from collections import Counter
2022-06-07 20:42:17 +00:00
from concurrent.futures import ProcessPoolExecutor
2022-10-17 18:54:42 +00:00
from datetime import datetime
2022-09-27 06:19:29 +00:00
from modules.logs import MyLogger
2021-12-17 14:24:46 +00:00
if sys.version_info[0] != 3 or sys.version_info[1] < 8:
print("Version Error: Version: %s.%s.%s incompatible please use Python 3.8+" % (sys.version_info[0], sys.version_info[1], sys.version_info[2]))
2023-03-10 15:07:25 +00:00
sys.exit(0)
try:
2023-03-10 15:07:25 +00:00
import plexapi, psutil, requests, schedule
2023-04-07 20:03:59 +00:00
from dotenv import load_dotenv
2022-07-10 08:04:06 +00:00
from PIL import ImageFile
2022-06-30 13:14:09 +00:00
from plexapi import server
2022-01-26 14:27:22 +00:00
from plexapi.exceptions import NotFound
from plexapi.video import Show, Season
except (ModuleNotFoundError, ImportError):
print("Requirements Error: Requirements are not installed")
sys.exit(0)
2023-08-23 13:40:08 +00:00
default_dir = os.path.join(os.path.dirname(os.path.abspath(__file__)), "config")
load_dotenv(os.path.join(default_dir, ".env"))
arguments = {
"config": {"args": "c", "type": "str", "help": "Run with desired *.yml file"},
"times": {"args": ["t", "time"], "type": "str", "default": "05:00", "help": "Times to update each day use format HH:MM (Default: 05:00) (comma-separated list)"},
"run": {"args": "r", "type": "bool", "help": "Run without the scheduler"},
2023-12-25 08:16:04 +00:00
"tests": {"args": ["ts", "rt", "test", "run-test", "run-tests"], "type": "bool", "help": "Run in debug mode with only collections that have test: true"},
2023-08-23 13:40:08 +00:00
"debug": {"args": "db", "type": "bool", "help": "Run with Debug Logs Reporting to the Command Window"},
"trace": {"args": "tr", "type": "bool", "help": "Run with extra Trace Debug Logs"},
"log-requests": {"args": ["lr", "log-request"], "type": "bool", "help": "Run with all Requests printed"},
"timeout": {"args": "ti", "type": "int", "default": 180, "help": "PMM Global Timeout (Default: 180)"},
"collections-only": {"args": ["co", "collection-only"], "type": "bool", "help": "Run only collection files"},
"metadata-only": {"args": ["mo", "metadatas-only"], "type": "bool", "help": "Run only metadata files"},
"playlists-only": {"args": ["po", "playlist-only"], "type": "bool", "help": "Run only playlist files"},
2023-08-23 13:40:08 +00:00
"operations-only": {"args": ["op", "operation", "operations", "lo", "library-only", "libraries-only", "operation-only"], "type": "bool", "help": "Run only operations"},
"overlays-only": {"args": ["ov", "overlay", "overlays", "overlay-only"], "type": "bool", "help": "Run only overlay files"},
2023-08-23 13:40:08 +00:00
"run-collections": {"args": ["rc", "cl", "collection", "collections", "run-collection"], "type": "str", "help": "Process only specified collections (pipe-separated list '|')"},
"run-libraries": {"args": ["rl", "l", "library", "libraries", "run-library"], "type": "str", "help": "Process only specified libraries (pipe-separated list '|')"},
"run-files": {"args": ["rf", "rm", "m", "run-file", "metadata", "metadata-files", "run-metadata-files"], "type": "str", "help": "Process only specified Files (pipe-separated list '|')"},
2023-08-23 13:40:08 +00:00
"ignore-schedules": {"args": "is", "type": "bool", "help": "Run ignoring collection schedules"},
"ignore-ghost": {"args": "ig", "type": "bool", "help": "Run ignoring ghost logging"},
"delete-collections": {"args": ["dc", "delete", "delete-collection"], "type": "bool", "help": "Deletes all Collections in the Plex Library before running"},
"delete-labels": {"args": ["dl", "delete-label"], "type": "bool", "help": "Deletes all Labels in the Plex Library before running"},
"resume": {"args": "re", "type": "str", "help": "Resume collection run from a specific collection"},
"no-countdown": {"args": "nc", "type": "bool", "help": "Run without displaying the countdown"},
"no-missing": {"args": "nm", "type": "bool", "help": "Run without running the missing section"},
"no-report": {"args": "nr", "type": "bool", "help": "Run without saving a report"},
"read-only-config": {"args": "ro", "type": "bool", "help": "Run without writing to the config"},
"divider": {"args": "d", "type": "str", "default": "=", "help": "Character that divides the sections (Default: '=')"},
"width": {"args": "w", "type": "int", "default": 100, "help": "Screen Width (Default: 100)"},
}
2021-01-20 21:37:59 +00:00
parser = argparse.ArgumentParser()
2023-08-23 13:40:08 +00:00
for arg_key, arg_data in arguments.items():
temp_args = arg_data["args"] if isinstance(arg_data["args"], list) else [arg_data["args"]]
args = [f"--{arg_key}"] + [f"--{a}" if len(a) > 2 else f"-{a}" for a in temp_args]
kwargs = {"dest": arg_key.replace("-", "_"), "help": arg_data["help"]}
if arg_data["type"] == "bool":
kwargs["action"] = "store_true"
kwargs["default"] = False # noqa
2023-08-23 13:40:08 +00:00
else:
kwargs["type"] = int if arg_data["type"] == "int" else str
if "default" in arg_data:
kwargs["default"] = arg_data["default"]
parser.add_argument(*args, **kwargs)
2023-04-07 20:03:59 +00:00
args, unknown = parser.parse_known_args()
2021-01-20 21:37:59 +00:00
2023-04-07 20:03:59 +00:00
2023-08-23 13:40:08 +00:00
def get_env(env_str, default, arg_bool=False, arg_int=False):
2022-04-18 18:16:39 +00:00
env_vars = [env_str] if not isinstance(env_str, list) else env_str
final_value = None
2023-04-07 20:03:59 +00:00
static_envs.extend(env_vars)
2022-04-18 18:16:39 +00:00
for env_var in env_vars:
env_value = os.environ.get(env_var)
if env_value is not None:
final_value = env_value
break
2023-01-04 16:43:32 +00:00
if final_value or (arg_int and final_value == 0):
2021-09-08 03:04:36 +00:00
if arg_bool:
2022-04-18 18:16:39 +00:00
if final_value is True or final_value is False:
return final_value
elif final_value.lower() in ["t", "true"]:
2021-09-08 03:04:36 +00:00
return True
else:
return False
elif arg_int:
2023-01-04 16:43:32 +00:00
try:
return int(final_value)
except ValueError:
return default
2021-03-16 21:00:21 +00:00
else:
2022-04-18 18:16:39 +00:00
return str(final_value)
2021-03-16 21:00:21 +00:00
else:
return default
2023-08-23 13:40:08 +00:00
static_envs = []
run_args = {}
for arg_key, arg_data in arguments.items():
temp_args = arg_data["args"] if isinstance(arg_data["args"], list) else [arg_data["args"]]
final_vars = [f"PMM_{arg_key.replace('-', '_').upper()}"] + [f"PMM_{a.replace('-', '_').upper()}" for a in temp_args if len(a) > 2]
run_args[arg_key] = get_env(final_vars, getattr(args, arg_key.replace("-", "_")), arg_bool=arg_data["type"] == "bool", arg_int=arg_data["type"] == "int")
env_version = get_env("BRANCH_NAME", "master")
is_docker = get_env("PMM_DOCKER", False, arg_bool=True)
is_linuxserver = get_env("PMM_LINUXSERVER", False, arg_bool=True)
2021-09-08 03:04:36 +00:00
2023-04-07 20:03:59 +00:00
secret_args = {}
2023-04-09 06:46:27 +00:00
plex_url = None
plex_token = None
2023-04-07 20:03:59 +00:00
i = 0
while i < len(unknown):
2023-07-24 14:03:55 +00:00
test_var = str(unknown[i]).lower().replace("_", "-")
2023-04-09 06:46:27 +00:00
if test_var.startswith("--pmm-") or test_var in ["-pu", "--plex-url", "-pt", "--plex-token"]:
if test_var in ["-pu", "--plex-url"]:
plex_url = str(unknown[i + 1])
elif test_var in ["-pt", "--plex-token"]:
plex_token = str(unknown[i + 1])
else:
secret_args[test_var[6:]] = str(unknown[i + 1])
2023-04-07 20:03:59 +00:00
i += 1
i += 1
2023-08-23 13:40:08 +00:00
plex_url = get_env("PMM_PLEX_URL", plex_url)
plex_token = get_env("PMM_PLEX_TOKEN", plex_token)
2023-04-09 06:46:27 +00:00
2023-07-24 14:03:55 +00:00
env_secrets = []
2023-04-07 20:03:59 +00:00
for env_name, env_data in os.environ.items():
if str(env_name).upper().startswith("PMM_") and str(env_name).upper() not in static_envs:
2023-07-24 14:03:55 +00:00
secret_args[str(env_name).lower()[4:].replace("_", "-")] = env_data
run_arg = " ".join([f'"{s}"' if " " in s else s for s in sys.argv[:]])
2024-01-04 20:24:53 +00:00
for _, sv in secret_args.items():
if sv in run_arg:
run_arg = run_arg.replace(sv, "(redacted)")
2023-04-07 20:03:59 +00:00
2023-08-23 13:40:08 +00:00
try:
from git import Repo, InvalidGitRepositoryError # noqa
2023-08-23 13:40:08 +00:00
try:
git_branch = Repo(path=".").head.ref.name # noqa
except InvalidGitRepositoryError:
git_branch = None
except ImportError:
git_branch = None
if run_args["run-collections"]:
run_args["collections-only"] = True
2022-04-25 20:18:04 +00:00
2023-08-23 13:40:08 +00:00
if run_args["width"] < 90 or run_args["width"] > 300:
print(f"Argument Error: width argument invalid: {run_args['width']} must be an integer between 90 and 300 using the default 100")
run_args["width"] = 100
2021-09-08 03:04:36 +00:00
2023-08-23 13:40:08 +00:00
if run_args["config"] and os.path.exists(run_args["config"]):
default_dir = os.path.join(os.path.dirname(os.path.abspath(run_args["config"])))
elif run_args["config"] and not os.path.exists(run_args["config"]):
print(f"Config Error: config not found at {os.path.abspath(run_args['config'])}")
2021-09-08 03:04:36 +00:00
sys.exit(0)
elif not os.path.exists(os.path.join(default_dir, "config.yml")):
print(f"Config Error: config not found at {os.path.abspath(default_dir)}")
sys.exit(0)
2021-01-20 21:37:59 +00:00
2023-08-23 13:40:08 +00:00
logger = MyLogger("Plex Meta Manager", default_dir, run_args["width"], run_args["divider"][0], run_args["ignore-ghost"],
run_args["tests"] or run_args["debug"], run_args["trace"], run_args["log-requests"])
2021-01-20 21:37:59 +00:00
from modules import util
util.logger = logger
from modules.builder import CollectionBuilder
from modules.config import ConfigFile
2022-11-12 16:42:53 +00:00
from modules.util import Failed, FilterFailed, NonExisting, NotScheduled, Deleted
2021-01-20 21:37:59 +00:00
def my_except_hook(exctype, value, tb):
2022-06-07 20:42:17 +00:00
if issubclass(exctype, KeyboardInterrupt):
sys.__excepthook__(exctype, value, tb)
else:
logger.critical("Uncaught Exception", exc_info=(exctype, value, tb))
2021-01-20 21:37:59 +00:00
sys.excepthook = my_except_hook
2021-01-20 21:37:59 +00:00
2022-08-01 20:26:10 +00:00
old_send = requests.Session.send
def new_send(*send_args, **kwargs):
if kwargs.get("timeout", None) is None:
2023-08-23 13:40:08 +00:00
kwargs["timeout"] = run_args["timeout"]
2022-08-01 20:26:10 +00:00
return old_send(*send_args, **kwargs)
requests.Session.send = new_send
2022-03-16 21:01:39 +00:00
version = ("Unknown", "Unknown", 0)
2021-10-21 16:00:00 +00:00
with open(os.path.join(os.path.dirname(os.path.abspath(__file__)), "VERSION")) as handle:
2021-10-19 04:23:52 +00:00
for line in handle.readlines():
line = line.strip()
if len(line) > 0:
2022-03-18 13:00:42 +00:00
version = util.parse_version(line)
2021-10-19 04:23:52 +00:00
break
2023-01-19 19:38:55 +00:00
branch = util.guess_branch(version, env_version, git_branch)
version = (version[0].replace("develop", branch), version[1].replace("develop", branch), version[2])
2021-10-19 04:23:52 +00:00
2022-04-21 04:39:38 +00:00
uuid_file = os.path.join(default_dir, "UUID")
uuid_num = None
if os.path.exists(uuid_file):
with open(uuid_file) as handle:
for line in handle.readlines():
line = line.strip()
if len(line) > 0:
uuid_num = line
break
if not uuid_num:
uuid_num = uuid.uuid4()
with open(uuid_file, "w") as handle:
handle.write(str(uuid_num))
plexapi.BASE_HEADERS["X-Plex-Client-Identifier"] = str(uuid_num)
2022-07-10 08:04:06 +00:00
ImageFile.LOAD_TRUNCATED_IMAGES = True
2022-02-08 06:19:35 +00:00
2022-06-07 20:42:17 +00:00
def process(attrs):
with ProcessPoolExecutor(max_workers=1) as executor:
executor.submit(start, *[attrs])
2021-10-04 17:51:32 +00:00
def start(attrs):
logger.add_main_handler()
logger.separator()
2021-10-19 04:23:52 +00:00
logger.info("")
logger.info_center(" ____ _ __ __ _ __ __ ")
logger.info_center("| _ \\| | _____ __ | \\/ | ___| |_ __ _ | \\/ | __ _ _ __ __ _ __ _ ___ _ __ ")
logger.info_center("| |_) | |/ _ \\ \\/ / | |\\/| |/ _ \\ __/ _` | | |\\/| |/ _` | '_ \\ / _` |/ _` |/ _ \\ '__|")
logger.info_center("| __/| | __/> < | | | | __/ || (_| | | | | | (_| | | | | (_| | (_| | __/ | ")
logger.info_center("|_| |_|\\___/_/\\_\\ |_| |_|\\___|\\__\\__,_| |_| |_|\\__,_|_| |_|\\__,_|\\__, |\\___|_| ")
logger.info_center(" |___/ ")
2022-12-07 19:44:15 +00:00
system_ver = "Docker" if is_docker else "Linuxserver" if is_linuxserver else f"Python {platform.python_version()}"
2023-01-19 19:38:55 +00:00
logger.info(f" Version: {version[0]} ({system_ver}){f' (Git: {git_branch})' if git_branch else ''}")
2023-01-24 14:29:25 +00:00
latest_version = util.current_version(version, branch=branch)
2022-03-18 13:00:42 +00:00
new_version = latest_version[0] if latest_version and (version[1] != latest_version[1] or (version[2] and version[2] < latest_version[2])) else None
2022-03-16 21:01:39 +00:00
if new_version:
logger.info(f" Newest Version: {new_version}")
2023-11-30 18:25:28 +00:00
required_version = None
if not is_docker and not is_linuxserver:
try:
with open("requirements.txt", "r") as file:
2024-01-04 20:24:53 +00:00
required_version = next(ln.strip()[9:] for ln in file.readlines() if ln.strip().startswith("PlexAPI=="))
2023-11-30 18:25:28 +00:00
except FileNotFoundError:
logger.error(" File Error: requirements.txt not found")
2023-11-30 16:57:09 +00:00
logger.info(f" PlexAPI Version: {plexapi.VERSION}")
2023-11-30 18:25:28 +00:00
if required_version is not None and required_version != plexapi.VERSION:
2023-11-30 16:57:09 +00:00
logger.info(f" PlexAPI Requires an Update to Version: {required_version}")
2022-11-19 20:17:43 +00:00
logger.info(f" Platform: {platform.platform()}")
logger.info(f" Memory: {round(psutil.virtual_memory().total / (1024.0 ** 3))} GB")
2021-10-07 15:17:20 +00:00
if "time" in attrs and attrs["time"]: start_type = f"{attrs['time']} "
2023-08-23 13:40:08 +00:00
elif run_args["tests"]: start_type = "Test "
2021-10-07 15:17:20 +00:00
elif "collections" in attrs and attrs["collections"]: start_type = "Collections "
elif "libraries" in attrs and attrs["libraries"]: start_type = "Libraries "
else: start_type = ""
2021-02-28 03:56:49 +00:00
start_time = datetime.now()
2021-10-04 17:51:32 +00:00
if "time" not in attrs:
attrs["time"] = start_time.strftime("%H:%M")
2021-12-01 05:38:20 +00:00
attrs["time_obj"] = start_time
2022-03-18 13:00:42 +00:00
attrs["version"] = version
2023-01-24 14:29:25 +00:00
attrs["branch"] = branch
2023-08-23 13:40:08 +00:00
attrs["config_file"] = run_args["config"]
attrs["ignore_schedules"] = run_args["ignore-schedules"]
attrs["read_only"] = run_args["read-only-config"]
attrs["no_missing"] = run_args["no-missing"]
attrs["no_report"] = run_args["no-report"]
attrs["collection_only"] = run_args["collections-only"]
attrs["metadata_only"] = run_args["metadata-only"]
2023-08-23 13:40:08 +00:00
attrs["playlist_only"] = run_args["playlists-only"]
attrs["operations_only"] = run_args["operations-only"]
attrs["overlays_only"] = run_args["overlays-only"]
2022-12-20 01:04:14 +00:00
attrs["plex_url"] = plex_url
attrs["plex_token"] = plex_token
logger.separator(debug=True)
logger.debug(f"Run Command: {run_arg}")
2023-08-23 13:40:08 +00:00
for akey, adata in arguments.items():
2023-11-30 22:08:14 +00:00
if isinstance(adata["help"], str):
ext = '"' if adata["type"] == "str" and run_args[akey] not in [None, "None"] else ""
logger.debug(f"--{akey} (PMM_{akey.upper()}): {ext}{run_args[akey]}{ext}")
logger.debug("")
if secret_args:
logger.debug("PMM Secrets Read:")
for sec in secret_args:
logger.debug(f"--pmm-{sec} (PMM_{sec.upper().replace('-', '_')}): (redacted)")
logger.debug("")
logger.separator(f"Starting {start_type}Run")
2021-11-03 14:36:11 +00:00
config = None
2022-05-28 05:32:02 +00:00
stats = {"created": 0, "modified": 0, "deleted": 0, "added": 0, "unchanged": 0, "removed": 0, "radarr": 0, "sonarr": 0, "names": []}
2021-01-20 21:37:59 +00:00
try:
2023-04-07 20:03:59 +00:00
config = ConfigFile(default_dir, attrs, secret_args)
2021-01-20 21:37:59 +00:00
except Exception as e:
logger.stacktrace()
logger.critical(e)
2021-10-04 17:51:32 +00:00
else:
try:
2022-05-28 05:32:02 +00:00
stats = run_config(config, stats)
2021-10-04 17:51:32 +00:00
except Exception as e:
config.notify(e)
logger.stacktrace()
logger.critical(e)
2021-01-20 21:37:59 +00:00
logger.info("")
2021-12-01 05:38:20 +00:00
end_time = datetime.now()
2022-03-25 20:13:51 +00:00
run_time = str(end_time - start_time).split(".")[0]
2021-11-03 14:36:11 +00:00
if config:
2021-11-28 07:14:35 +00:00
try:
2022-03-18 13:00:42 +00:00
config.Webhooks.end_time_hooks(start_time, end_time, run_time, stats)
2021-11-28 07:14:35 +00:00
except Failed as e:
logger.stacktrace()
2021-11-28 07:14:35 +00:00
logger.error(f"Webhooks Error: {e}")
2022-03-16 21:01:39 +00:00
version_line = f"Version: {version[0]}"
if new_version:
version_line = f"{version_line} Newest Version: {new_version}"
2023-01-09 21:22:39 +00:00
try:
log_data = {}
2024-01-04 20:24:53 +00:00
no_overlays = []
no_overlays_count = 0
convert_errors = {}
other_log_groups = [
("No Items found for", r"No Items found for .* \(\d+\) (.*)"),
("Convert Warning: No TVDb ID or IMDb ID found for AniDB ID:", r"Convert Warning: No TVDb ID or IMDb ID found for AniDB ID: (.*)"),
("Convert Warning: No AniDB ID Found for AniList ID:", r"Convert Warning: No AniDB ID Found for AniList ID: (.*)"),
("Convert Warning: No AniDB ID Found for MyAnimeList ID:", r"Convert Warning: No AniDB ID Found for MyAnimeList ID: (.*)"),
("Convert Warning: No IMDb ID Found for TMDb ID:", r"Convert Warning: No IMDb ID Found for TMDb ID: (.*)"),
("Convert Warning: No TMDb ID Found for IMDb ID:", r"Convert Warning: No TMDb ID Found for IMDb ID: (.*)"),
("Convert Warning: No TVDb ID Found for TMDb ID:", r"Convert Warning: No TVDb ID Found for TMDb ID: (.*)"),
("Convert Warning: No TMDb ID Found for TVDb ID:", r"Convert Warning: No TMDb ID Found for TVDb ID: (.*)"),
("Convert Warning: No IMDb ID Found for TVDb ID:", r"Convert Warning: No IMDb ID Found for TVDb ID: (.*)"),
("Convert Warning: No TVDb ID Found for IMDb ID:", r"Convert Warning: No TVDb ID Found for IMDb ID: (.*)"),
("Convert Warning: No AniDB ID to Convert to MyAnimeList ID for Guid:", r"Convert Warning: No AniDB ID to Convert to MyAnimeList ID for Guid: (.*)"),
("Convert Warning: No MyAnimeList Found for AniDB ID:", r"Convert Warning: No MyAnimeList Found for AniDB ID: (.*) of Guid: .*"),
]
other_message = {}
2023-01-09 21:22:39 +00:00
with open(logger.main_log, encoding="utf-8") as f:
for log_line in f:
for err_type in ["WARNING", "ERROR", "CRITICAL"]:
if f"[{err_type}]" in log_line:
log_line = log_line.split("|")[1].strip()
2024-01-04 20:24:53 +00:00
other = False
for key, reg in other_log_groups:
if log_line.startswith(key):
other = True
_name = re.match(reg, log_line).group(1)
if key not in other_message:
other_message[key] = {"list": [], "count": 0}
other_message[key]["count"] += 1
if _name not in other_message[key]:
other_message[key]["list"].append(_name)
if other is False:
if err_type not in log_data:
log_data[err_type] = []
log_data[err_type].append(log_line)
if "No Items found for" in other_message:
logger.separator(f"Overlay Errors Summary", space=False, border=False)
logger.info("")
logger.info(f"No Items found for {other_message['No Items found for']['count']} Overlays: {other_message['No Items found for']['list']}")
logger.info("")
convert_title = False
for key, _ in other_log_groups:
if key.startswith("Convert Warning") and key in other_message:
if convert_title is False:
logger.separator("Convert Summary", space=False, border=False)
logger.info("")
convert_title = True
logger.info(f"{key[17:]}")
logger.info(", ".join(other_message[key]["list"]))
if convert_title:
logger.info("")
2023-01-09 21:22:39 +00:00
for err_type in ["WARNING", "ERROR", "CRITICAL"]:
if err_type not in log_data:
continue
logger.separator(f"{err_type.lower().capitalize()} Summary", space=False, border=False)
logger.info("")
2023-01-11 18:59:58 +00:00
logger.info("Count | Message")
2023-01-09 21:22:39 +00:00
logger.separator(f"{logger.separating_character * 5}|", space=False, border=False, side_space=False, left=True)
for k, v in Counter(log_data[err_type]).most_common():
logger.info(f"{v:>5} | {k}")
logger.info("")
except Failed as e:
logger.stacktrace()
logger.error(f"Report Error: {e}")
2022-03-15 06:47:50 +00:00
logger.separator(f"Finished {start_type}Run\n{version_line}\nFinished: {end_time.strftime('%H:%M:%S %Y-%m-%d')} Run Time: {run_time}")
logger.remove_main_handler()
2021-01-20 21:37:59 +00:00
2022-05-28 05:32:02 +00:00
def run_config(config, stats):
library_status = run_libraries(config)
2022-05-11 04:37:42 +00:00
playlist_status = {}
playlist_stats = {}
if (config.playlist_files or config.general["playlist_report"]) and not run_args["overlays-only"] and not run_args["metadata-only"] and not run_args["operations-only"] and not run_args["collections-only"] and not config.requested_files:
2023-10-19 14:02:53 +00:00
#logger.add_playlists_handler()
2022-05-11 04:37:42 +00:00
if config.playlist_files:
playlist_status, playlist_stats = run_playlists(config)
if config.general["playlist_report"]:
ran = []
for library in config.libraries:
if library.PlexServer.machineIdentifier in ran:
continue
ran.append(library.PlexServer.machineIdentifier)
logger.info("")
logger.separator(f"{library.PlexServer.friendlyName} Playlist Report")
logger.info("")
report = library.playlist_report()
max_length = 0
for playlist_name in report:
if len(playlist_name) > max_length:
max_length = len(playlist_name)
logger.info(f"{'Playlist Title':<{max_length}} | Users")
logger.separator(f"{logger.separating_character * max_length}|", space=False, border=False, side_space=False, left=True)
for playlist_name, users in report.items():
logger.info(f"{playlist_name:<{max_length}} | {'all' if len(users) == len(library.users) + 1 else ', '.join(users)}")
2023-10-19 14:02:53 +00:00
#logger.remove_playlists_handler()
2022-05-11 04:37:42 +00:00
amount_added = 0
2023-08-23 13:40:08 +00:00
if not run_args["operations-only"] and not run_args["overlays-only"] and not run_args["playlists-only"]:
2022-05-11 04:37:42 +00:00
has_run_again = False
for library in config.libraries:
if library.run_again:
has_run_again = True
break
if has_run_again:
logger.info("")
logger.separator("Run Again")
logger.info("")
2023-12-25 08:16:04 +00:00
for x in range(0, config.general["run_again_delay"]):
logger.ghost(f"Waiting to run again in {config.general['run_again_delay'] - x} minutes")
time.sleep(60)
2022-05-11 04:37:42 +00:00
logger.exorcise()
for library in config.libraries:
if library.run_again:
try:
2023-10-19 14:02:53 +00:00
#logger.re_add_library_handler(library.mapping_name)
2022-05-11 04:37:42 +00:00
os.environ["PLEXAPI_PLEXAPI_TIMEOUT"] = str(library.timeout)
logger.info("")
logger.separator(f"{library.name} Library Run Again")
logger.info("")
2022-06-01 20:15:15 +00:00
library.map_guids(library.cache_items())
2022-05-11 04:37:42 +00:00
for builder in library.run_again:
logger.info("")
logger.separator(f"{builder.name} Collection in {library.name}")
logger.info("")
try:
amount_added += builder.run_collections_again()
except Failed as e:
library.notify(e, collection=builder.name, critical=False)
logger.stacktrace()
logger.error(e)
2023-10-19 14:02:53 +00:00
#logger.remove_library_handler(library.mapping_name)
2022-05-11 04:37:42 +00:00
except Exception as e:
library.notify(e)
logger.stacktrace()
logger.critical(e)
2023-08-23 13:40:08 +00:00
if not run_args["collections-only"] and not run_args["overlays-only"] and not run_args["playlists-only"]:
2022-05-11 04:37:42 +00:00
used_url = []
for library in config.libraries:
if library.url not in used_url:
used_url.append(library.url)
if library.empty_trash:
library.query(library.PlexServer.library.emptyTrash)
if library.clean_bundles:
library.query(library.PlexServer.library.cleanBundles)
if library.optimize:
library.query(library.PlexServer.library.optimize)
longest = 20
for library in config.libraries:
for title in library.status:
if len(title) > longest:
longest = len(title)
if playlist_status:
for title in playlist_status:
if len(title) > longest:
longest = len(title)
def print_status(status):
logger.info(f"{'Title':^{longest}} | + | = | - | Run Time | {'Status'}")
breaker = f"{logger.separating_character * longest}|{logger.separating_character * 7}|{logger.separating_character * 7}|{logger.separating_character * 7}|{logger.separating_character * 10}|"
logger.separator(breaker, space=False, border=False, side_space=False, left=True)
for name, data in status.items():
logger.info(f"{name:<{longest}} | {data['added']:>5} | {data['unchanged']:>5} | {data['removed']:>5} | {data['run_time']:>8} | {data['status']}")
if data["errors"]:
for error in data["errors"]:
logger.info(error)
logger.info("")
logger.info("")
logger.separator("Summary")
for library in config.libraries:
logger.info("")
logger.separator(f"{library.name} Summary", space=False, border=False)
logger.info("")
logger.info(f"{'Title':<27} | Run Time |")
logger.info(f"{logger.separating_character * 27} | {logger.separating_character * 8} |")
if library.name in library_status:
for text, value in library_status[library.name].items():
logger.info(f"{text:<27} | {value:>8} |")
2022-05-11 04:37:42 +00:00
logger.info("")
print_status(library.status)
if playlist_status:
logger.info("")
logger.separator(f"Playlists Summary", space=False, border=False)
logger.info("")
print_status(playlist_status)
stats["added"] += amount_added
for library in config.libraries:
stats["created"] += library.stats["created"]
stats["modified"] += library.stats["modified"]
stats["deleted"] += library.stats["deleted"]
stats["added"] += library.stats["added"]
stats["unchanged"] += library.stats["unchanged"]
stats["removed"] += library.stats["removed"]
stats["radarr"] += library.stats["radarr"]
stats["sonarr"] += library.stats["sonarr"]
stats["names"].extend([{"name": n, "library": library.name} for n in library.stats["names"]])
if playlist_stats:
stats["created"] += playlist_stats["created"]
stats["modified"] += playlist_stats["modified"]
stats["deleted"] += playlist_stats["deleted"]
stats["added"] += playlist_stats["added"]
stats["unchanged"] += playlist_stats["unchanged"]
stats["removed"] += playlist_stats["removed"]
stats["radarr"] += playlist_stats["radarr"]
stats["sonarr"] += playlist_stats["sonarr"]
stats["names"].extend([{"name": n, "library": "PLAYLIST"} for n in playlist_stats["names"]])
return stats
def run_libraries(config):
2022-05-05 13:21:17 +00:00
library_status = {}
2021-05-08 02:30:28 +00:00
for library in config.libraries:
2021-12-18 03:02:24 +00:00
if library.skip_library:
logger.info("")
2023-01-31 19:22:33 +00:00
logger.separator(f"Skipping {library.original_mapping_name} Library")
2021-12-18 03:02:24 +00:00
continue
2022-05-05 13:21:17 +00:00
library_status[library.name] = {}
2021-10-04 17:51:32 +00:00
try:
2023-10-19 14:02:53 +00:00
#logger.add_library_handler(library.mapping_name)
2021-10-26 05:40:23 +00:00
plexapi.server.TIMEOUT = library.timeout
2022-06-30 13:14:09 +00:00
os.environ["PLEXAPI_PLEXAPI_TIMEOUT"] = str(library.timeout)
logger.info("")
2023-01-31 19:22:33 +00:00
logger.separator(f"{library.original_mapping_name} Library")
logger.debug("")
2023-01-31 19:22:33 +00:00
logger.debug(f"Library Name: {library.name}")
2023-11-30 22:08:14 +00:00
logger.debug(f"Run Order: {', '.join(library.run_order)}")
logger.debug(f"Folder Name: {library.mapping_name}")
for ad in library.asset_directory:
logger.debug(f"Asset Directory: {ad}")
logger.debug(f"Asset Folders: {library.asset_folders}")
2024-01-09 16:59:17 +00:00
logger.debug(f"Asset Depth: {library.asset_depth}")
logger.debug(f"Create Asset Folders: {library.create_asset_folders}")
2024-01-09 16:59:17 +00:00
logger.debug(f"Prioritize Assets: {library.prioritize_assets}")
logger.debug(f"Dimensional Asset Rename: {library.dimensional_asset_rename}")
2022-01-11 01:23:56 +00:00
logger.debug(f"Download URL Assets: {library.download_url_assets}")
2024-01-09 16:59:17 +00:00
logger.debug(f"Show Missing Assets: {library.show_missing_assets}")
logger.debug(f"Show Missing Season Assets: {library.show_missing_season_assets}")
logger.debug(f"Show Missing Episode Assets: {library.show_missing_episode_assets}")
logger.debug(f"Show Assets Not Needed: {library.show_asset_not_needed}")
logger.debug(f"Sync Mode: {library.sync_mode}")
logger.debug(f"Minimum Items: {library.minimum_items}")
logger.debug(f"Delete Below Minimum: {library.delete_below_minimum}")
logger.debug(f"Delete Not Scheduled: {library.delete_not_scheduled}")
2022-01-11 01:23:56 +00:00
logger.debug(f"Default Collection Order: {library.default_collection_order}")
logger.debug(f"Missing Only Released: {library.missing_only_released}")
logger.debug(f"Only Filter Missing: {library.only_filter_missing}")
logger.debug(f"Show Unmanaged: {library.show_unmanaged}")
2024-01-09 16:59:17 +00:00
logger.debug(f"Show Unconfigured: {library.show_unconfigured}")
logger.debug(f"Show Filtered: {library.show_filtered}")
2024-01-09 16:59:17 +00:00
logger.debug(f"Show Options: {library.show_options}")
logger.debug(f"Show Missing: {library.show_missing}")
logger.debug(f"Save Report: {library.save_report}")
logger.debug(f"Report Path: {library.report_path}")
2024-01-09 16:59:17 +00:00
logger.debug(f"Ignore IDs: {library.ignore_ids}")
logger.debug(f"Ignore IMDb IDs: {library.ignore_imdb_ids}")
logger.debug(f"Item Refresh Delay: {library.item_refresh_delay}")
logger.debug(f"Clean Bundles: {library.clean_bundles}")
logger.debug(f"Empty Trash: {library.empty_trash}")
logger.debug(f"Optimize: {library.optimize}")
logger.debug(f"Timeout: {library.timeout}")
2023-08-23 13:40:08 +00:00
if run_args["delete-collections"] and not run_args["playlists-only"]:
2022-05-05 13:21:17 +00:00
time_start = datetime.now()
2021-12-31 04:34:04 +00:00
logger.info("")
logger.separator(f"Deleting all Collections from the {library.name} Library", space=False, border=False)
2021-12-31 04:34:04 +00:00
logger.info("")
for collection in library.get_all_collections():
2022-10-17 16:06:32 +00:00
try:
2022-11-28 22:00:54 +00:00
library.delete(collection)
2022-10-17 16:06:32 +00:00
logger.info(f"Collection {collection.title} Deleted")
2022-11-28 22:00:54 +00:00
except Failed as e:
logger.error(e)
2022-05-05 13:21:17 +00:00
library_status[library.name]["All Collections Deleted"] = str(datetime.now() - time_start).split('.')[0]
2022-04-20 21:30:31 +00:00
2023-08-23 13:40:08 +00:00
if run_args["delete-labels"] and not run_args["playlists-only"]:
2022-11-01 18:20:00 +00:00
time_start = datetime.now()
logger.info("")
logger.separator(f"Deleting all Labels from All items in the {library.name} Library", space=False, border=False)
logger.info("")
if library.is_show:
library_types = ["show", "season", "episode"]
elif library.is_music:
library_types = ["artist", "album", "track"]
else:
library_types = ["movie"]
for library_type in library_types:
for item in library.get_all(builder_level=library_type):
try:
2024-01-04 20:24:53 +00:00
sync = ["Overlay"] if "Overlay" in [lbl.tag for lbl in item.labels] else []
library.edit_tags("label", item, sync_tags=sync)
2022-11-01 18:20:00 +00:00
except NotFound:
logger.error(f"{item.title[:25]:<25} | Labels Failed to be Removed")
library_status[library.name]["All Labels Deleted"] = str(datetime.now() - time_start).split('.')[0]
2022-05-10 15:47:13 +00:00
time_start = datetime.now()
temp_items = None
list_key = None
2022-07-21 18:19:25 +00:00
if config.Cache:
2023-11-30 22:08:14 +00:00
list_key, _ = config.Cache.query_list_cache("library", library.mapping_name, 1)
2022-05-10 15:47:13 +00:00
if not temp_items:
temp_items = library.cache_items()
2022-07-21 18:19:25 +00:00
if config.Cache and list_key:
config.Cache.delete_list_ids(list_key)
2022-05-15 20:29:54 +00:00
if not library.is_music:
logger.info("")
2023-01-31 19:22:33 +00:00
logger.separator(f"Mapping {library.original_mapping_name} Library", space=False, border=False)
logger.info("")
2022-04-20 21:30:31 +00:00
library.map_guids(temp_items)
2022-05-10 15:47:13 +00:00
library_status[library.name]["Library Loading and Mapping"] = str(datetime.now() - time_start).split('.')[0]
2022-04-20 21:30:31 +00:00
2023-12-01 14:49:11 +00:00
runs = {
"metadata": all([not run_args[x] for x in ["tests", "operations-only", "overlays-only", "playlists-only", "collections-only"]]),
"collections": all([not run_args[x] for x in ["operations-only", "overlays-only", "playlists-only", "metadata-only"]]),
"operations": all([not run_args[x] for x in ["tests", "collections-only", "overlays-only", "playlists-only", "metadata-only"]]),
"overlays": all([not run_args[x] for x in ["tests", "collections-only", "operations-only", "playlists-only", "metadata-only"]]),
2023-12-01 14:49:11 +00:00
}
2023-11-30 22:08:14 +00:00
for run_type in library.run_order:
if run_type == "collections" and runs[run_type]:
time_start = datetime.now()
for metadata in library.collection_files:
metadata_name = metadata.get_file_name()
if config.requested_files and metadata_name not in config.requested_files:
logger.info("")
logger.separator(f"Skipping {metadata_name} Collection File")
continue
logger.info("")
logger.separator(f"Running {metadata_name} Collection File\n{metadata.path}")
collections_to_run = metadata.get_collections(config.requested_collections)
if run_args["resume"] and run_args["resume"] not in collections_to_run:
logger.info("")
logger.warning(f"Collection: {run_args['resume']} not in Collection File: {metadata.path}")
continue
if collections_to_run:
logger.info("")
logger.separator(f"{'Test ' if run_args['tests'] else ''}Collections")
# logger.remove_library_handler(library.mapping_name)
run_collection(config, library, metadata, collections_to_run)
# logger.re_add_library_handler(library.mapping_name)
library_status[library.name]["Library Collection Files"] = str(datetime.now() - time_start).split('.')[0]
elif run_type == "metadata" and runs[run_type]:
2023-12-01 14:49:11 +00:00
time_start = datetime.now()
for images in library.images_files:
images_name = images.get_file_name()
if config.requested_files and images_name not in config.requested_files:
2023-11-30 22:08:14 +00:00
logger.info("")
2023-12-01 14:49:11 +00:00
logger.separator(f"Skipping {images_name} Images File")
continue
logger.info("")
logger.separator(f"Running {images_name} Images File\n{images.path}")
if not run_args["tests"] and not run_args["resume"] and not run_args["collections-only"]:
try:
images.update_metadata()
except Failed as e:
library.notify(e)
logger.error(e)
library_status[library.name]["Library Images Files"] = str(datetime.now() - time_start).split('.')[0]
2022-07-14 13:46:33 +00:00
2023-12-01 14:49:11 +00:00
time_start = datetime.now()
for metadata in library.metadata_files:
metadata_name = metadata.get_file_name()
if config.requested_files and metadata_name not in config.requested_files:
2023-12-01 14:49:11 +00:00
logger.info("")
logger.separator(f"Skipping {metadata_name} Metadata File")
continue
logger.info("")
logger.separator(f"Running {metadata_name} Metadata File\n{metadata.path}")
try:
metadata.update_metadata()
except Failed as e:
library.notify(e)
logger.error(e)
2023-12-01 14:49:11 +00:00
library_status[library.name]["Library Metadata Files"] = str(datetime.now() - time_start).split('.')[0]
elif run_type == "operations" and runs[run_type] and not config.requested_files and library.library_operation:
2023-12-01 14:49:11 +00:00
library_status[library.name]["Library Operations"] = library.Operations.run_operations()
elif run_type == "overlays" and runs[run_type] and (library.overlay_files or (library.remove_overlays and not config.requested_files)):
library_status[library.name]["Library Overlay Files"] = library.Overlays.run_overlays()
2023-10-19 14:02:53 +00:00
#logger.remove_library_handler(library.mapping_name)
2021-10-04 17:51:32 +00:00
except Exception as e:
library.notify(e)
logger.stacktrace()
logger.critical(e)
2022-05-11 04:37:42 +00:00
return library_status
2021-12-30 21:29:16 +00:00
2021-05-26 14:45:33 +00:00
def run_collection(config, library, metadata, requested_collections):
2021-05-19 17:12:34 +00:00
logger.info("")
2021-05-26 04:35:54 +00:00
for mapping_name, collection_attrs in requested_collections.items():
2021-05-20 20:38:48 +00:00
collection_start = datetime.now()
2023-08-23 13:40:08 +00:00
if run_args["tests"] and ("test" not in collection_attrs or collection_attrs["test"] is not True):
2021-05-12 14:25:48 +00:00
no_template_test = True
if "template" in collection_attrs and collection_attrs["template"]:
for data_template in util.get_list(collection_attrs["template"], split=False):
if "name" in data_template \
and data_template["name"] \
and metadata.templates \
and data_template["name"] in metadata.templates \
2022-04-28 14:47:40 +00:00
and metadata.templates[data_template["name"]][0] \
and "test" in metadata.templates[data_template["name"]][0] \
and metadata.templates[data_template["name"]][0]["test"] is True:
2021-05-12 14:25:48 +00:00
no_template_test = False
if no_template_test:
continue
2023-08-23 18:01:28 +00:00
if run_args["resume"] and run_args["resume"] != mapping_name:
2021-05-19 17:12:34 +00:00
continue
2023-08-23 18:01:28 +00:00
elif run_args["resume"] == mapping_name:
run_args["resume"] = None
2021-05-12 14:25:48 +00:00
logger.info("")
logger.separator(f"Resuming Collections")
2021-05-19 17:12:34 +00:00
2021-05-19 21:30:20 +00:00
if "name_mapping" in collection_attrs and collection_attrs["name_mapping"]:
2021-05-20 20:38:48 +00:00
collection_log_name, output_str = util.validate_filename(collection_attrs["name_mapping"])
2021-05-19 21:30:20 +00:00
else:
2021-05-20 20:38:48 +00:00
collection_log_name, output_str = util.validate_filename(mapping_name)
2023-10-19 14:02:53 +00:00
#logger.add_collection_handler(library.mapping_name, collection_log_name)
2022-06-07 07:06:20 +00:00
library.status[str(mapping_name)] = {"status": "Unchanged", "errors": [], "added": 0, "unchanged": 0, "removed": 0, "radarr": 0, "sonarr": 0}
2021-05-19 21:30:20 +00:00
2021-05-19 17:12:34 +00:00
try:
builder = CollectionBuilder(config, metadata, mapping_name, collection_attrs, library=library, extra=output_str)
library.stats["names"].append(builder.name)
2023-03-30 19:51:04 +00:00
if builder.build_collection:
library.collection_names.append(builder.name)
2021-05-24 03:38:46 +00:00
logger.info("")
2023-03-30 19:51:04 +00:00
logger.separator(f"Running {builder.name} Collection", space=False, border=False)
2021-05-12 14:25:48 +00:00
if len(builder.schedule) > 0:
logger.info(builder.schedule)
2021-05-12 14:25:48 +00:00
2021-05-24 03:38:46 +00:00
if len(builder.smart_filter_details) > 0:
2021-05-29 02:33:55 +00:00
logger.info("")
logger.info(builder.smart_filter_details)
2022-11-19 20:17:43 +00:00
logger.info("")
logger.info(f"Items Found: {builder.beginning_count}")
2021-05-24 03:38:46 +00:00
2021-11-03 14:36:11 +00:00
items_added = 0
items_removed = 0
2022-01-27 21:24:50 +00:00
if not builder.smart_url and builder.builders and not builder.blank_collection:
2021-05-12 14:25:48 +00:00
logger.info("")
2021-05-19 17:12:34 +00:00
logger.info(f"Sync Mode: {'sync' if builder.sync else 'append'}")
2022-04-21 20:56:32 +00:00
for method, value in builder.builders:
logger.debug("")
logger.debug(f"Builder: {method}: {value}")
logger.info("")
2022-07-19 13:25:48 +00:00
try:
builder.filter_and_save_items(builder.gather_ids(method, value))
2023-07-24 16:55:41 +00:00
except Failed as e:
2022-07-19 13:25:48 +00:00
if builder.ignore_blank_results:
logger.warning(e)
else:
raise Failed(e)
2022-10-03 19:32:50 +00:00
builder.display_filters()
2021-05-12 14:25:48 +00:00
2022-11-07 17:05:21 +00:00
if len(builder.found_items) > 0 and len(builder.found_items) + builder.beginning_count >= builder.minimum and builder.build_collection:
2021-12-30 21:29:16 +00:00
items_added, items_unchanged = builder.add_to_collection()
library.stats["added"] += items_added
2022-06-07 07:06:20 +00:00
library.status[str(mapping_name)]["added"] = items_added
2021-12-30 21:29:16 +00:00
library.stats["unchanged"] += items_unchanged
2022-06-07 07:06:20 +00:00
library.status[str(mapping_name)]["unchanged"] = items_unchanged
2021-11-03 14:36:11 +00:00
items_removed = 0
2021-10-20 19:58:25 +00:00
if builder.sync:
2021-11-03 14:36:11 +00:00
items_removed = builder.sync_collection()
2021-12-30 21:29:16 +00:00
library.stats["removed"] += items_removed
2022-06-07 07:06:20 +00:00
library.status[str(mapping_name)]["removed"] = items_removed
2021-10-20 19:58:25 +00:00
2021-08-16 05:41:04 +00:00
if builder.do_missing and (len(builder.missing_movies) > 0 or len(builder.missing_shows) > 0):
2021-11-03 14:36:11 +00:00
radarr_add, sonarr_add = builder.run_missing()
2021-12-30 21:29:16 +00:00
library.stats["radarr"] += radarr_add
2022-06-07 07:06:20 +00:00
library.status[str(mapping_name)]["radarr"] += radarr_add
2021-12-30 21:29:16 +00:00
library.stats["sonarr"] += sonarr_add
2022-06-07 07:06:20 +00:00
library.status[str(mapping_name)]["sonarr"] += sonarr_add
2021-05-12 14:25:48 +00:00
if not builder.found_items and not builder.ignore_blank_results:
raise NonExisting(f"{builder.Type} Warning: No items found")
valid = True
2022-11-19 20:17:43 +00:00
if builder.build_collection and not builder.blank_collection and items_added + builder.beginning_count < builder.minimum:
logger.info("")
2022-10-18 13:43:09 +00:00
logger.info(f"{builder.Type} Minimum: {builder.minimum} not met for {mapping_name} Collection")
2022-05-04 16:27:17 +00:00
delete_status = f"Minimum {builder.minimum} Not Met"
valid = False
if builder.details["delete_below_minimum"] and builder.obj:
logger.info("")
logger.info(builder.delete())
2022-05-04 16:27:17 +00:00
library.stats["deleted"] += 1
delete_status = f"Deleted; {delete_status}"
2022-06-07 07:06:20 +00:00
library.status[str(mapping_name)]["status"] = delete_status
2021-09-15 03:16:59 +00:00
run_item_details = True
2022-01-27 21:24:50 +00:00
if valid and builder.build_collection and (builder.builders or builder.smart_url or builder.blank_collection):
2021-09-15 03:16:59 +00:00
try:
builder.load_collection()
2021-11-03 14:36:11 +00:00
if builder.created:
2021-12-30 21:29:16 +00:00
library.stats["created"] += 1
2022-06-07 07:06:20 +00:00
library.status[str(mapping_name)]["status"] = "Created"
2021-11-03 14:36:11 +00:00
elif items_added > 0 or items_removed > 0:
2021-12-30 21:29:16 +00:00
library.stats["modified"] += 1
2022-06-07 07:06:20 +00:00
library.status[str(mapping_name)]["status"] = "Modified"
2021-09-15 03:16:59 +00:00
except Failed:
logger.stacktrace()
2021-09-15 03:16:59 +00:00
run_item_details = False
logger.info("")
2022-10-20 20:38:53 +00:00
logger.separator(f"No {builder.Type} to Update", space=False, border=False)
2021-09-15 03:16:59 +00:00
else:
2022-05-04 16:27:17 +00:00
details_list = builder.update_details()
if details_list:
pre = ""
2022-06-07 07:06:20 +00:00
if library.status[str(mapping_name)]["status"] != "Unchanged":
pre = f"{library.status[str(mapping_name)]['status']} and "
library.status[str(mapping_name)]["status"] = f"{pre}Updated {', '.join(details_list)}"
if builder.server_preroll is not None:
library.set_server_preroll(builder.server_preroll)
logger.info("")
logger.info(f"Plex Server Movie pre-roll video updated to {builder.server_preroll}")
2022-09-29 21:03:55 +00:00
if valid and run_item_details and (builder.item_details or builder.custom_sort or builder.sync_to_trakt_list):
2021-09-15 03:16:59 +00:00
try:
builder.load_collection_items()
except Failed:
logger.info("")
logger.separator("No Items Found", space=False, border=False)
2021-09-15 03:16:59 +00:00
else:
2021-12-17 14:24:46 +00:00
if builder.item_details:
builder.update_item_details()
if builder.custom_sort:
builder.sort_collection()
if builder.sync_to_trakt_list:
builder.sync_trakt_list()
builder.send_notifications()
if builder.run_again and (len(builder.run_again_movies) > 0 or len(builder.run_again_shows) > 0):
2021-05-12 14:25:48 +00:00
library.run_again.append(builder)
2022-07-19 13:25:48 +00:00
except NonExisting as e:
logger.warning(e)
library.status[str(mapping_name)]["status"] = "Ignored"
except NotScheduled as e:
logger.info(e)
2022-05-04 16:27:17 +00:00
if str(e).endswith("and was deleted"):
2022-12-28 18:58:12 +00:00
library.notify_delete(e)
2022-05-04 16:27:17 +00:00
library.stats["deleted"] += 1
2022-06-07 07:06:20 +00:00
library.status[str(mapping_name)]["status"] = "Deleted Not Scheduled"
elif str(e).startswith("Skipped because run_definition"):
library.status[str(mapping_name)]["status"] = "Skipped Run Definition"
2022-05-04 16:27:17 +00:00
else:
2022-06-07 07:06:20 +00:00
library.status[str(mapping_name)]["status"] = "Not Scheduled"
2022-11-12 16:42:53 +00:00
except FilterFailed:
pass
2021-05-12 14:25:48 +00:00
except Failed as e:
2021-10-04 17:51:32 +00:00
library.notify(e, collection=mapping_name)
logger.stacktrace()
logger.error(e)
2022-06-07 07:06:20 +00:00
library.status[str(mapping_name)]["status"] = "PMM Failure"
library.status[str(mapping_name)]["errors"].append(e)
2021-05-12 14:25:48 +00:00
except Exception as e:
2021-10-04 17:51:32 +00:00
library.notify(f"Unknown Error: {e}", collection=mapping_name)
logger.stacktrace()
2021-05-12 14:25:48 +00:00
logger.error(f"Unknown Error: {e}")
2022-06-07 07:06:20 +00:00
library.status[str(mapping_name)]["status"] = "Unknown Error"
library.status[str(mapping_name)]["errors"].append(e)
2022-05-04 16:27:17 +00:00
collection_run_time = str(datetime.now() - collection_start).split('.')[0]
2022-06-07 07:06:20 +00:00
library.status[str(mapping_name)]["run_time"] = collection_run_time
2021-05-19 17:12:34 +00:00
logger.info("")
2022-05-04 16:27:17 +00:00
logger.separator(f"Finished {mapping_name} Collection\nCollection Run Time: {collection_run_time}")
2023-10-19 14:02:53 +00:00
#logger.remove_collection_handler(library.mapping_name, collection_log_name)
2021-05-08 02:30:28 +00:00
2021-12-23 17:02:07 +00:00
def run_playlists(config):
stats = {"created": 0, "modified": 0, "deleted": 0, "added": 0, "unchanged": 0, "removed": 0, "radarr": 0, "sonarr": 0, "names": []}
2021-12-30 21:29:16 +00:00
status = {}
2021-12-23 17:02:07 +00:00
logger.info("")
logger.separator("Playlists")
2021-12-23 17:02:07 +00:00
logger.info("")
for playlist_file in config.playlist_files:
for mapping_name, playlist_attrs in playlist_file.playlists.items():
playlist_start = datetime.now()
2023-08-23 13:40:08 +00:00
if run_args["tests"] and ("test" not in playlist_attrs or playlist_attrs["test"] is not True):
2021-12-23 17:02:07 +00:00
no_template_test = True
if "template" in playlist_attrs and playlist_attrs["template"]:
for data_template in util.get_list(playlist_attrs["template"], split=False):
if "name" in data_template \
and data_template["name"] \
and playlist_file.templates \
and data_template["name"] in playlist_file.templates \
2022-04-28 14:47:40 +00:00
and playlist_file.templates[data_template["name"]][0] \
and "test" in playlist_file.templates[data_template["name"]][0] \
and playlist_file.templates[data_template["name"]][0]["test"] is True:
2021-12-23 17:02:07 +00:00
no_template_test = False
if no_template_test:
continue
if "name_mapping" in playlist_attrs and playlist_attrs["name_mapping"]:
playlist_log_name, output_str = util.validate_filename(playlist_attrs["name_mapping"])
else:
playlist_log_name, output_str = util.validate_filename(mapping_name)
2023-10-19 14:02:53 +00:00
#logger.add_playlist_handler(playlist_log_name)
2022-05-04 16:27:17 +00:00
status[mapping_name] = {"status": "Unchanged", "errors": [], "added": 0, "unchanged": 0, "removed": 0, "radarr": 0, "sonarr": 0}
2021-12-23 17:02:07 +00:00
server_name = None
try:
builder = CollectionBuilder(config, playlist_file, mapping_name, playlist_attrs, extra=output_str)
stats["names"].append(builder.name)
2021-12-23 17:02:07 +00:00
logger.info("")
2024-01-26 17:16:10 +00:00
server_name = builder.libraries[0].PlexServer.friendlyName
2021-12-23 17:02:07 +00:00
logger.separator(f"Running {mapping_name} Playlist", space=False, border=False)
2021-12-23 17:02:07 +00:00
if len(builder.schedule) > 0:
logger.info(builder.schedule)
2021-12-23 17:02:07 +00:00
items_added = 0
items_removed = 0
valid = True
logger.info("")
logger.info(f"Sync Mode: {'sync' if builder.sync else 'append'}")
method, value = builder.builders[0]
logger.debug("")
logger.debug(f"Builder: {method}: {value}")
logger.info("")
2022-09-08 20:11:11 +00:00
if method == "plex_watchlist":
ids = builder.libraries[0].get_rating_keys(method, value, True)
elif "plex" in method:
2022-01-24 22:36:40 +00:00
ids = []
2022-02-08 22:57:36 +00:00
for pl_library in builder.libraries:
2022-09-08 16:07:15 +00:00
try:
2022-09-08 20:11:11 +00:00
ids.extend(pl_library.get_rating_keys(method, value, True))
2022-09-08 16:07:15 +00:00
except Failed as e:
if builder.validate_builders:
raise
else:
logger.error(e)
2022-01-24 22:36:40 +00:00
elif "tautulli" in method:
ids = []
2022-02-08 22:57:36 +00:00
for pl_library in builder.libraries:
2022-09-08 16:07:15 +00:00
try:
ids.extend(pl_library.Tautulli.get_rating_keys(value, True))
except Failed as e:
if builder.validate_builders:
raise
else:
logger.error(e)
2022-01-24 22:36:40 +00:00
else:
ids = builder.gather_ids(method, value)
2021-12-23 17:02:07 +00:00
2022-10-03 19:32:50 +00:00
builder.display_filters()
2022-02-08 22:57:36 +00:00
builder.filter_and_save_items(ids)
2021-12-23 17:02:07 +00:00
2022-11-07 17:05:21 +00:00
if len(builder.found_items) > 0 and len(builder.found_items) + builder.beginning_count >= builder.minimum:
2021-12-30 21:29:16 +00:00
items_added, items_unchanged = builder.add_to_collection()
2021-12-23 17:02:07 +00:00
stats["added"] += items_added
2021-12-30 21:29:16 +00:00
status[mapping_name]["added"] += items_added
stats["unchanged"] += items_unchanged
status[mapping_name]["unchanged"] += items_unchanged
2021-12-23 17:02:07 +00:00
items_removed = 0
if builder.sync:
items_removed = builder.sync_collection()
stats["removed"] += items_removed
2021-12-30 21:29:16 +00:00
status[mapping_name]["removed"] += items_removed
2022-11-07 17:05:21 +00:00
elif len(builder.found_items) < builder.minimum:
2021-12-23 17:02:07 +00:00
logger.info("")
logger.info(f"Playlist Minimum: {builder.minimum} not met for {mapping_name} Playlist")
2022-05-04 16:27:17 +00:00
delete_status = f"Minimum {builder.minimum} Not Met"
2021-12-23 17:02:07 +00:00
valid = False
if builder.details["delete_below_minimum"] and builder.obj:
2021-12-26 15:48:52 +00:00
logger.info("")
logger.info(builder.delete())
2022-05-04 16:27:17 +00:00
stats["deleted"] += 1
delete_status = f"Deleted; {delete_status}"
status[mapping_name]["status"] = delete_status
2021-12-23 17:02:07 +00:00
if builder.do_missing and (len(builder.missing_movies) > 0 or len(builder.missing_shows) > 0):
radarr_add, sonarr_add = builder.run_missing()
stats["radarr"] += radarr_add
2021-12-30 21:29:16 +00:00
status[mapping_name]["radarr"] += radarr_add
2021-12-23 17:02:07 +00:00
stats["sonarr"] += sonarr_add
2021-12-30 21:29:16 +00:00
status[mapping_name]["sonarr"] += sonarr_add
2021-12-23 17:02:07 +00:00
run_item_details = True
2022-05-04 16:27:17 +00:00
if valid and builder.builders:
try:
builder.load_collection()
if builder.created:
stats["created"] += 1
status[mapping_name]["status"] = "Created"
elif items_added > 0 or items_removed > 0:
stats["modified"] += 1
status[mapping_name]["status"] = "Modified"
except Failed:
logger.stacktrace()
run_item_details = False
logger.info("")
logger.separator("No Playlist to Update", space=False, border=False)
else:
details_list = builder.update_details()
if details_list:
pre = ""
if status[mapping_name]["status"] != "Unchanged":
pre = f"{status[mapping_name]['status']} and "
status[mapping_name]["status"] = f"{pre}Updated {', '.join(details_list)}"
2021-12-23 17:02:07 +00:00
if valid and run_item_details and builder.builders and (builder.item_details or builder.custom_sort):
try:
builder.load_collection_items()
except Failed:
logger.info("")
logger.separator("No Items Found", space=False, border=False)
2021-12-23 17:02:07 +00:00
else:
if builder.item_details:
builder.update_item_details()
if builder.custom_sort:
builder.sort_collection()
if valid:
2021-12-26 15:48:52 +00:00
builder.sync_playlist()
2021-12-23 17:02:07 +00:00
builder.send_notifications(playlist=True)
2022-05-28 16:01:51 +00:00
except Deleted as e:
logger.info(e)
2022-06-21 06:12:47 +00:00
status[mapping_name]["status"] = "Deleted"
2024-01-26 17:16:10 +00:00
config.notify_delete(e, server=server_name)
2021-12-23 17:02:07 +00:00
except NotScheduled as e:
logger.info(e)
2022-05-04 16:27:17 +00:00
if str(e).endswith("and was deleted"):
stats["deleted"] += 1
status[mapping_name]["status"] = "Deleted Not Scheduled"
2022-12-28 18:58:12 +00:00
config.notify_delete(e)
2022-05-04 16:27:17 +00:00
else:
status[mapping_name]["status"] = "Not Scheduled"
2021-12-23 17:02:07 +00:00
except Failed as e:
2024-01-26 17:16:10 +00:00
config.notify(e, server=server_name, playlist=mapping_name)
logger.stacktrace()
logger.error(e)
2021-12-30 21:29:16 +00:00
status[mapping_name]["status"] = "PMM Failure"
status[mapping_name]["errors"].append(e)
2021-12-23 17:02:07 +00:00
except Exception as e:
2024-01-26 17:16:10 +00:00
config.notify(f"Unknown Error: {e}", server=server_name, playlist=mapping_name)
logger.stacktrace()
2021-12-23 17:02:07 +00:00
logger.error(f"Unknown Error: {e}")
2021-12-30 21:29:16 +00:00
status[mapping_name]["status"] = "Unknown Error"
status[mapping_name]["errors"].append(e)
2021-12-23 17:02:07 +00:00
logger.info("")
2022-05-04 16:27:17 +00:00
playlist_run_time = str(datetime.now() - playlist_start).split('.')[0]
status[mapping_name]["run_time"] = playlist_run_time
logger.info("")
logger.separator(f"Finished {mapping_name} Playlist\nPlaylist Run Time: {playlist_run_time}")
2023-10-19 14:02:53 +00:00
#logger.remove_playlist_handler(playlist_log_name)
2021-12-30 21:29:16 +00:00
return status, stats
2021-12-23 17:02:07 +00:00
2022-06-07 20:42:17 +00:00
if __name__ == "__main__":
try:
if run_args["run"] or run_args["tests"] or run_args["run-collections"] or run_args["run-libraries"] or run_args["run-files"] or run_args["resume"]:
process({"collections": run_args["run-collections"], "libraries": run_args["run-libraries"], "files": run_args["run-files"]})
2022-06-07 20:42:17 +00:00
else:
2023-08-23 13:40:08 +00:00
times_to_run = util.get_list(run_args["times"])
2022-06-07 20:42:17 +00:00
valid_times = []
for time_to_run in times_to_run:
try:
final_time = datetime.strftime(datetime.strptime(time_to_run, "%H:%M"), "%H:%M")
if final_time not in valid_times:
valid_times.append(final_time)
2022-06-07 20:42:17 +00:00
except ValueError:
if time_to_run:
raise Failed(f"Argument Error: time argument invalid: {time_to_run} must be in the HH:MM format between 00:00-23:59")
else:
raise Failed(f"Argument Error: blank time argument")
for time_to_run in valid_times:
2023-08-23 13:40:08 +00:00
schedule.every().day.at(time_to_run).do(process, {"time": time_to_run})
2022-06-07 20:42:17 +00:00
while True:
schedule.run_pending()
2023-08-23 13:40:08 +00:00
if not run_args["no-countdown"]:
2022-06-07 20:42:17 +00:00
current_time = datetime.now().strftime("%H:%M")
seconds = None
og_time_str = ""
for time_to_run in valid_times:
new_seconds = (datetime.strptime(time_to_run, "%H:%M") - datetime.strptime(current_time, "%H:%M")).total_seconds()
if new_seconds < 0:
new_seconds += 86400
if (seconds is None or new_seconds < seconds) and new_seconds > 0:
seconds = new_seconds
og_time_str = time_to_run
if seconds is not None:
hours = int(seconds // 3600)
minutes = int((seconds % 3600) // 60)
time_str = f"{hours} Hour{'s' if hours > 1 else ''} and " if hours > 0 else ""
time_str += f"{minutes} Minute{'s' if minutes > 1 else ''}"
logger.ghost(f"Current Time: {current_time} | {time_str} until the next run at {og_time_str} | Runs: {', '.join(valid_times)}")
2022-06-07 20:42:17 +00:00
else:
logger.error(f"Time Error: {valid_times}")
time.sleep(60)
except KeyboardInterrupt:
logger.separator("Exiting Plex Meta Manager")