2023-04-28 03:43:26 +00:00
import math , operator , os , re
2021-05-07 05:06:40 +00:00
from datetime import datetime
2024-06-05 18:44:33 +00:00
from modules import plex , ergast , util , letterboxd
2024-05-28 20:22:51 +00:00
from modules . request import quote
from modules . util import Failed , NotScheduled
2022-03-25 20:13:51 +00:00
from plexapi . exceptions import NotFound , BadRequest
2021-05-07 05:06:40 +00:00
2022-02-13 16:33:57 +00:00
logger = util . logger
2021-05-07 05:06:40 +00:00
2022-05-19 16:09:28 +00:00
all_auto = [ " genre " , " number " , " custom " ]
2022-03-11 15:33:22 +00:00
ms_auto = [
2022-03-30 21:24:58 +00:00
" actor " , " year " , " content_rating " , " original_language " , " tmdb_popular_people " , " trakt_user_lists " , " studio " ,
2024-01-02 13:51:40 +00:00
" trakt_liked_lists " , " trakt_people_list " , " subtitle_language " , " audio_language " , " resolution " , " decade " , " imdb_awards "
2022-03-11 15:33:22 +00:00
]
2022-01-28 21:59:26 +00:00
auto = {
2024-06-05 18:44:33 +00:00
" Movie " : [ " tmdb_collection " , " edition " , " country " , " director " , " producer " , " writer " , " letterboxd_user_lists " ] + all_auto + ms_auto ,
2023-01-18 19:20:49 +00:00
" Show " : [ " network " , " origin_country " , " episode_year " ] + all_auto + ms_auto ,
2022-12-02 17:08:01 +00:00
" Artist " : [ " mood " , " style " , " country " , " album_genre " , " album_mood " , " album_style " , " track_mood " ] + all_auto ,
2022-03-11 15:33:22 +00:00
" Video " : [ " country " , " content_rating " ] + all_auto
2022-01-28 21:59:26 +00:00
}
2022-04-14 13:29:54 +00:00
dynamic_attributes = [
2023-04-09 06:46:27 +00:00
" type " , " data " , " exclude " , " addons " , " template " , " template_variables " , " other_template " , " remove_suffix " , " custom_keys " ,
2022-04-14 13:29:54 +00:00
" remove_prefix " , " title_format " , " key_name_override " , " title_override " , " test " , " sync " , " include " , " other_name "
]
2022-07-18 15:44:16 +00:00
auto_type_translation = {
" content_rating " : " contentRating " , " subtitle_language " : " subtitleLanguage " , " audio_language " : " audioLanguage " ,
2023-01-18 19:20:49 +00:00
" album_genre " : " album.genre " , " album_style " : " album.style " , " album_mood " : " album.mood " , " track_mood " : " track.mood " ,
" edition " : " editionTitle " , " episode_year " : " episode.year "
2022-07-18 15:44:16 +00:00
}
2022-01-31 08:37:49 +00:00
default_templates = {
2022-03-26 02:56:56 +00:00
" original_language " : { " plex_all " : True , " filters " : { " original_language " : " <<value>> " } } ,
" origin_country " : { " plex_all " : True , " filters " : { " origin_country " : " <<value>> " } } ,
" tmdb_collection " : { " tmdb_collection_details " : " <<value>> " , " minimum_items " : 2 } ,
" trakt_user_lists " : { " trakt_list_details " : " <<value>> " } ,
" trakt_liked_lists " : { " trakt_list_details " : " <<value>> " } ,
2024-06-05 18:44:33 +00:00
" letterboxd_user_lists " : { " letterboxd_list_details " : " <<value>> " } ,
2022-11-15 21:30:46 +00:00
" tmdb_popular_people " : { " tmdb_person " : " <<value>> " , " plex_search " : { " all " : { " actor " : " tmdb " } } } ,
" trakt_people_list " : { " tmdb_person " : " <<value>> " , " plex_search " : { " all " : { " actor " : " tmdb " } } }
2022-01-31 08:37:49 +00:00
}
2022-01-28 21:59:26 +00:00
2022-05-09 20:40:39 +00:00
def get_dict ( attribute , attr_data , check_list = None , make_str = False ) :
2021-12-21 21:33:37 +00:00
if check_list is None :
check_list = [ ]
if attr_data and attribute in attr_data :
if attr_data [ attribute ] :
if isinstance ( attr_data [ attribute ] , dict ) :
new_dict = { }
for _name , _data in attr_data [ attribute ] . items ( ) :
2022-05-09 20:40:39 +00:00
if make_str and str ( _name ) in check_list or not make_str and _name in check_list :
new_name = f ' " { str ( _name ) } " ' if make_str or not isinstance ( _name , int ) else _name
logger . warning ( f " Config Warning: Skipping duplicate { attribute [ : - 1 ] if attribute [ - 1 ] == ' s ' else attribute } : { new_name } " )
2021-12-21 21:33:37 +00:00
elif _data is None :
2022-05-28 05:32:02 +00:00
continue
2022-06-03 17:46:38 +00:00
elif attribute != " queues " and not isinstance ( _data , dict ) :
2022-01-27 15:22:58 +00:00
logger . warning ( f " Config Warning: { attribute [ : - 1 ] if attribute [ - 1 ] == ' s ' else attribute } : { _name } must be a dictionary " )
2022-04-28 14:47:40 +00:00
elif attribute == " templates " :
new_dict [ str ( _name ) ] = ( _data , { } )
2021-12-21 21:33:37 +00:00
else :
2022-05-09 20:40:39 +00:00
new_dict [ str ( _name ) if make_str else _name ] = _data
2021-12-21 21:33:37 +00:00
return new_dict
else :
2022-01-25 07:45:31 +00:00
logger . error ( f " Config Error: { attribute } must be a dictionary " )
2021-12-21 21:33:37 +00:00
else :
2022-01-25 07:45:31 +00:00
logger . error ( f " Config Error: { attribute } attribute is blank " )
2022-01-28 18:36:21 +00:00
return { }
2021-12-21 21:33:37 +00:00
class DataFile :
2024-02-28 20:45:11 +00:00
def __init__ ( self , config , file_type , path , temp_vars , asset_directory , data_type ) :
if file_type != " Data " :
logger . info ( " " )
logger . info ( f " Loading { data_type } { file_type } : { path } " )
logger . info ( " " )
2021-06-02 15:18:37 +00:00
self . config = config
2022-04-03 21:22:58 +00:00
self . library = None
2021-05-07 05:06:40 +00:00
self . type = file_type
self . path = path
2022-04-09 03:04:33 +00:00
self . temp_vars = temp_vars
2023-03-30 19:51:04 +00:00
self . language = " en "
if " language " in self . temp_vars and self . temp_vars [ " language " ] :
if self . temp_vars [ " language " ] . lower ( ) not in self . config . GitHub . translation_keys :
logger . warning ( f " Config Error: Language: { self . temp_vars [ ' language ' ] . lower ( ) } Not Found using en. Options: { ' , ' . join ( self . config . GitHub . translation_keys ) } " )
else :
self . language = self . temp_vars [ " language " ] . lower ( )
2022-04-21 05:35:07 +00:00
self . asset_directory = asset_directory
2021-12-21 21:33:37 +00:00
self . data_type = " "
self . templates = { }
2023-12-12 22:03:15 +00:00
filename = self . get_file_name ( )
if config . requested_files and filename not in config . requested_files :
raise NotScheduled ( filename )
2021-12-21 21:33:37 +00:00
2022-01-24 15:57:48 +00:00
def get_file_name ( self ) :
2022-07-13 05:13:25 +00:00
data = f " { self . config . GitHub . configs_url } { self . path } .yml " if self . type == " GIT " else self . path
2022-01-24 15:57:48 +00:00
if " / " in data :
2022-06-26 19:08:17 +00:00
if data . endswith ( " .yml " ) :
return data [ data . rfind ( " / " ) + 1 : - 4 ]
2022-10-02 17:19:32 +00:00
elif data . endswith ( " .yaml " ) :
return data [ data . rfind ( " / " ) + 1 : - 5 ]
2022-06-26 19:08:17 +00:00
else :
return data [ data . rfind ( " / " ) + 1 : ]
2022-01-24 15:57:48 +00:00
elif " \\ " in data :
2022-06-26 19:08:17 +00:00
if data . endswith ( " .yml " ) :
return data [ data . rfind ( " \\ " ) + 1 : - 4 ]
2022-10-02 17:19:32 +00:00
elif data . endswith ( " .yaml " ) :
return data [ data . rfind ( " / " ) + 1 : - 5 ]
2022-06-26 19:08:17 +00:00
else :
return data [ data . rfind ( " \\ " ) + 1 : ]
2022-01-24 15:57:48 +00:00
else :
return data
2023-03-04 20:20:52 +00:00
def load_file ( self , file_type , file_path , overlay = False , translation = False , images = False , folder = " " ) :
2022-10-02 17:19:32 +00:00
if translation :
if file_path . endswith ( " .yml " ) :
file_path = file_path [ : - 4 ]
elif file_path . endswith ( " .yaml " ) :
file_path = file_path [ : - 5 ]
if not translation and not file_path . endswith ( ( " .yml " , " .yaml " ) ) :
2022-09-27 06:19:29 +00:00
file_path = f " { file_path } .yml "
2024-04-22 14:20:12 +00:00
if file_type in [ " URL " , " Git " , " Repo " ] or ( images and file_type == " Default " ) :
2022-05-12 19:10:03 +00:00
if file_type == " Repo " and not self . config . custom_repo :
raise Failed ( " Config Error: No custom_repo defined " )
2023-03-04 20:20:52 +00:00
if file_type == " URL " :
content_path = file_path
elif file_type == " Repo " :
content_path = f " { self . config . custom_repo } { file_path } "
2024-04-22 14:20:12 +00:00
elif file_type == " Default " :
2023-03-04 20:20:52 +00:00
content_path = f " { self . config . GitHub . images_raw_url } { folder } { file_path } "
else :
content_path = f " { self . config . GitHub . configs_url } { file_path } "
2022-09-29 21:03:55 +00:00
dir_path = content_path
if translation :
content_path = f " { content_path } /default.yml "
2024-05-28 20:22:51 +00:00
yaml = self . config . Requests . get_yaml ( content_path , check_empty = True )
2022-05-12 19:10:03 +00:00
else :
2024-04-22 14:20:12 +00:00
if file_type == " Default " :
2023-02-24 21:48:08 +00:00
if not overlay and file_path . startswith ( ( " movie/ " , " chart/ " , " award/ " ) ) :
2022-09-27 06:19:29 +00:00
file_path = file_path [ 6 : ]
2023-02-24 21:48:08 +00:00
elif not overlay and file_path . startswith ( ( " show/ " , " both/ " ) ) :
2022-09-27 06:19:29 +00:00
file_path = file_path [ 5 : ]
2023-02-24 21:48:08 +00:00
elif overlay and file_path . startswith ( " overlays/ " ) :
2022-09-27 06:19:29 +00:00
file_path = file_path [ 9 : ]
defaults_path = os . path . join ( os . path . dirname ( os . path . dirname ( os . path . abspath ( __file__ ) ) ) , " defaults " )
if overlay :
defaults_path = os . path . join ( defaults_path , " overlays " )
2023-02-24 15:52:23 +00:00
if os . path . exists ( os . path . join ( defaults_path , file_path ) ) :
file_path = os . path . join ( defaults_path , file_path )
2022-10-03 15:24:58 +00:00
elif self . library :
for default_folder in [ self . library . type . lower ( ) , " both " , " chart " , " award " ] :
2023-02-24 15:52:23 +00:00
if os . path . exists ( os . path . join ( defaults_path , default_folder , file_path ) ) :
file_path = os . path . join ( defaults_path , default_folder , file_path )
2022-10-02 03:03:58 +00:00
break
2023-02-13 23:13:54 +00:00
content_path = os . path . abspath ( os . path . join ( file_path , " default.yml " ) if translation else file_path )
2022-09-29 21:03:55 +00:00
dir_path = file_path
if not os . path . exists ( content_path ) :
2024-04-22 14:20:12 +00:00
if file_type == " Default " :
2022-10-04 19:24:01 +00:00
raise Failed ( f " File Error: Default does not exist { file_path } " )
else :
raise Failed ( f " File Error: File does not exist { content_path } " )
2024-05-28 20:22:51 +00:00
yaml = self . config . Requests . file_yaml ( content_path , check_empty = True )
2022-09-29 21:03:55 +00:00
if not translation :
2022-09-30 15:56:13 +00:00
logger . debug ( f " File Loaded From: { content_path } " )
2022-09-29 21:03:55 +00:00
return yaml . data
if " translations " not in yaml . data :
raise Failed ( f " URL Error: Top Level translations attribute not found in { content_path } " )
2022-10-02 03:03:58 +00:00
translations = { k : { " default " : v } for k , v in yaml . data [ " translations " ] . items ( ) }
2022-10-05 14:49:39 +00:00
lib_type = self . library . type . lower ( ) if self . library else " item "
2023-01-16 20:32:14 +00:00
logger . trace ( f " Translations Loaded From: { dir_path } " )
2022-10-20 13:32:09 +00:00
key_names = { }
variables = { k : { " default " : v [ lib_type ] } for k , v in yaml . data [ " variables " ] . items ( ) }
2022-09-29 21:03:55 +00:00
2024-05-28 20:22:51 +00:00
def add_translation ( yaml_path , yaml_key , url = False ) :
if url :
yaml_content = self . config . Requests . get_yaml ( yaml_path , check_empty = True )
else :
yaml_content = self . config . Requests . file_yaml ( yaml_path , check_empty = True )
2022-10-05 14:49:39 +00:00
if " variables " in yaml_content . data and yaml_content . data [ " variables " ] :
2022-10-05 17:27:24 +00:00
for var_key , var_value in yaml_content . data [ " variables " ] . items ( ) :
2022-10-05 14:49:39 +00:00
if lib_type in var_value :
2022-10-20 13:32:09 +00:00
if var_key not in variables :
variables [ var_key ] = { }
variables [ var_key ] [ yaml_key ] = var_value [ lib_type ]
2022-10-05 14:49:39 +00:00
if " translations " in yaml_content . data and yaml_content . data [ " translations " ] :
2022-12-23 07:45:51 +00:00
for translation_key in translations :
if translation_key in yaml_content . data [ " translations " ] :
translations [ translation_key ] [ yaml_key ] = yaml_content . data [ " translations " ] [ translation_key ]
2022-09-29 21:03:55 +00:00
else :
2023-01-16 20:32:14 +00:00
logger . trace ( f " Translation Error: translations attribute { translation_key } not found in { yaml_path } " )
2022-09-29 21:03:55 +00:00
else :
2023-01-16 20:32:14 +00:00
logger . trace ( f " Config Error: Top Level translations attribute not found in { yaml_path } " )
2022-12-23 07:45:51 +00:00
2022-10-03 15:01:38 +00:00
if " key_names " in yaml_content . data and yaml_content . data [ " key_names " ] :
2022-10-02 00:14:55 +00:00
for kn , vn in yaml_content . data [ " key_names " ] . items ( ) :
2022-10-05 14:49:39 +00:00
if kn not in key_names :
2022-10-02 00:14:55 +00:00
key_names [ kn ] = { }
key_names [ kn ] [ yaml_key ] = vn
2022-09-29 21:03:55 +00:00
if file_type in [ " URL " , " Git " , " Repo " ] :
if " languages " in yaml . data and isinstance ( yaml . data [ " language " ] , list ) :
for language in yaml . data [ " language " ] :
2024-05-28 20:22:51 +00:00
try :
add_translation ( f " { dir_path } / { language } .yml " , language , url = True )
except Failed :
2022-09-29 21:03:55 +00:00
logger . error ( f " URL Error: Language file not found at { dir_path } / { language } .yml " )
else :
for file in os . listdir ( dir_path ) :
if file . endswith ( " .yml " ) and file != " default.yml " :
add_translation ( os . path . abspath ( f " { dir_path } / { file } " ) , file [ : - 4 ] )
2022-10-20 13:32:09 +00:00
return translations , key_names , variables
2021-12-21 21:33:37 +00:00
2022-10-13 06:12:24 +00:00
def apply_template ( self , call_name , mapping_name , data , template_call , extra_variables ) :
2021-12-21 21:33:37 +00:00
if not self . templates :
raise Failed ( f " { self . data_type } Error: No templates found " )
2022-01-06 07:12:32 +00:00
elif not template_call :
2021-12-21 21:33:37 +00:00
raise Failed ( f " { self . data_type } Error: template attribute is blank " )
else :
2022-04-03 21:22:58 +00:00
new_attributes = { }
2022-12-01 22:28:30 +00:00
for original_variables in util . get_list ( template_call , split = False ) :
2022-12-02 17:27:27 +00:00
if original_variables [ " name " ] not in self . templates :
2022-12-01 22:28:30 +00:00
raise Failed ( f " { self . data_type } Error: template { original_variables [ ' name ' ] } not found " )
elif not isinstance ( self . templates [ original_variables [ " name " ] ] [ 0 ] , dict ) :
raise Failed ( f " { self . data_type } Error: template { original_variables [ ' name ' ] } is not a dictionary " )
2021-05-07 05:06:40 +00:00
else :
2022-12-02 17:27:27 +00:00
logger . separator ( f " Template { original_variables [ ' name ' ] } " , space = False , border = False , trace = True )
2022-10-24 21:01:58 +00:00
logger . trace ( " " )
2022-12-01 22:28:30 +00:00
logger . trace ( f " Original: { original_variables } " )
2022-07-18 19:24:23 +00:00
2022-01-06 20:46:37 +00:00
remove_variables = [ ]
2022-09-20 19:56:03 +00:00
optional = [ ]
2023-01-24 19:07:42 +00:00
for tm in original_variables . copy ( ) :
2022-12-01 22:28:30 +00:00
if original_variables [ tm ] is None :
2022-01-06 20:46:37 +00:00
remove_variables . append ( tm )
2022-12-01 22:28:30 +00:00
original_variables . pop ( tm )
2022-09-20 19:56:03 +00:00
optional . append ( str ( tm ) )
2022-01-06 20:46:37 +00:00
2022-12-01 22:28:30 +00:00
template , temp_vars = self . templates [ original_variables [ " name " ] ]
2022-07-14 19:37:10 +00:00
2022-10-13 06:12:24 +00:00
if call_name :
name = call_name
elif " name " in template :
2022-07-14 19:37:10 +00:00
name = template [ " name " ]
2022-10-13 06:12:24 +00:00
else :
2022-07-15 23:30:09 +00:00
name = mapping_name
2022-07-14 19:37:10 +00:00
2022-07-16 03:01:25 +00:00
name_var = f " { self . data_type . lower ( ) } _name "
2023-02-22 16:59:27 +00:00
now = datetime . now ( )
2022-12-01 22:28:30 +00:00
original_variables [ name_var ] = str ( name )
original_variables [ " mapping_name " ] = mapping_name
2023-02-22 16:59:27 +00:00
original_variables [ " current_year " ] = now . year
original_variables [ " current_month " ] = now . month
original_variables [ " current_day " ] = now . day
2022-12-01 22:28:30 +00:00
original_variables [ " library_type " ] = self . library . type . lower ( ) if self . library else " item "
original_variables [ " library_typeU " ] = self . library . type if self . library else " Item "
original_variables [ " library_name " ] = self . library . name if self . library else " playlist "
def replace_var ( input_item , search_dicts ) :
if not isinstance ( search_dicts , list ) :
search_dicts = [ search_dicts ]
return_item = input_item
for search_dict in search_dicts :
for rk , rv in search_dict . items ( ) :
2023-04-01 19:02:32 +00:00
if rk not in [ " name_format " , " summary_format " ] :
if f " << { rk } >> " == str ( return_item ) :
return_item = rv
if f " << { rk } >> " in str ( return_item ) :
return_item = str ( return_item ) . replace ( f " << { rk } >> " , str ( rv ) )
2022-12-01 22:28:30 +00:00
return return_item
2021-12-21 21:33:37 +00:00
2022-10-17 06:48:08 +00:00
conditionals = { }
if " conditionals " in template :
if not template [ " conditionals " ] :
raise Failed ( f " { self . data_type } Error: template sub-attribute conditionals is blank " )
if not isinstance ( template [ " conditionals " ] , dict ) :
raise Failed ( f " { self . data_type } Error: template sub-attribute conditionals is not a dictionary " )
for ck , cv in template [ " conditionals " ] . items ( ) :
conditionals [ ck ] = cv
2022-10-26 21:14:08 +00:00
added_vars = { }
2022-10-27 18:56:40 +00:00
init_defaults = { }
if " default " in template :
if not template [ " default " ] :
raise Failed ( f " { self . data_type } Error: template sub-attribute default is blank " )
if not isinstance ( template [ " default " ] , dict ) :
raise Failed ( f " { self . data_type } Error: template sub-attribute default is not a dictionary " )
init_defaults = template [ " default " ]
2022-11-09 20:13:04 +00:00
all_init_defaults = { k : v for k , v in init_defaults . items ( ) }
2022-11-11 15:57:36 +00:00
2022-12-01 22:28:30 +00:00
variables = { }
2022-11-11 15:57:36 +00:00
temp_conditionals = { }
2022-10-26 21:14:08 +00:00
for input_dict , input_type , overwrite_call in [
2022-12-01 22:28:30 +00:00
( original_variables , " Call " , False ) ,
2022-10-17 06:48:08 +00:00
( temp_vars , " External " , False ) ,
( extra_variables , " Definition " , False ) ,
( self . temp_vars , " Config " , True )
] :
2022-10-24 21:01:58 +00:00
logger . trace ( " " )
2022-10-17 06:48:08 +00:00
logger . trace ( f " { input_type } : { input_dict } " )
for input_key , input_value in input_dict . items ( ) :
if input_key == " conditionals " :
if not input_value :
raise Failed ( f " { self . data_type } Error: { input_type } template sub-attribute conditionals is blank " )
if not isinstance ( input_value , dict ) :
raise Failed ( f " { self . data_type } Error: { input_type } template sub-attribute conditionals is not a dictionary " )
for ck , cv in input_value . items ( ) :
2022-11-11 15:57:36 +00:00
temp_conditionals [ ck ] = cv
2022-10-27 18:56:40 +00:00
elif input_key == " default " :
if not input_value :
raise Failed ( f " { self . data_type } Error: { input_type } template sub-attribute default is blank " )
if not isinstance ( input_value , dict ) :
raise Failed ( f " { self . data_type } Error: { input_type } template sub-attribute default is not a dictionary " )
for dk , dv in input_value . items ( ) :
2022-11-09 20:13:04 +00:00
all_init_defaults [ dk ] = dv
2022-10-26 21:14:08 +00:00
else :
2022-12-01 22:28:30 +00:00
input_key = replace_var ( input_key , original_variables )
if input_value is None :
optional . append ( str ( input_key ) )
if input_key in variables :
variables . pop ( input_key )
if input_key in added_vars :
added_vars . pop ( input_key )
elif overwrite_call :
variables [ input_key ] = input_value
2023-04-14 04:16:35 +00:00
elif input_key not in added_vars :
2022-12-01 22:28:30 +00:00
added_vars [ input_key ] = input_value
2022-10-26 21:14:08 +00:00
for k , v in added_vars . items ( ) :
if k not in variables :
variables [ k ] = v
2022-11-11 15:57:36 +00:00
for k , v in temp_conditionals . items ( ) :
if k not in variables :
conditionals [ k ] = v
2021-12-21 21:33:37 +00:00
2022-10-20 20:38:53 +00:00
if " key_name " in variables :
2022-11-15 07:00:59 +00:00
variables [ " original_key_name " ] = variables [ " key_name " ]
2022-11-19 20:45:38 +00:00
first_letter = str ( variables [ " key_name " ] ) . upper ( ) [ 0 ]
variables [ " key_name_first_letter " ] = first_letter if first_letter . isalpha ( ) else " # "
2022-09-29 21:03:55 +00:00
2021-12-21 21:33:37 +00:00
default = { }
2022-11-09 20:13:04 +00:00
if all_init_defaults :
var_default = { replace_var ( dk , variables ) : replace_var ( dv , variables ) for dk , dv in all_init_defaults . items ( ) if dk not in variables }
2023-04-28 03:43:26 +00:00
for d_key , d_value in var_default . items ( ) :
final_key = replace_var ( d_key , var_default )
2022-10-27 18:56:40 +00:00
if final_key not in optional and final_key not in variables and final_key not in conditionals :
2023-04-28 03:43:26 +00:00
default [ final_key ] = d_value
if " << " in str ( d_value ) :
default [ f " { final_key } _encoded " ] = re . sub ( r ' <<(.+)>> ' , r ' << \ 1_encoded>> ' , d_value )
2023-01-17 16:50:23 +00:00
else :
2024-05-28 20:22:51 +00:00
default [ f " { final_key } _encoded " ] = quote ( d_value )
2022-07-06 01:32:57 +00:00
2021-12-21 21:33:37 +00:00
if " optional " in template :
if template [ " optional " ] :
for op in util . get_list ( template [ " optional " ] ) :
2022-08-23 15:30:53 +00:00
op = replace_var ( op , variables )
2022-10-18 20:54:28 +00:00
if op not in default and op not in conditionals :
2021-12-21 21:33:37 +00:00
optional . append ( str ( op ) )
2022-04-02 23:29:30 +00:00
optional . append ( f " { op } _encoded " )
2022-11-09 20:13:04 +00:00
elif op in init_defaults :
2022-07-18 19:24:23 +00:00
logger . debug ( " " )
logger . debug ( f " Template Warning: variable { op } cannot be optional if it has a default " )
2021-12-21 21:33:37 +00:00
else :
raise Failed ( f " { self . data_type } Error: template sub-attribute optional is blank " )
2022-10-17 06:48:08 +00:00
for con_key , con_value in conditionals . items ( ) :
logger . debug ( " " )
logger . debug ( f " Conditional: { con_key } " )
if not isinstance ( con_value , dict ) :
2022-11-07 20:00:29 +00:00
raise Failed ( f " { self . data_type } Error: conditional { con_key } is not a dictionary " )
2022-10-17 06:48:08 +00:00
final_key = replace_var ( con_key , [ variables , default ] )
if final_key != con_key :
2022-11-15 07:00:59 +00:00
logger . trace ( f " Variable: { final_key } " )
if final_key in variables :
logger . debug ( f ' Conditional Variable: { final_key } overwritten to " { variables [ final_key ] } " ' )
continue
2022-10-17 06:48:08 +00:00
if " conditions " not in con_value :
2022-11-07 20:00:29 +00:00
raise Failed ( f " { self . data_type } Error: conditions sub-attribute required " )
2022-10-17 06:48:08 +00:00
conditions = con_value [ " conditions " ]
if isinstance ( conditions , dict ) :
conditions = [ conditions ]
if not isinstance ( conditions , list ) :
raise Failed ( f " { self . data_type } Error: conditions sub-attribute must be a list or dictionary " )
condition_found = False
for i , condition in enumerate ( conditions , 1 ) :
if not isinstance ( condition , dict ) :
raise Failed ( f " { self . data_type } Error: each condition must be a dictionary " )
if " value " not in condition :
raise Failed ( f " { self . data_type } Error: each condition must have a result value " )
condition_passed = True
for var_key , var_value in condition . items ( ) :
if var_key == " value " :
continue
2023-01-24 20:06:33 +00:00
error_text = " "
2023-01-25 03:18:19 +00:00
con_var_value = " "
2022-10-17 06:48:08 +00:00
var_key = replace_var ( var_key , [ variables , default ] )
var_value = replace_var ( var_value , [ variables , default ] )
2022-11-03 19:44:01 +00:00
if var_key . endswith ( " .exists " ) :
2023-01-25 03:18:19 +00:00
con_var_value = util . parse ( self . data_type , var_key , var_value , datatype = " bool " , default = False )
if con_var_value :
2024-04-08 22:04:10 +00:00
if var_key [ : - 7 ] not in variables or variables [ var_key [ : - 7 ] ] is None :
2023-01-24 20:06:33 +00:00
error_text = " - does not exist "
2024-04-08 22:04:10 +00:00
elif var_key [ : - 7 ] in variables and variables [ var_key [ : - 7 ] ] is not None :
2023-01-24 20:06:33 +00:00
error_text = " - exists "
2024-04-08 22:04:10 +00:00
con_var_value = var_key [ : - 7 ]
2022-11-03 19:44:01 +00:00
elif var_key . endswith ( " .not " ) :
2024-04-08 22:04:10 +00:00
if var_key [ : - 4 ] in variables :
con_var_value = variables [ var_key [ : - 4 ] ]
if isinstance ( var_value , list ) :
if con_var_value in var_value :
error_text = f ' in { var_value } '
elif str ( con_var_value ) == str ( var_value ) :
error_text = f ' is " { var_value } " '
elif var_key . endswith ( " .notdefault " ) :
2024-04-08 22:50:24 +00:00
var_name = var_key [ : - 11 ]
2024-04-05 01:46:21 +00:00
if var_name in variables or var_name in default :
con_var_value = variables [ var_name ] if var_name in variables else default [ var_name ]
2022-11-03 19:44:01 +00:00
if isinstance ( var_value , list ) :
2023-01-25 03:18:19 +00:00
if con_var_value in var_value :
2023-01-24 20:06:33 +00:00
error_text = f ' in { var_value } '
2023-01-25 03:18:19 +00:00
elif str ( con_var_value ) == str ( var_value ) :
2023-01-24 20:06:33 +00:00
error_text = f ' is " { var_value } " '
elif var_key in variables or var_key in default :
2023-01-25 03:18:19 +00:00
con_var_value = variables [ var_key ] if var_key in variables else default [ var_key ]
2023-01-24 20:06:33 +00:00
if isinstance ( var_value , list ) :
2023-01-25 03:18:19 +00:00
if con_var_value not in var_value :
2023-01-24 20:06:33 +00:00
error_text = f ' not in { var_value } '
2023-01-25 03:18:19 +00:00
elif str ( con_var_value ) != str ( var_value ) :
2023-01-24 20:06:33 +00:00
error_text = f ' is not " { var_value } " '
2022-07-10 08:02:39 +00:00
else :
2023-01-24 22:03:30 +00:00
error_text = " is not a variable provided or a default variable "
2023-01-24 20:06:33 +00:00
if error_text :
2023-01-25 03:18:19 +00:00
if con_var_value :
error_text = f ' : " { con_var_value } " { error_text } '
2023-01-24 22:03:30 +00:00
logger . trace ( f ' Condition { i } Failed: { var_key } { error_text } ' )
2022-10-17 06:48:08 +00:00
condition_passed = False
if condition_passed :
2023-03-21 19:37:16 +00:00
logger . trace ( f ' Conditional Variable: { final_key } is " { condition [ " value " ] } " ' )
2022-10-17 06:48:08 +00:00
condition_found = True
2023-01-17 16:50:23 +00:00
if condition [ " value " ] is not None :
variables [ final_key ] = condition [ " value " ]
2024-05-28 20:22:51 +00:00
variables [ f " { final_key } _encoded " ] = quote ( condition [ " value " ] )
2023-01-17 16:50:23 +00:00
else :
optional . append ( final_key )
2022-10-17 06:48:08 +00:00
break
if not condition_found :
2022-11-15 07:00:59 +00:00
if " default " in con_value :
2023-03-21 19:37:16 +00:00
logger . trace ( f ' Conditional Variable: { final_key } defaults to " { con_value [ " default " ] } " ' )
2022-10-17 06:48:08 +00:00
variables [ final_key ] = con_value [ " default " ]
2024-05-28 20:22:51 +00:00
variables [ f " { final_key } _encoded " ] = quote ( con_value [ " default " ] )
2022-11-15 07:00:59 +00:00
else :
2023-03-21 19:37:16 +00:00
logger . trace ( f " Conditional Variable: { final_key } added as optional variable " )
2022-10-17 06:48:08 +00:00
optional . append ( str ( final_key ) )
optional . append ( f " { final_key } _encoded " )
2022-07-10 08:02:39 +00:00
2022-05-10 23:52:04 +00:00
sort_name = None
2024-03-20 19:29:52 +00:00
sort_mapping = None
2021-12-24 00:16:46 +00:00
if " move_prefix " in template or " move_collection_prefix " in template :
prefix = None
if " move_prefix " in template :
prefix = template [ " move_prefix " ]
elif " move_collection_prefix " in template :
2022-10-24 21:01:58 +00:00
logger . debug ( " " )
2022-07-18 19:24:23 +00:00
logger . debug ( f " { self . data_type } Warning: template sub-attribute move_collection_prefix will run as move_prefix " )
2021-12-24 00:16:46 +00:00
prefix = template [ " move_collection_prefix " ]
if prefix :
for op in util . get_list ( prefix ) :
2024-03-20 20:38:15 +00:00
if not sort_name and variables [ name_var ] . startswith ( f " { op } " ) :
2022-06-26 19:26:08 +00:00
sort_name = f " { variables [ name_var ] [ len ( op ) : ] . strip ( ) } , { op } "
2024-03-20 20:38:15 +00:00
if not sort_mapping and variables [ " mapping_name " ] . startswith ( f " { op } " ) :
2024-03-20 19:29:52 +00:00
sort_mapping = f " { variables [ ' mapping_name ' ] [ len ( op ) : ] . strip ( ) } , { op } "
2024-03-26 16:51:10 +00:00
if sort_name and sort_mapping :
2024-05-28 20:22:51 +00:00
break
2021-12-21 21:33:37 +00:00
else :
2021-12-24 00:16:46 +00:00
raise Failed ( f " { self . data_type } Error: template sub-attribute move_prefix is blank " )
2022-05-10 23:52:04 +00:00
variables [ f " { self . data_type . lower ( ) } _sort " ] = sort_name if sort_name else variables [ name_var ]
2024-03-20 19:29:52 +00:00
variables [ " mapping_sort " ] = sort_mapping if sort_mapping else variables [ " mapping_name " ]
2021-12-21 21:33:37 +00:00
2022-10-18 02:57:42 +00:00
for key , value in variables . copy ( ) . items ( ) :
if " << " in key and " >> " in key :
for k , v in variables . items ( ) :
if f " << { k } >> " in key :
2023-09-26 04:05:59 +00:00
key = key . replace ( f " << { k } >> " , f " { v } " )
2022-10-18 02:57:42 +00:00
for k , v in default . items ( ) :
if f " << { k } >> " in key :
2023-09-26 04:05:59 +00:00
key = key . replace ( f " << { k } >> " , f " { v } " )
2023-01-06 15:00:59 +00:00
if key not in variables :
variables [ key ] = value
2022-10-18 02:57:42 +00:00
for key , value in variables . copy ( ) . items ( ) :
2024-05-28 20:22:51 +00:00
variables [ f " { key } _encoded " ] = quote ( value )
2022-10-18 02:57:42 +00:00
2022-10-18 03:30:22 +00:00
default = { k : v for k , v in default . items ( ) if k not in variables }
2023-09-19 19:28:03 +00:00
og_optional = optional
optional = [ ]
for key in og_optional :
if " << " in key and " >> " in key :
for k , v in variables . items ( ) :
if f " << { k } >> " in key :
2023-09-26 04:05:59 +00:00
key = key . replace ( f " << { k } >> " , f " { v } " )
2023-09-19 19:28:03 +00:00
for k , v in default . items ( ) :
if f " << { k } >> " in key :
2023-09-26 04:05:59 +00:00
key = key . replace ( f " << { k } >> " , f " { v } " )
2023-09-19 19:28:03 +00:00
if key not in variables and key not in default :
optional . append ( key )
2022-10-18 02:57:42 +00:00
2022-10-24 21:01:58 +00:00
logger . trace ( " " )
logger . trace ( f " Variables: { variables } " )
logger . trace ( " " )
logger . trace ( f " Defaults: { default } " )
logger . trace ( " " )
logger . trace ( f " Optional: { optional } " )
2022-10-07 13:06:30 +00:00
logger . trace ( " " )
2022-07-18 15:44:16 +00:00
2023-01-06 20:41:35 +00:00
def check_for_var ( _method , _data , _debug ) :
2024-05-15 20:52:32 +00:00
def scan_text ( og_txt , var , actual_value , second = False ) :
2022-07-10 08:02:39 +00:00
if og_txt is None :
return og_txt
elif str ( og_txt ) == f " << { var } >> " :
return actual_value
2023-01-09 22:01:02 +00:00
elif f " << { var } " in str ( og_txt ) :
final = str ( og_txt ) . replace ( f " << { var } >> " , str ( actual_value ) ) if f " << { var } >> " in str ( og_txt ) else str ( og_txt )
2024-05-15 20:52:32 +00:00
if f " << { var } " in final and second :
2023-04-28 03:43:26 +00:00
match = re . search ( f " <<( { var } ([+-])( \\ d+))>> " , final )
2023-01-09 22:01:02 +00:00
if match :
try :
final = final . replace ( f " << { match . group ( 1 ) } >> " , str ( int ( actual_value ) + ( int ( match . group ( 3 ) ) * ( - 1 if match . group ( 2 ) == " - " else 1 ) ) ) )
2023-01-18 19:20:49 +00:00
except ( ValueError , TypeError ) :
logger . error ( f " Template Error: { actual_value } must be a number to use { match . group ( 1 ) } " )
raise Failed
2023-01-09 22:01:02 +00:00
return final
2022-07-10 08:02:39 +00:00
else :
return og_txt
2023-01-06 20:41:35 +00:00
if _debug :
logger . trace ( f " Start { _method } : { _data } " )
2023-01-12 20:30:31 +00:00
try :
for i_check in range ( 8 ) :
for option in optional :
2023-03-31 22:19:06 +00:00
if option not in variables and f " << { option } >> " in str ( _data ) :
2023-01-12 20:30:31 +00:00
raise Failed
2024-05-15 20:52:32 +00:00
for option in [ False , True ] :
for variable , variable_data in variables . items ( ) :
if ( variable == " collection_name " or variable == " playlist_name " ) and _method in [ " radarr_tag " , " item_radarr_tag " , " sonarr_tag " , " item_sonarr_tag " ] :
_data = scan_text ( _data , variable , variable_data . replace ( " , " , " " ) , second = option )
elif ( variable == " name_format " and _method != " name " ) or ( variable == " summary_format " and _method != " summary " ) :
continue
elif variable != " name " and ( _method not in [ " name " , " summary " ] or variable != " key_name " ) :
_data = scan_text ( _data , variable , variable_data , second = option )
2023-01-12 20:30:31 +00:00
for dm , dd in default . items ( ) :
2023-04-01 19:02:32 +00:00
if ( dm == " name_format " and _method != " name " ) or ( dm == " summary_format " and _method != " summary " ) :
continue
elif _method not in [ " name " , " summary " ] or dm != " key_name " :
_data = scan_text ( _data , dm , dd )
2023-01-12 20:30:31 +00:00
except Failed :
if _debug :
logger . trace ( f " Failed { _method } : { _data } " )
raise
2023-01-06 20:41:35 +00:00
if _debug :
logger . trace ( f " End { _method } : { _data } " )
2022-07-10 08:02:39 +00:00
return _data
2023-01-06 20:41:35 +00:00
def check_data ( _method , _data , _debug ) :
2021-12-21 21:33:37 +00:00
if isinstance ( _data , dict ) :
final_data = { }
for sm , sd in _data . items ( ) :
try :
2023-01-06 20:41:35 +00:00
final_data [ check_for_var ( _method , sm , _debug ) ] = check_data ( _method , sd , _debug )
2021-12-21 21:33:37 +00:00
except Failed :
continue
2022-05-25 16:12:01 +00:00
if not final_data :
raise Failed
2021-12-21 21:33:37 +00:00
elif isinstance ( _data , list ) :
final_data = [ ]
for li in _data :
try :
2023-01-06 20:41:35 +00:00
final_data . append ( check_data ( _method , li , _debug ) )
2021-12-21 21:33:37 +00:00
except Failed :
continue
2022-05-25 16:12:01 +00:00
if not final_data :
raise Failed
2021-12-21 21:33:37 +00:00
else :
2023-01-06 20:41:35 +00:00
final_data = check_for_var ( _method , _data , _debug )
2021-12-21 21:33:37 +00:00
return final_data
for method_name , attr_data in template . items ( ) :
2022-07-10 08:02:39 +00:00
if method_name not in data and method_name not in [ " default " , " optional " , " conditionals " , " move_collection_prefix " , " move_prefix " ] :
try :
2023-01-06 20:41:35 +00:00
debug_template = False
new_name = check_for_var ( method_name , method_name , debug_template )
2022-07-10 08:02:39 +00:00
if new_name in new_attributes :
2023-08-14 15:37:32 +00:00
logger . trace ( " " )
logger . trace ( f " Template Warning: template attribute: { new_name } from { variables [ ' name ' ] } skipped " )
2022-07-10 08:02:39 +00:00
else :
2023-01-06 20:41:35 +00:00
new_attributes [ new_name ] = check_data ( new_name , attr_data , debug_template )
2022-07-10 08:02:39 +00:00
except Failed :
continue
2023-02-22 17:13:28 +00:00
logger . trace ( f " Current Final: { new_attributes } " )
logger . trace ( " " )
2022-10-27 18:56:40 +00:00
logger . separator ( f " Final Template Attributes " , space = False , border = False , debug = True )
logger . debug ( " " )
logger . debug ( new_attributes )
2022-07-18 18:14:32 +00:00
logger . debug ( " " )
2022-04-03 21:22:58 +00:00
return new_attributes
2021-12-21 21:33:37 +00:00
2022-10-17 06:48:08 +00:00
def external_templates ( self , data , overlay = False ) :
2022-04-26 15:28:04 +00:00
if data and " external_templates " in data and data [ " external_templates " ] :
2023-04-18 18:37:08 +00:00
files , _ = util . load_files ( data [ " external_templates " ] , " external_templates " )
2022-04-13 04:30:59 +00:00
if not files :
logger . error ( " Config Error: No Paths Found for external_templates " )
2022-04-21 05:35:07 +00:00
for file_type , template_file , temp_vars , _ in files :
2022-10-17 06:48:08 +00:00
temp_data = self . load_file ( file_type , template_file , overlay = overlay )
2022-04-10 01:45:33 +00:00
if temp_data and isinstance ( temp_data , dict ) and " templates " in temp_data and temp_data [ " templates " ] and isinstance ( temp_data [ " templates " ] , dict ) :
2022-10-20 13:32:09 +00:00
self . templates . update ( { k : ( v , temp_vars ) for k , v in temp_data [ " templates " ] . items ( ) if k not in self . templates } )
2021-12-21 21:33:37 +00:00
class MetadataFile ( DataFile ) :
2023-12-12 22:03:15 +00:00
def __init__ ( self , config , library , file_type , path , temp_vars , asset_directory , file_style ) :
self . file_style = file_style
self . type_str = f " { file_style . capitalize ( ) } File "
2024-02-28 20:45:11 +00:00
super ( ) . __init__ ( config , file_type , path , temp_vars , asset_directory , self . type_str )
2021-12-21 21:33:37 +00:00
self . data_type = " Collection "
self . library = library
2023-03-04 20:20:52 +00:00
self . metadata = None
self . collections = None
2023-12-13 19:32:02 +00:00
self . dynamic_collections = [ ]
2023-03-04 20:20:52 +00:00
self . templates = None
2023-03-15 21:03:05 +00:00
self . update_collections = True
2023-03-04 20:20:52 +00:00
self . update_seasons = True
self . update_episodes = True
self . set_collections = { }
2023-03-15 21:03:05 +00:00
self . style_priority = [ ]
2023-12-12 22:03:15 +00:00
if self . file_style == " image " :
2023-03-04 20:20:52 +00:00
self . metadata = { }
2024-04-22 14:20:12 +00:00
if self . type == " Default " :
2023-03-21 07:12:50 +00:00
if self . path . endswith ( " .yml " ) :
self . path = self . path [ : - 4 ]
elif self . path . endswith ( " .yaml " ) :
self . path = self . path [ : - 5 ]
data = self . load_file ( self . type , " set " , images = True , folder = f " { self . path } / " )
else :
data = self . load_file ( self . type , self . path , images = True )
2023-03-04 20:20:52 +00:00
methods = { t . lower ( ) : t for t in self . temp_vars }
2023-03-15 21:03:05 +00:00
2023-03-04 20:20:52 +00:00
use_all = True if " use_all " in methods and self . temp_vars [ methods [ " use_all " ] ] else False
2023-03-21 07:12:50 +00:00
logger . info ( f " Use All Sections: { use_all } " )
2023-03-15 21:03:05 +00:00
exclude = [ ]
if " exclude " in methods :
if not use_all :
raise Failed ( f " Image Set Error: exclude only works when use_all is true " )
exclude = util . parse ( " Images " , " exclude " , self . temp_vars , datatype = " list " , methods = methods )
logger . info ( f " Exclude: { exclude } " )
include = [ ]
if " include " in methods :
if use_all :
raise Failed ( f " Image Set Error: include only works when use_all is false " )
include = util . parse ( " Images " , " include " , self . temp_vars , datatype = " list " , methods = methods )
logger . info ( f " Include: { include } " )
if " style_priority " in methods :
self . style_priority = util . parse ( " Images " , " style_priority " , self . temp_vars , datatype = " list " , methods = methods )
logger . info ( f " Style Priority: { self . style_priority } " )
if " update_collections " in methods :
self . update_collections = util . parse ( " Images " , " update_collections " , self . temp_vars , datatype = " bool " , methods = methods , default = True )
logger . info ( f " Update Collections: { self . update_collections } " )
2023-03-04 20:20:52 +00:00
if " update_seasons " in methods :
self . update_seasons = util . parse ( " Images " , " update_seasons " , self . temp_vars , datatype = " bool " , methods = methods , default = True )
logger . info ( f " Update Seasons: { self . update_seasons } " )
2023-03-15 21:03:05 +00:00
2023-03-04 20:20:52 +00:00
if " update_episodes " in methods :
self . update_episodes = util . parse ( " Images " , " update_episodes " , self . temp_vars , datatype = " bool " , methods = methods , default = True )
logger . info ( f " Update Episodes: { self . update_episodes } " )
2023-03-18 18:49:46 +00:00
item_attr = " movies " if self . library . is_movie else " shows "
2023-03-27 19:56:00 +00:00
for section_key , section_data in get_dict ( " sections " , data ) . items ( ) :
if not isinstance ( section_data , dict ) :
raise Failed ( " Image Set Error: Section Data must be a dictionary " )
if " builders " not in section_data or not section_data [ " builders " ] :
logger . trace ( f " Skipping No Builder for Section: { section_key } " )
2023-03-21 14:31:43 +00:00
continue
2023-03-27 19:56:00 +00:00
elif item_attr not in section_data :
raise Failed ( f " Section Data must have the { item_attr } attribute " )
elif not section_data [ item_attr ] :
raise Failed ( f " Section Data attribute { item_attr } is empty " )
elif " styles " not in section_data :
raise Failed ( " Image Section Error: Section Data must have the styles attribute " )
styles = util . parse ( " Section Data " , " styles " , section_data [ " styles " ] , datatype = " dictlist " )
if not styles :
raise Failed ( " Image Section Error: Section Data styles attribute is empty " )
default_style = None
for sk , sv in styles . items ( ) :
2023-03-27 23:27:31 +00:00
default_style = sk
break
2023-03-27 19:56:00 +00:00
if not default_style :
raise Failed ( f " Image Section Error: No styles found for section: { section_key } " )
2023-03-04 20:20:52 +00:00
use_key = None
2023-03-27 19:56:00 +00:00
if f " use_ { section_key } " in methods :
2023-04-28 03:43:26 +00:00
use_key = util . parse ( " Images " , f " use_ { section_key } " , self . temp_vars , datatype = " bool " , methods = methods , default = False )
2023-03-27 19:56:00 +00:00
logger . info ( f " Use { section_key } : { use_key } " )
2023-03-15 21:03:05 +00:00
if use_key is False :
2023-03-27 19:56:00 +00:00
logger . trace ( f " Skipped as use_ { section_key } is false " )
2023-03-15 21:03:05 +00:00
continue
2023-03-27 19:56:00 +00:00
elif use_all and section_key in exclude :
logger . trace ( f " Skipped as { section_key } is in the exclude list " )
2023-03-04 20:20:52 +00:00
continue
2023-03-27 19:56:00 +00:00
elif not use_all and use_key is None and section_key not in include :
logger . trace ( f " Skipped as use_all is false and use_ { section_key } is not set { f ' and { section_key } not in the include list ' if include else ' ' } " )
2023-03-15 21:03:05 +00:00
continue
prioritized_style = None
for ps in self . style_priority :
if ps in styles :
prioritized_style = ps
break
2023-03-27 19:56:00 +00:00
if f " style_ { section_key } " in methods :
style_key = util . parse ( " Images " , f " style_ { section_key } " , self . temp_vars , methods = methods , default = default_style )
logger . info ( f " Style { section_key } : { style_key } " )
2023-03-21 07:12:50 +00:00
if style_key not in styles :
2023-03-27 19:56:00 +00:00
p_warning = f " Image Section Warning: { section_key } has no style: { style_key } using "
2023-03-15 21:03:05 +00:00
if prioritized_style :
logger . warning ( f " { p_warning } Prioritized Style: { prioritized_style } " )
2023-03-21 07:12:50 +00:00
style_key = prioritized_style
2023-03-15 21:03:05 +00:00
else :
logger . warning ( f " { p_warning } default: { default_style } . Options: { ' , ' . join ( [ s for s in styles ] ) } " )
2023-03-21 07:12:50 +00:00
style_key = default_style
2023-03-15 21:03:05 +00:00
elif prioritized_style :
logger . info ( f " Using Prioritized Style: { prioritized_style } " )
2023-03-21 07:12:50 +00:00
style_key = prioritized_style
2023-03-15 21:03:05 +00:00
else :
2023-03-21 07:12:50 +00:00
style_key = default_style
2023-03-27 19:56:00 +00:00
if self . update_collections and " collections " in section_data and section_data [ " collections " ] :
self . set_collections [ section_key ] = section_data [ " collections " ]
2023-03-21 07:12:50 +00:00
2023-03-27 19:56:00 +00:00
if f " style_file_ { section_key } " in methods :
style_file = self . temp_vars [ methods [ f " style_file_ { section_key } " ] ]
2023-03-21 07:12:50 +00:00
elif not styles [ style_key ] :
2023-03-27 19:56:00 +00:00
style_file = [ { " pmm " : f " { section_key } / { style_key } " } ]
2023-03-18 18:49:46 +00:00
else :
2023-03-21 07:12:50 +00:00
style_file = styles [ style_key ]
if not style_file :
raise Failed ( " Image Style Error: style file call attribute is blank " )
style_dict = style_file [ 0 ] if isinstance ( style_file , list ) else style_file
if not isinstance ( style_dict , dict ) :
2023-04-05 21:01:29 +00:00
raise Failed ( f " Image Style Error: style file call attribute: { style_dict } is not a dictionary " )
2023-03-21 07:12:50 +00:00
elif not style_dict :
raise Failed ( " Image Style Error: style file call attribute dictionary is empty " )
2023-03-27 19:56:00 +00:00
style_data = self . get_style_data ( style_dict , section_key , items_data = section_data [ item_attr ] )
for item_name , item_data in section_data [ item_attr ] . items ( ) :
2023-03-21 07:12:50 +00:00
if item_name not in style_data or not style_data [ item_name ] :
continue
2023-03-04 20:20:52 +00:00
if isinstance ( item_data , dict ) :
if " mapping_id " not in item_data :
2023-03-27 19:56:00 +00:00
raise Failed ( f " Image Section Error: { section_key } : { item_name } : No mapping ID found " )
2023-03-04 20:20:52 +00:00
meta_data = item_data
else :
meta_data = { " mapping_id " : item_data }
2023-03-21 07:12:50 +00:00
meta_data [ " style_data " ] = style_data [ item_name ]
2023-03-27 19:56:00 +00:00
meta_data [ " section_key " ] = section_key
2023-03-21 07:12:50 +00:00
meta_data [ " style_key " ] = style_key
if " seasons " in style_data [ item_name ] and style_data [ item_name ] [ " seasons " ] :
2023-03-06 16:30:16 +00:00
season_dict = { }
2023-03-21 07:12:50 +00:00
for season_num , season_data in style_data [ item_name ] [ " seasons " ] . items ( ) :
2023-03-06 16:30:16 +00:00
season_dict [ season_num ] = { }
if season_data and " episodes " in season_data :
episode_dict = { }
for episode_num in season_data [ " episodes " ] :
episode_dict [ episode_num ] = { }
season_dict [ season_num ] [ " episodes " ] = episode_dict
meta_data [ " seasons " ] = season_dict
2023-03-04 20:20:52 +00:00
self . metadata [ item_name ] = meta_data
if not self . metadata :
raise Failed ( f " { self . type_str } Error: No metadata items added " )
logger . info ( " " )
2023-12-12 22:03:15 +00:00
logger . info ( " Images File Loaded Successfully " )
2023-03-04 20:20:52 +00:00
elif file_type == " Data " :
2021-12-06 07:52:08 +00:00
self . collections = get_dict ( " collections " , path , library . collections )
self . templates = get_dict ( " templates " , path )
else :
2021-12-21 21:33:37 +00:00
logger . info ( " " )
2023-12-12 22:03:15 +00:00
logger . separator ( f " Loading { self . type_str } { file_type } : { path } " )
2023-02-13 23:13:54 +00:00
logger . info ( " " )
2022-10-03 15:24:58 +00:00
data = self . load_file ( self . type , self . path )
2023-12-13 19:32:02 +00:00
if self . file_style == " metadata " :
self . metadata = get_dict ( " metadata " , data , library . metadatas )
2021-12-21 21:33:37 +00:00
self . templates = get_dict ( " templates " , data )
2022-04-10 04:33:59 +00:00
self . external_templates ( data )
2023-12-13 19:32:02 +00:00
if self . file_style == " collection " :
self . collections = get_dict ( " collections " , data , library . collections )
self . dynamic_collections = get_dict ( " dynamic_collections " , data )
col_names = library . collections + [ c for c in self . collections ]
for map_name , dynamic in self . dynamic_collections . items ( ) :
logger . info ( " " )
logger . separator ( f " Building { map_name } Dynamic Collections " , space = False , border = False )
logger . info ( " " )
try :
methods = { dm . lower ( ) : dm for dm in dynamic }
for m in methods :
if m not in dynamic_attributes :
logger . warning ( f " Config Warning: { methods [ m ] } attribute is invalid. Options: { ' , ' . join ( dynamic_attributes ) } " )
if " type " not in methods :
raise Failed ( f " Config Error: { map_name } type attribute not found " )
elif not dynamic [ methods [ " type " ] ] :
raise Failed ( f " Config Error: { map_name } type attribute is blank " )
elif dynamic [ methods [ " type " ] ] . lower ( ) not in auto [ library . type ] :
raise Failed ( f " Config Error: { map_name } type attribute { dynamic [ methods [ ' type ' ] ] . lower ( ) } invalid Options: { auto [ library . type ] } " )
elif dynamic [ methods [ " type " ] ] . lower ( ) == " network " and library . agent not in plex . new_plex_agents :
raise Failed ( f " Config Error: { map_name } type attribute: network only works with the New Plex TV Agent " )
elif dynamic [ methods [ " type " ] ] . lower ( ) . startswith ( " trakt " ) and not self . config . Trakt :
raise Failed ( f " Config Error: { map_name } type attribute: { dynamic [ methods [ ' type ' ] ] } requires trakt to be configured " )
auto_type = dynamic [ methods [ " type " ] ] . lower ( )
og_exclude = [ ]
if " exclude " in self . temp_vars :
og_exclude = util . parse ( " Config " , " exclude " , self . temp_vars [ " exclude " ] , parent = " template_variables " , datatype = " strlist " )
elif " exclude " in methods :
og_exclude = util . parse ( " Config " , " exclude " , dynamic , parent = map_name , methods = methods , datatype = " strlist " )
if " append_exclude " in self . temp_vars :
og_exclude . extend ( util . parse ( " Config " , " append_exclude " , self . temp_vars [ " append_exclude " ] , parent = " template_variables " , datatype = " strlist " ) )
if " remove_exclude " in self . temp_vars :
for word in util . parse ( " Config " , " remove_exclude " , self . temp_vars [ " remove_exclude " ] , parent = " template_variables " , datatype = " strlist " ) :
og_exclude . remove ( word )
include = [ ]
if " include " in self . temp_vars :
include = util . parse ( " Config " , " include " , self . temp_vars [ " include " ] , parent = " template_variables " , datatype = " strlist " )
elif " include " in methods :
include = [ i for i in util . parse ( " Config " , " include " , dynamic , parent = map_name , methods = methods , datatype = " strlist " ) if i not in og_exclude ]
if " append_include " in self . temp_vars :
include . extend ( util . parse ( " Config " , " append_include " , self . temp_vars [ " append_include " ] , parent = " template_variables " , datatype = " strlist " ) )
if " remove_include " in self . temp_vars :
for word in util . parse ( " Config " , " remove_include " , self . temp_vars [ " remove_include " ] , parent = " template_variables " , datatype = " strlist " ) :
include . remove ( word )
2022-10-11 21:02:49 +00:00
addons = { }
2023-12-13 19:32:02 +00:00
if " addons " in self . temp_vars :
addons = util . parse ( " Config " , " addons " , self . temp_vars [ " addons " ] , parent = " template_variables " , datatype = " dictliststr " )
elif " addons " in methods :
addons = util . parse ( " Config " , " addons " , dynamic , parent = map_name , methods = methods , datatype = " dictliststr " )
if " append_addons " in self . temp_vars :
append_addons = util . parse ( " Config " , " append_addons " , self . temp_vars [ " append_addons " ] , parent = map_name , methods = methods , datatype = " dictliststr " )
for k , v in append_addons . items ( ) :
if k not in addons :
addons [ k ] = [ ]
addons [ k ] . extend ( v )
if " remove_addons " in self . temp_vars :
remove_addons = util . parse ( " Config " , " remove_addons " , self . temp_vars [ " remove_addons " ] , parent = map_name , methods = methods , datatype = " dictliststr " )
for k , v in remove_addons . items ( ) :
if k in addons :
for word in v :
addons [ k ] . remove ( word )
exclude = [ str ( e ) for e in og_exclude ]
for k , v in addons . items ( ) :
if k in v :
logger . warning ( f " Config Warning: { k } cannot be an addon for itself " )
exclude . extend ( [ y for y in v if y != k and y not in exclude ] )
default_title_format = " <<key_name>> "
default_template = None
auto_list = { }
2022-11-11 06:57:22 +00:00
all_keys = { }
2024-01-02 13:51:40 +00:00
extra_template_vars = { }
2023-12-13 19:32:02 +00:00
dynamic_data = None
def _check_dict ( check_dict ) :
for ck , cv in check_dict . items ( ) :
all_keys [ str ( ck ) ] = cv
if str ( ck ) not in exclude and str ( cv ) not in exclude :
auto_list [ str ( ck ) ] = cv
if auto_type == " decade " and library . is_show :
all_items = library . get_all ( )
if addons :
raise Failed ( f " Config Error: addons cannot be used with show decades " )
addons = { }
2022-11-11 06:57:22 +00:00
all_keys = { }
2023-12-13 19:32:02 +00:00
for i , item in enumerate ( all_items , 1 ) :
logger . ghost ( f " Processing: { i } / { len ( all_items ) } { item . title } " )
if item . year :
decade = str ( int ( math . floor ( item . year / 10 ) * 10 ) )
if decade not in addons :
addons [ decade ] = [ ]
if str ( item . year ) not in addons [ decade ] :
addons [ decade ] . append ( str ( item . year ) )
all_keys [ str ( item . year ) ] = str ( item . year )
auto_list = { str ( k ) : f " { k } s " for k in addons if str ( k ) not in exclude and f " { k } s " not in exclude }
default_template = { " smart_filter " : { " limit " : 50 , " sort_by " : " critic_rating.desc " , " any " : { " year " : " <<value>> " } } }
2023-12-31 16:45:00 +00:00
default_title_format = " Best <<library_type>>s of the <<key_name>> "
2023-12-13 19:32:02 +00:00
elif auto_type in [ " genre " , " mood " , " style " , " album_genre " , " album_mood " , " album_style " , " track_mood " , " country " , " studio " , " edition " , " network " , " year " , " episode_year " , " decade " , " content_rating " , " subtitle_language " , " audio_language " , " resolution " ] :
search_tag = auto_type_translation [ auto_type ] if auto_type in auto_type_translation else auto_type
if library . is_show and auto_type in [ " resolution " , " subtitle_language " , " audio_language " ] :
tags = library . get_tags ( f " episode. { search_tag } " )
else :
tags = library . get_tags ( search_tag )
if auto_type in [ " subtitle_language " , " audio_language " ] :
all_keys = { }
auto_list = { }
for i in tags :
final_title = self . config . TMDb . TMDb . _iso_639_1 [ str ( i . key ) ] . english_name if str ( i . key ) in self . config . TMDb . TMDb . _iso_639_1 else str ( i . title ) # noqa
all_keys [ str ( i . key ) ] = final_title
if all ( [ x not in exclude for x in [ final_title , str ( i . title ) , str ( i . key ) ] ] ) :
auto_list [ str ( i . key ) ] = final_title
elif auto_type in [ " resolution " , " decade " ] :
all_keys = { str ( i . key ) : i . title for i in tags }
auto_list = { str ( i . key ) : i . title for i in tags if str ( i . title ) not in exclude and str ( i . key ) not in exclude }
else :
all_keys = { str ( i . title ) : i . title for i in tags }
auto_list = { str ( i . title ) : i . title for i in tags if str ( i . title ) not in exclude }
if library . is_music :
final_var = auto_type if auto_type . startswith ( ( " album " , " track " ) ) else f " artist_ { auto_type } "
default_template = { " smart_filter " : { " limit " : 50 if auto_type . startswith ( " track " ) else 10 , " sort_by " : " plays.desc " , " any " : { final_var : " <<value>> " } } }
music_type = " <<library_type>> "
if auto_type . startswith ( ( " album " , " track " ) ) :
default_template [ " builder_level " ] = " album " if auto_type . startswith ( " album " ) else " track "
music_type = " Album " if auto_type . startswith ( " album " ) else " Track "
default_title_format = f " Most Played <<key_name>> { music_type } s "
elif auto_type == " resolution " :
default_template = { " smart_filter " : { " sort_by " : " title.asc " , " any " : { auto_type : " <<value>> " } } }
default_title_format = " <<key_name>> <<library_type>>s "
else :
default_template = { " smart_filter " : { " limit " : 50 , " sort_by " : " critic_rating.desc " , " any " : { f " { auto_type } .is " if auto_type == " studio " else auto_type : " <<value>> " } } }
2023-12-31 16:45:00 +00:00
if auto_type . startswith ( " episode " ) :
default_template [ " builder_level " ] = " episode "
default_title_format = " Best Episodes of <<key_name>> "
elif auto_type == " year " :
default_title_format = " Best <<library_type>>s of <<key_name>> "
elif auto_type == " decade " :
default_title_format = " Best <<library_type>>s of the <<key_name>> "
else :
default_title_format = " Top <<key_name>> <<library_type>>s "
2023-12-13 19:32:02 +00:00
elif auto_type == " tmdb_collection " :
all_items = library . get_all ( )
for i , item in enumerate ( all_items , 1 ) :
logger . ghost ( f " Processing: { i } / { len ( all_items ) } { item . title } " )
tmdb_id , tvdb_id , imdb_id = library . get_ids ( item )
tmdb_item = config . TMDb . get_item ( item , tmdb_id , tvdb_id , imdb_id , is_movie = True )
if tmdb_item and tmdb_item . collection_id and tmdb_item . collection_name :
all_keys [ str ( tmdb_item . collection_id ) ] = tmdb_item . collection_name
if str ( tmdb_item . collection_id ) not in exclude and tmdb_item . collection_name not in exclude :
auto_list [ str ( tmdb_item . collection_id ) ] = tmdb_item . collection_name
logger . exorcise ( )
elif auto_type == " original_language " :
all_items = library . get_all ( )
for i , item in enumerate ( all_items , 1 ) :
logger . ghost ( f " Processing: { i } / { len ( all_items ) } { item . title } " )
tmdb_id , tvdb_id , imdb_id = library . get_ids ( item )
tmdb_item = config . TMDb . get_item ( item , tmdb_id , tvdb_id , imdb_id , is_movie = library . type == " Movie " )
if tmdb_item and tmdb_item . language_iso :
all_keys [ tmdb_item . language_iso ] = tmdb_item . language_name
if tmdb_item . language_iso not in exclude and tmdb_item . language_name not in exclude :
auto_list [ tmdb_item . language_iso ] = tmdb_item . language_name
logger . exorcise ( )
2022-03-11 07:07:37 +00:00
default_title_format = " <<key_name>> <<library_type>>s "
2023-12-13 19:32:02 +00:00
elif auto_type == " origin_country " :
all_items = library . get_all ( )
for i , item in enumerate ( all_items , 1 ) :
logger . ghost ( f " Processing: { i } / { len ( all_items ) } { item . title } " )
tmdb_id , tvdb_id , imdb_id = library . get_ids ( item )
tmdb_item = config . TMDb . get_item ( item , tmdb_id , tvdb_id , imdb_id , is_movie = library . type == " Movie " )
if tmdb_item and tmdb_item . countries :
for country in tmdb_item . countries :
all_keys [ country . iso_3166_1 . lower ( ) ] = country . name
if country . iso_3166_1 . lower ( ) not in exclude and country . name not in exclude :
auto_list [ country . iso_3166_1 . lower ( ) ] = country . name
logger . exorcise ( )
default_title_format = " <<key_name>> <<library_type>>s "
elif auto_type in [ " actor " , " director " , " writer " , " producer " ] :
people = { }
2024-01-04 03:47:22 +00:00
dynamic_data = util . parse ( " Config " , " data " , dynamic , parent = map_name , methods = methods , datatype = " dict " )
if " data " in self . temp_vars :
temp_data = util . parse ( " Config " , " data " , self . temp_vars [ " data " ] , datatype = " dict " )
for k , v in temp_data . items ( ) :
dynamic_data [ k ] = v
2023-12-13 19:32:02 +00:00
person_methods = { am . lower ( ) : am for am in dynamic_data }
if " actor_depth " in person_methods :
person_methods [ " depth " ] = person_methods . pop ( " actor_depth " )
if " actor_minimum " in person_methods :
person_methods [ " minimum " ] = person_methods . pop ( " actor_minimum " )
if " number_of_actors " in person_methods :
person_methods [ " limit " ] = person_methods . pop ( " number_of_actors " )
person_depth = util . parse ( " Config " , " depth " , dynamic_data , parent = f " { map_name } data " , methods = person_methods , datatype = " int " , default = 3 , minimum = 1 )
person_minimum = util . parse ( " Config " , " minimum " , dynamic_data , parent = f " { map_name } data " , methods = person_methods , datatype = " int " , default = 3 , minimum = 1 ) if " minimum " in person_methods else None
person_limit = util . parse ( " Config " , " limit " , dynamic_data , parent = f " { map_name } data " , methods = person_methods , datatype = " int " , default = 25 , minimum = 1 ) if " limit " in person_methods else None
lib_all = library . get_all ( )
include_cols = [ ]
for i , item in enumerate ( lib_all , 1 ) :
logger . ghost ( f " Scanning: { i } / { len ( lib_all ) } { item . title } " )
try :
item = self . library . reload ( item )
for person in getattr ( item , f " { auto_type } s " ) [ : person_depth ] :
if person . tag in include :
if person . tag not in include_cols :
include_cols . append ( person . tag )
else :
if person . id not in people :
people [ person . id ] = { " name " : person . tag , " count " : 0 }
people [ person . id ] [ " count " ] + = 1
except Failed as e :
logger . error ( f " Plex Error: { e } " )
roles = [ data for _ , data in people . items ( ) ]
roles . sort ( key = operator . itemgetter ( ' count ' ) , reverse = True )
if not person_minimum :
person_minimum = 1 if person_limit else 3
if not person_limit :
person_limit = len ( roles )
person_count = 0
for inc in include_cols :
auto_list [ inc ] = inc
all_keys [ inc ] = inc
2022-10-11 21:02:49 +00:00
person_count + = 1
2023-12-13 19:32:02 +00:00
for role in roles :
if person_count < person_limit and role [ " count " ] > = person_minimum and role [ " name " ] not in exclude :
auto_list [ role [ " name " ] ] = role [ " name " ]
all_keys [ role [ " name " ] ] = role [ " name " ]
person_count + = 1
default_template = { " plex_search " : { " any " : { auto_type : " <<value>> " } } }
2024-01-02 13:51:40 +00:00
elif auto_type == " imdb_awards " :
2024-01-04 03:47:22 +00:00
dynamic_data = util . parse ( " Config " , " data " , dynamic , parent = map_name , methods = methods , datatype = " dict " )
if " data " in self . temp_vars :
temp_data = util . parse ( " Config " , " data " , self . temp_vars [ " data " ] , datatype = " dict " )
for k , v in temp_data . items ( ) :
dynamic_data [ k ] = v
2024-01-02 13:51:40 +00:00
award_methods = { am . lower ( ) : am for am in dynamic_data }
event_id = util . parse ( " Config " , " event_id " , dynamic_data , parent = f " { map_name } data " , methods = award_methods , regex = ( r " (ev \ d+) " , " ev0000003 " ) )
extra_template_vars [ " event_id " ] = event_id
if event_id not in self . config . IMDb . events_validation :
raise Failed ( f " Config Error: { map_name } data only specific Event IDs work with imdb_awards. Event Options: [ { ' , ' . join ( [ k for k in self . config . IMDb . events_validation ] ) } ] " )
2024-01-03 13:24:48 +00:00
_ , event_years = self . config . IMDb . get_event_years ( event_id )
year_options = [ event_years [ len ( event_years ) - i ] for i in range ( 1 , len ( event_years ) + 1 ) ]
2024-01-02 22:04:33 +00:00
2024-01-11 17:08:29 +00:00
def get_position ( attr ) :
2024-01-03 13:24:48 +00:00
if attr not in award_methods :
return 0 if attr == " starting " else len ( year_options )
position_value = str ( dynamic_data [ award_methods [ attr ] ] )
if not position_value :
raise Failed ( f " Config Error: { map_name } data { attr } attribute is blank " )
2024-01-04 03:47:22 +00:00
if position_value . startswith ( ( " first " , " latest " , " current_year " ) ) :
2024-01-11 17:08:29 +00:00
is_first = position_value . startswith ( " first " )
int_values = position_value . split ( " + " if is_first else " - " )
2024-01-03 13:24:48 +00:00
try :
if len ( int_values ) == 1 :
2024-01-11 17:08:29 +00:00
return 1 if is_first else len ( year_options )
2024-01-03 13:24:48 +00:00
else :
2024-01-11 17:08:29 +00:00
return ( int ( int_values [ 1 ] . strip ( ) ) + ( 1 if is_first else 0 ) ) * ( 1 if is_first else - 1 )
2024-01-03 13:24:48 +00:00
except ValueError :
raise Failed ( f " Config Error: { map_name } data { attr } attribute modifier invalid ' { int_values [ 1 ] } ' " )
elif position_value in year_options :
2024-01-11 17:08:29 +00:00
return year_options . index ( position_value ) + 1
2024-01-03 13:24:48 +00:00
else :
raise Failed ( f " Config Error: { map_name } data { attr } attribute invalid: { position_value } " )
2024-01-02 22:04:33 +00:00
2024-01-11 17:08:29 +00:00
found_options = year_options [ get_position ( " starting " ) - 1 : get_position ( " ending " ) ]
2024-01-02 22:04:33 +00:00
2024-01-03 13:24:48 +00:00
if not found_options :
raise Failed ( f " Config Error: { map_name } data starting/ending range found no valid events " )
for option in event_years :
2024-01-02 13:51:40 +00:00
all_keys [ option ] = option
2024-01-03 13:24:48 +00:00
if option not in exclude and option in found_options :
2024-01-02 13:51:40 +00:00
auto_list [ option ] = option
2024-01-03 13:24:48 +00:00
default_template = { " imdb_award " : { " event_id " : " <<event_id>> " , " event_year " : " <<value>> " , " winning " : True } }
2023-12-13 19:32:02 +00:00
elif auto_type == " number " :
2024-01-04 03:47:22 +00:00
dynamic_data = util . parse ( " Config " , " data " , dynamic , parent = map_name , methods = methods , datatype = " dict " )
if " data " in self . temp_vars :
temp_data = util . parse ( " Config " , " data " , self . temp_vars [ " data " ] , datatype = " dict " )
for k , v in temp_data . items ( ) :
dynamic_data [ k ] = v
2023-12-13 19:32:02 +00:00
number_methods = { nm . lower ( ) : nm for nm in dynamic_data }
if " starting " in number_methods and str ( dynamic_data [ number_methods [ " starting " ] ] ) . startswith ( " current_year " ) :
year_values = str ( dynamic_data [ number_methods [ " starting " ] ] ) . split ( " - " )
try :
starting = datetime . now ( ) . year - ( 0 if len ( year_values ) == 1 else int ( year_values [ 1 ] . strip ( ) ) )
except ValueError :
2024-01-02 13:51:40 +00:00
raise Failed ( f " Config Error: { map_name } data starting attribute modifier invalid ' { year_values [ 1 ] } ' " )
2023-12-13 19:32:02 +00:00
else :
starting = util . parse ( " Config " , " starting " , dynamic_data , parent = f " { map_name } data " , methods = number_methods , datatype = " int " , default = 0 , minimum = 0 )
if " ending " in number_methods and str ( dynamic_data [ number_methods [ " ending " ] ] ) . startswith ( " current_year " ) :
year_values = str ( dynamic_data [ number_methods [ " ending " ] ] ) . split ( " - " )
try :
ending = datetime . now ( ) . year - ( 0 if len ( year_values ) == 1 else int ( year_values [ 1 ] . strip ( ) ) )
except ValueError :
2024-01-02 13:51:40 +00:00
raise Failed ( f " Config Error: { map_name } data ending attribute modifier invalid ' { year_values [ 1 ] } ' " )
2023-12-13 19:32:02 +00:00
else :
ending = util . parse ( " Config " , " ending " , dynamic_data , parent = f " { map_name } data " , methods = number_methods , datatype = " int " , default = 0 , minimum = 1 )
increment = util . parse ( " Config " , " increment " , dynamic_data , parent = f " { map_name } data " , methods = number_methods , datatype = " int " , default = 1 , minimum = 1 ) if " increment " in number_methods else 1
if starting > ending :
raise Failed ( f " Config Error: { map_name } data ending must be greater than starting " )
current = starting
while current < = ending :
all_keys [ str ( current ) ] = str ( current )
if str ( current ) not in exclude and current not in exclude :
auto_list [ str ( current ) ] = str ( current )
current + = increment
elif auto_type == " custom " :
2024-01-04 03:47:22 +00:00
if " data " in self . temp_vars :
dynamic_data = util . parse ( " Config " , " data " , self . temp_vars [ " data " ] , datatype = " strdict " )
else :
dynamic_data = util . parse ( " Config " , " data " , dynamic , parent = map_name , methods = methods , datatype = " strdict " )
if " remove_data " in self . temp_vars :
for k in util . parse ( " Config " , " remove_data " , self . temp_vars [ " remove_data " ] , datatype = " strlist " ) :
if k in dynamic_data :
dynamic_data . pop ( k )
if " append_data " in self . temp_vars :
for k , v in util . parse ( " Config " , " append_data " , self . temp_vars [ " append_data " ] , datatype = " strdict " ) . items ( ) :
dynamic_data [ k ] = v
for k , v in dynamic_data . items ( ) :
2023-12-13 19:32:02 +00:00
all_keys [ k ] = v
if k not in exclude and v not in exclude :
auto_list [ k ] = v
elif auto_type == " trakt_liked_lists " :
_check_dict ( self . config . Trakt . all_liked_lists ( ) )
2024-06-05 18:44:33 +00:00
elif auto_type == " letterboxd_user_lists " :
dynamic_data = util . parse ( " Config " , " data " , dynamic , parent = map_name , methods = methods , datatype = " dict " )
if " data " in self . temp_vars :
temp_data = util . parse ( " Config " , " data " , self . temp_vars [ " data " ] , datatype = " dict " )
for k , v in temp_data . items ( ) :
dynamic_data [ k ] = v
letter_methods = { am . lower ( ) : am for am in dynamic_data }
users = util . parse ( " Config " , " username " , dynamic_data , parent = f " { map_name } data " , methods = letter_methods , datatype = " strlist " )
sort = util . parse ( " Config " , " sort_by " , dynamic_data , parent = f " { map_name } data " , methods = letter_methods , options = letterboxd . sort_options , default = " updated " )
limit = util . parse ( " Config " , " limit " , dynamic_data , parent = f " { map_name } data " , methods = letter_methods , datatype = " int " , minimum = 0 , default = 0 )
final = { }
for user in users :
out = self . config . Letterboxd . get_user_lists ( user , sort , self . language )
if limit != 0 :
out = out [ : limit ]
for url , name in out :
final [ url ] = name
_check_dict ( final )
2023-12-13 19:32:02 +00:00
elif auto_type == " tmdb_popular_people " :
2024-01-04 03:47:22 +00:00
if " data " in self . temp_vars :
dynamic_data = util . parse ( " Config " , " data " , self . temp_vars [ " data " ] , datatype = " int " , minimum = 1 )
else :
dynamic_data = util . parse ( " Config " , " data " , dynamic , parent = map_name , methods = methods , datatype = " int " , minimum = 1 )
2023-12-13 19:32:02 +00:00
_check_dict ( self . config . TMDb . get_popular_people ( dynamic_data ) )
2024-01-04 03:47:22 +00:00
elif auto_type in [ " trakt_people_list " , " trakt_user_lists " ] :
if " data " in self . temp_vars :
dynamic_data = util . parse ( " Config " , " data " , self . temp_vars [ " data " ] , datatype = " strlist " )
else :
dynamic_data = util . parse ( " Config " , " data " , dynamic , parent = map_name , methods = methods , datatype = " strlist " )
if " remove_data " in self . temp_vars :
for k in util . parse ( " Config " , " remove_data " , self . temp_vars [ " remove_data " ] , datatype = " strlist " ) :
if k in dynamic_data :
dynamic_data . remove ( k )
if " append_data " in self . temp_vars :
for k in util . parse ( " Config " , " append_data " , self . temp_vars [ " append_data " ] , datatype = " strlist " ) :
if k not in dynamic_data :
dynamic_data . append ( k )
2023-12-13 19:32:02 +00:00
for option in dynamic_data :
2024-01-04 03:47:22 +00:00
if auto_type == " trakt_user_lists " :
_check_dict ( { self . config . Trakt . build_user_url ( u [ 0 ] , u [ 1 ] ) : u [ 2 ] for u in self . config . Trakt . all_user_lists ( option ) } )
else :
_check_dict ( self . config . Trakt . get_people ( option ) )
2022-05-18 17:07:15 +00:00
else :
2023-12-13 19:32:02 +00:00
raise Failed ( f " Config Error: { map_name } type attribute { dynamic [ methods [ ' type ' ] ] } invalid " )
2022-03-09 23:59:31 +00:00
2023-12-13 19:32:02 +00:00
if " append_data " in self . temp_vars :
for k , v in util . parse ( " Config " , " append_data " , self . temp_vars [ " append_data " ] , parent = map_name , methods = methods , datatype = " strdict " ) . items ( ) :
all_keys [ k ] = v
if k not in exclude and v not in exclude :
auto_list [ k ] = v
custom_keys = True
if " custom_keys " in self . temp_vars :
custom_keys = util . parse ( " Config " , " custom_keys " , self . temp_vars [ " custom_keys " ] , parent = " template_variables " , default = custom_keys )
elif " custom_keys " in methods :
custom_keys = util . parse ( " Config " , " custom_keys " , dynamic , parent = map_name , methods = methods , default = custom_keys )
for add_key , combined_keys in addons . items ( ) :
if add_key not in all_keys and add_key not in og_exclude :
final_keys = [ ck for ck in combined_keys if ck in all_keys ]
if custom_keys and final_keys :
auto_list [ add_key ] = add_key
addons [ add_key ] = final_keys
elif custom_keys :
logger . trace ( f " Config Warning: { add_key } Custom Key must have at least one Key " )
else :
for final_key in final_keys :
auto_list [ final_key ] = all_keys [ final_key ]
title_format = default_title_format
if " title_format " in self . temp_vars :
title_format = util . parse ( " Config " , " title_format " , self . temp_vars [ " title_format " ] , parent = " template_variables " , default = default_title_format )
elif " title_format " in methods :
title_format = util . parse ( " Config " , " title_format " , dynamic , parent = map_name , methods = methods , default = default_title_format )
if " <<key_name>> " not in title_format and " <<title>> " not in title_format :
logger . error ( f " Config Error: <<key_name>> not in title_format: { title_format } using default: { default_title_format } " )
title_format = default_title_format
if " post_format_override " in methods :
methods [ " title_override " ] = methods . pop ( " post_format_override " )
if " pre_format_override " in methods :
methods [ " key_name_override " ] = methods . pop ( " pre_format_override " )
title_override = { }
if " title_override " in self . temp_vars :
title_override = util . parse ( " Config " , " title_override " , self . temp_vars [ " title_override " ] , parent = " template_variables " , datatype = " strdict " )
elif " title_override " in methods :
title_override = util . parse ( " Config " , " title_override " , dynamic , parent = map_name , methods = methods , datatype = " strdict " )
key_name_override = { }
if " key_name_override " in self . temp_vars :
key_name_override = util . parse ( " Config " , " key_name_override " , self . temp_vars [ " key_name_override " ] , parent = " template_variables " , datatype = " strdict " )
elif " key_name_override " in methods :
key_name_override = util . parse ( " Config " , " key_name_override " , dynamic , parent = map_name , methods = methods , datatype = " strdict " )
test_override = [ ]
for k , v in key_name_override . items ( ) :
if v in test_override :
logger . warning ( f " Config Warning: { v } can only be used once skipping { k } : { v } " )
key_name_override . pop ( k )
else :
test_override . append ( v )
2024-01-02 13:51:40 +00:00
test = False
if " test " in self . temp_vars :
test = util . parse ( " Config " , " test " , self . temp_vars [ " test " ] , parent = " template_variables " , datatype = " bool " )
elif " test " in methods :
test = util . parse ( " Config " , " test " , dynamic , parent = map_name , methods = methods , default = False , datatype = " bool " )
2023-12-13 19:32:02 +00:00
sync = False
if " sync " in self . temp_vars :
sync = util . parse ( " Config " , " sync " , self . temp_vars [ " sync " ] , parent = " template_variables " , datatype = " bool " )
elif " sync " in methods :
sync = util . parse ( " Config " , " sync " , dynamic , parent = map_name , methods = methods , default = False , datatype = " bool " )
if " <<library_type>> " in title_format :
title_format = title_format . replace ( " <<library_type>> " , library . type . lower ( ) )
if " <<library_typeU>> " in title_format :
title_format = title_format . replace ( " <<library_typeU>> " , library . type )
if " limit " in self . temp_vars and " <<limit>> " in title_format :
2024-03-06 16:31:44 +00:00
title_format = title_format . replace ( " <<limit>> " , str ( self . temp_vars [ " limit " ] ) )
2023-12-13 19:32:02 +00:00
template_variables = util . parse ( " Config " , " template_variables " , dynamic , parent = map_name , methods = methods , datatype = " dictdict " ) if " template_variables " in methods else { }
if " template " in methods :
template_names = util . parse ( " Config " , " template " , dynamic , parent = map_name , methods = methods , datatype = " strlist " )
has_var = False
for template_name in template_names :
if template_name not in self . templates :
raise Failed ( f " Config Error: { map_name } template: { template_name } not found " )
if any ( [ a in str ( self . templates [ template_name ] [ 0 ] ) for a in [ " <<value " , " <<key " , f " << { auto_type } " ] ] ) :
has_var = True
if not has_var :
raise Failed ( f " Config Error: One { map_name } template: { template_names } is required to have the template variable <<value>> " )
elif auto_type in [ " number " , " list " ] :
raise Failed ( f " Config Error: { map_name } template required for type: { auto_type } " )
2022-01-30 07:48:56 +00:00
else :
2023-12-13 19:32:02 +00:00
self . templates [ map_name ] = ( default_template if default_template else default_templates [ auto_type ] , { } )
template_names = [ map_name ]
remove_prefix = [ ]
if " remove_prefix " in self . temp_vars :
remove_prefix = util . parse ( " Config " , " remove_prefix " , self . temp_vars [ " remove_prefix " ] , parent = " template_variables " , datatype = " commalist " )
elif " remove_prefix " in methods :
remove_prefix = util . parse ( " Config " , " remove_prefix " , dynamic , parent = map_name , methods = methods , datatype = " commalist " )
remove_suffix = [ ]
if " remove_suffix " in self . temp_vars :
remove_suffix = util . parse ( " Config " , " remove_suffix " , self . temp_vars [ " remove_suffix " ] , parent = " template_variables " , datatype = " commalist " )
elif " remove_suffix " in methods :
remove_suffix = util . parse ( " Config " , " remove_suffix " , dynamic , parent = map_name , methods = methods , datatype = " commalist " )
sync = { i . title : i for i in self . library . get_all_collections ( label = str ( map_name ) ) } if sync else { }
other_name = None
if " other_name " in self . temp_vars and include :
other_name = util . parse ( " Config " , " other_name " , self . temp_vars [ " remove_suffix " ] , parent = " template_variables " )
elif " other_name " in methods and include :
other_name = util . parse ( " Config " , " other_name " , dynamic , parent = map_name , methods = methods )
other_templates = util . parse ( " Config " , " other_template " , dynamic , parent = map_name , methods = methods , datatype = " strlist " ) if " other_template " in methods and include else None
if other_templates :
for other_template in other_templates :
if other_template not in self . templates :
raise Failed ( f " Config Error: { map_name } other template: { other_template } not found " )
2022-01-28 18:36:21 +00:00
else :
2023-12-13 19:32:02 +00:00
other_templates = template_names
other_keys = [ ]
logger . debug ( f " Mapping Name: { map_name } " )
logger . debug ( f " Type: { auto_type } " )
logger . debug ( f " Data: { dynamic_data } " )
logger . debug ( f " Exclude: { og_exclude } " )
logger . debug ( f " Exclude Final: { exclude } " )
logger . debug ( f " Addons: { addons } " )
logger . debug ( f " Template: { template_names } " )
logger . debug ( f " Other Template: { other_templates } " )
logger . debug ( f " Library Variables: { self . temp_vars } " )
logger . debug ( f " Template Variables: { template_variables } " )
logger . debug ( f " Remove Prefix: { remove_prefix } " )
logger . debug ( f " Remove Suffix: { remove_suffix } " )
logger . debug ( f " Title Format: { title_format } " )
logger . debug ( f " Key Name Override: { key_name_override } " )
logger . debug ( f " Title Override: { title_override } " )
logger . debug ( f " Custom Keys: { custom_keys } " )
logger . debug ( f " Test: { test } " )
logger . debug ( f " Sync: { sync } " )
logger . debug ( f " Include: { include } " )
logger . debug ( f " Other Name: { other_name } " )
logger . debug ( f " All Keys: { all_keys . keys ( ) } " )
if not auto_list :
raise Failed ( " No Keys found to create a set of Dynamic Collections " )
logger . debug ( f " Keys (Title): " )
for key , value in auto_list . items ( ) :
logger . debug ( f " - { key } { ' ' if key == value else f ' ( { value } ) ' } " )
used_keys = [ ]
for key , value in auto_list . items ( ) :
if include and key not in include :
if key not in exclude :
other_keys . append ( key )
continue
if key in key_name_override :
key_name = key_name_override [ key ]
else :
key_name = value
for prefix in remove_prefix :
if key_name . startswith ( prefix ) :
key_name = key_name [ len ( prefix ) : ] . strip ( )
for suffix in remove_suffix :
if key_name . endswith ( suffix ) :
key_name = key_name [ : - len ( suffix ) ] . strip ( )
key_value = [ key ] if key in all_keys else [ ]
if key in addons :
key_value . extend ( [ a for a in addons [ key ] if ( a in all_keys or auto_type == " custom " ) and a != key ] )
used_keys . extend ( key_value )
og_call = { " value " : key_value , auto_type : key_value , " key_name " : key_name , " key " : key }
for k , v in template_variables . items ( ) :
if k in self . temp_vars :
og_call [ k ] = self . temp_vars [ k ]
elif key in v :
og_call [ k ] = v [ key ]
elif " default " in v :
og_call [ k ] = v [ " default " ]
2024-01-02 13:51:40 +00:00
for k , v in extra_template_vars . items ( ) :
if k not in og_call :
og_call [ k ] = v
2023-12-13 19:32:02 +00:00
template_call = [ ]
for template_name in template_names :
new_call = og_call . copy ( )
new_call [ " name " ] = template_name
template_call . append ( new_call )
if key in title_override :
collection_title = title_override [ key ]
else :
collection_title = title_format . replace ( " <<title>> " , key_name ) . replace ( " <<key_name>> " , key_name )
if collection_title in col_names :
logger . warning ( f " Config Warning: Skipping duplicate collection: { collection_title } " )
else :
logger . info ( template_call )
col = { " template " : template_call , " append_label " : str ( map_name ) }
if test :
col [ " test " ] = True
if collection_title in sync :
sync . pop ( collection_title )
col_names . append ( collection_title )
self . collections [ collection_title ] = col
if other_name and not other_keys :
logger . warning ( f " Config Warning: Other Collection { other_name } not needed " )
elif other_name :
og_other = {
" value " : other_keys , " included_keys " : include , " used_keys " : used_keys ,
auto_type : other_keys , " key_name " : other_name , " key " : " other "
}
for k , v in template_variables . items ( ) :
if k in self . temp_vars and " other " in self . temp_vars [ k ] :
og_other [ k ] = self . temp_vars [ k ] [ " other " ]
elif k in self . temp_vars and " default " in self . temp_vars [ k ] :
og_other [ k ] = self . temp_vars [ k ] [ " default " ]
if " other " in v :
og_other [ k ] = v [ " other " ]
elif " default " in v :
og_other [ k ] = v [ " default " ]
other_call = [ ]
for other_template in other_templates :
new_call = og_other . copy ( )
new_call [ " name " ] = other_template
other_call . append ( new_call )
col = { " template " : other_call , " append_label " : str ( map_name ) }
2022-01-30 07:48:56 +00:00
if test :
col [ " test " ] = True
2023-12-13 19:32:02 +00:00
if other_name in sync :
sync . pop ( other_name )
self . collections [ other_name ] = col
for col_title , col in sync . items ( ) :
try :
self . library . delete ( col )
logger . info ( f " { map_name } Dynamic Collection: { col_title } Deleted " )
except Failed as e :
logger . error ( e )
except Failed as e :
logger . error ( e )
logger . error ( f " { map_name } Dynamic Collection Failed " )
continue
2021-12-21 21:33:37 +00:00
2023-12-12 22:03:15 +00:00
if self . file_style == " metadata " and not self . metadata :
raise Failed ( " YAML Error: metadata attribute is required " )
if self . file_style == " collection " and not self . collections :
raise Failed ( " YAML Error: collections or dynamic_collections attribute is required " )
2022-04-27 15:11:10 +00:00
logger . info ( " " )
2023-12-12 22:03:15 +00:00
logger . info ( f " { self . type_str } Loaded Successfully " )
2021-05-07 05:06:40 +00:00
2023-03-27 19:56:00 +00:00
def get_style_data ( self , style_file , section_key , items_data = None ) :
2023-03-21 07:12:50 +00:00
style_id = " "
for k , v in style_file . items ( ) :
style_id = f " { k } : { v } "
2023-03-06 16:30:16 +00:00
break
2023-03-21 07:12:50 +00:00
if style_id in self . library . image_styles :
return self . library . image_styles [ style_id ]
if " git_style " in style_file :
if not style_file [ " git_style " ] :
raise Failed ( " Image Set Error: git_style cannot be blank " )
if not items_data :
raise Failed ( " Image Set Error: items_data cannot be blank " )
top_tree , repo = self . config . GitHub . get_top_tree ( style_file [ " git_style " ] )
sub = style_file [ " git_subfolder " ] if " git_subfolder " in style_file and style_file [ " git_subfolder " ] else " "
2023-03-06 21:54:01 +00:00
sub = sub . replace ( " \\ " , " / " )
if sub . startswith ( " / " ) :
sub = sub [ 1 : ]
if sub . endswith ( " / " ) :
sub = sub [ : - 1 ]
if sub :
sub_str = " "
for folder in sub . split ( " / " ) :
2024-05-28 20:22:51 +00:00
folder_encode = quote ( folder )
2023-03-07 18:53:33 +00:00
sub_str + = f " { folder_encode } / "
2023-03-06 21:54:01 +00:00
if folder not in top_tree :
raise Failed ( f " Image Set Error: Subfolder { folder } Not Found at https://github.com { repo } tree/master/ { sub_str } " )
top_tree = self . config . GitHub . get_tree ( top_tree [ folder ] [ " url " ] )
2023-03-07 18:53:33 +00:00
sub = sub_str
2023-03-06 21:54:01 +00:00
def repo_url ( u ) :
return f " https://raw.githubusercontent.com { repo } master/ { sub } { u } "
def from_repo ( u ) :
2024-05-28 20:22:51 +00:00
return self . config . Requests . get ( repo_url ( u ) ) . content . decode ( ) . strip ( )
2023-03-06 21:54:01 +00:00
def check_for_definition ( check_key , check_tree , is_poster = True , git_name = None ) :
2023-03-06 22:01:07 +00:00
attr_name = " poster " if is_poster and ( git_name is None or " background " not in git_name ) else " background "
2023-03-06 21:54:01 +00:00
if ( git_name and git_name . lower ( ) . endswith ( " .tpdb " ) ) or ( not git_name and f " { attr_name } .tpdb " in check_tree ) :
2024-05-28 20:22:51 +00:00
return f " tpdb_ { attr_name } " , from_repo ( f " { check_key } / { quote ( git_name ) if git_name else f ' { attr_name } .tpdb ' } " )
2023-03-06 21:54:01 +00:00
elif ( git_name and git_name . lower ( ) . endswith ( " .url " ) ) or ( not git_name and f " { attr_name } .url " in check_tree ) :
2024-05-28 20:22:51 +00:00
return f " url_ { attr_name } " , from_repo ( f " { check_key } / { quote ( git_name ) if git_name else f ' { attr_name } .url ' } " )
2023-03-06 21:54:01 +00:00
elif git_name :
if git_name in check_tree :
2024-05-28 20:22:51 +00:00
return f " url_ { attr_name } " , repo_url ( f " { check_key } / { quote ( git_name ) } " )
2023-03-06 21:54:01 +00:00
else :
for ct in check_tree :
2023-03-07 18:53:33 +00:00
if ct . lower ( ) . startswith ( attr_name ) :
2024-05-28 20:22:51 +00:00
return f " url_ { attr_name } " , repo_url ( f " { check_key } / { quote ( ct ) } " )
2023-03-06 21:54:01 +00:00
return None , None
2023-03-06 22:01:07 +00:00
def init_set ( check_key , check_tree ) :
2023-03-06 21:54:01 +00:00
_data = { }
2023-03-06 22:01:07 +00:00
attr , attr_data = check_for_definition ( check_key , check_tree )
2023-03-06 21:54:01 +00:00
if attr :
_data [ attr ] = attr_data
2023-03-06 22:01:07 +00:00
attr , attr_data = check_for_definition ( check_key , check_tree , is_poster = False )
2023-03-06 21:54:01 +00:00
if attr :
_data [ attr ] = attr_data
return _data
2023-03-21 07:12:50 +00:00
style_data = { }
for k in items_data :
2023-03-06 21:54:01 +00:00
if k not in top_tree :
logger . info ( f " Image Set Warning: { k } not found at https://github.com { repo } tree/master/ { sub } " )
continue
2024-05-28 20:22:51 +00:00
k_encoded = quote ( k )
2023-03-06 21:54:01 +00:00
item_folder = self . config . GitHub . get_tree ( top_tree [ k ] [ " url " ] )
2023-03-07 18:53:33 +00:00
item_data = init_set ( k_encoded , item_folder )
2023-03-06 21:54:01 +00:00
seasons = { }
for ik in item_folder :
match = re . search ( r " ( \ d+) " , ik )
if match :
2024-05-28 20:22:51 +00:00
season_path = f " { k_encoded } / { quote ( ik ) } "
2023-03-06 21:54:01 +00:00
season_num = int ( match . group ( 1 ) )
season_folder = self . config . GitHub . get_tree ( item_folder [ ik ] [ " url " ] )
2023-03-07 18:53:33 +00:00
season_data = init_set ( season_path , season_folder )
2023-03-06 21:54:01 +00:00
episodes = { }
for sk in season_folder :
match = re . search ( r " ( \ d+)(?!.* \ d) " , sk )
if match :
episode_num = int ( match . group ( 1 ) )
2023-03-06 22:01:07 +00:00
if episode_num not in episodes :
episodes [ episode_num ] = { }
2023-03-07 18:53:33 +00:00
a , ad = check_for_definition ( season_path , season_folder , git_name = sk )
if a :
episodes [ episode_num ] [ a ] = ad
2023-03-06 21:54:01 +00:00
if episodes :
season_data [ " episodes " ] = episodes
seasons [ season_num ] = season_data
if seasons :
item_data [ " seasons " ] = seasons
2023-03-21 07:12:50 +00:00
style_data [ k ] = item_data
2023-03-06 21:54:01 +00:00
2023-03-21 07:12:50 +00:00
self . library . image_styles [ style_id ] = style_data
2023-03-06 21:54:01 +00:00
2023-03-27 19:56:00 +00:00
if section_key and section_key in self . set_collections and " collections " in top_tree :
2023-03-06 21:54:01 +00:00
collections_folder = self . config . GitHub . get_tree ( top_tree [ " collections " ] [ " url " ] )
2023-03-27 19:56:00 +00:00
for k , alts in self . set_collections [ section_key ] . items ( ) :
2023-03-06 21:54:01 +00:00
if k in collections_folder :
collection_data = init_set ( f " collections/ { k } " , self . config . GitHub . get_tree ( collections_folder [ k ] [ " url " ] ) )
self . library . collection_images [ k ] = collection_data
for alt in alts :
self . library . collection_images [ alt ] = collection_data
2023-03-21 07:12:50 +00:00
else :
2023-04-18 18:37:08 +00:00
files , _ = util . load_files ( style_file , " style_file " , err_type = self . type_str , single = True )
2023-03-06 21:54:01 +00:00
if not files :
2023-03-21 07:12:50 +00:00
raise Failed ( f " { self . type_str } Error: No Path Found for style_file " )
file_type , style_path , _ , _ = files [ 0 ]
temp_data = self . load_file ( file_type , style_path , images = True , folder = f " { self . path } /styles/ " )
item_attr = " movies " if self . library . is_movie else " shows "
2023-03-06 21:54:01 +00:00
if not isinstance ( temp_data , dict ) :
2023-03-21 07:12:50 +00:00
raise Failed ( " Image Style Error: base must be a dictionary " )
if item_attr not in temp_data :
raise Failed ( f " Image Style Error: Image Styles must use the base attribute { item_attr } " )
if not temp_data [ item_attr ] :
raise Failed ( f " Image Style Error: { item_attr } attribute is empty " )
if not isinstance ( temp_data [ item_attr ] , dict ) :
raise Failed ( f " Image Style Error: { item_attr } attribute must be a dictionary " )
self . library . image_styles [ style_id ] = temp_data [ item_attr ]
2023-03-27 19:56:00 +00:00
if section_key and section_key in self . set_collections and " collections " in temp_data and temp_data [ " collections " ] :
for k , alts in self . set_collections [ section_key ] . items ( ) :
2023-03-06 21:54:01 +00:00
if k in temp_data [ " collections " ] :
self . library . collection_images [ k ] = temp_data [ " collections " ] [ k ]
2023-03-15 13:26:58 +00:00
if alts :
for alt in alts :
self . library . collection_images [ alt ] = temp_data [ " collections " ] [ k ]
2023-03-21 07:12:50 +00:00
return self . library . image_styles [ style_id ]
2023-03-06 16:30:16 +00:00
2021-05-07 05:06:40 +00:00
def get_collections ( self , requested_collections ) :
if requested_collections :
return { c : self . collections [ c ] for c in util . get_list ( requested_collections ) if c in self . collections }
else :
return self . collections
2021-12-29 15:51:22 +00:00
def edit_tags ( self , attr , obj , group , alias , extra = None ) :
if attr in alias and f " { attr } .sync " in alias :
2023-03-04 20:20:52 +00:00
logger . error ( f " { self . type_str } Error: Cannot use { attr } and { attr } .sync together " )
2021-12-29 15:51:22 +00:00
elif f " { attr } .remove " in alias and f " { attr } .sync " in alias :
2023-03-04 20:20:52 +00:00
logger . error ( f " { self . type_str } Error: Cannot use { attr } .remove and { attr } .sync together " )
2022-05-10 13:56:29 +00:00
elif attr in alias and not group [ alias [ attr ] ] :
2023-03-04 20:20:52 +00:00
logger . warning ( f " { self . type_str } Error: { attr } attribute is blank " )
2022-05-10 13:56:29 +00:00
elif f " { attr } .remove " in alias and not group [ alias [ f " { attr } .remove " ] ] :
2023-03-04 20:20:52 +00:00
logger . warning ( f " { self . type_str } Error: { attr } .remove attribute is blank " )
2022-05-10 13:56:29 +00:00
elif f " { attr } .sync " in alias and not group [ alias [ f " { attr } .sync " ] ] :
2023-03-04 20:20:52 +00:00
logger . warning ( f " { self . type_str } Error: { attr } .sync attribute is blank " )
2021-12-29 15:51:22 +00:00
elif attr in alias or f " { attr } .remove " in alias or f " { attr } .sync " in alias :
add_tags = util . get_list ( group [ alias [ attr ] ] ) if attr in alias else [ ]
if extra :
add_tags . extend ( extra )
remove_tags = util . get_list ( group [ alias [ f " { attr } .remove " ] ] ) if f " { attr } .remove " in alias else None
2022-05-10 13:56:29 +00:00
sync_tags = util . get_list ( group [ alias [ f " { attr } .sync " ] ] ) if f " { attr } .sync " in alias else None
2022-03-25 20:13:51 +00:00
return len ( self . library . edit_tags ( attr , obj , add_tags = add_tags , remove_tags = remove_tags , sync_tags = sync_tags ) ) > 0
2021-12-29 15:51:22 +00:00
return False
2021-08-05 14:59:45 +00:00
def update_metadata ( self ) :
2021-05-24 03:38:46 +00:00
if not self . metadata :
return None
2021-05-07 05:06:40 +00:00
logger . info ( " " )
2023-03-04 20:20:52 +00:00
logger . separator ( f " Running { self . type_str } " )
2021-05-07 05:06:40 +00:00
logger . info ( " " )
2022-04-08 20:59:14 +00:00
next_year = datetime . now ( ) . year + 1
2021-05-07 05:06:40 +00:00
for mapping_name , meta in self . metadata . items ( ) :
2023-02-23 04:00:41 +00:00
try :
methods = { mm . lower ( ) : mm for mm in meta }
2021-05-07 05:06:40 +00:00
2022-05-09 20:40:39 +00:00
logger . info ( " " )
2023-03-04 20:20:52 +00:00
logger . separator ( f " { mapping_name } { self . type_str } " )
2023-02-23 04:00:41 +00:00
if " template " in methods :
2023-02-23 18:57:00 +00:00
logger . debug ( " " )
2023-02-23 04:00:41 +00:00
logger . separator ( f " Building Definition From Templates " , space = False , border = False )
2023-02-22 22:01:29 +00:00
logger . debug ( " " )
2023-02-23 04:00:41 +00:00
named_templates = [ ]
for original_variables in util . get_list ( meta [ methods [ " template " ] ] , split = False ) :
if not isinstance ( original_variables , dict ) :
2023-03-04 20:20:52 +00:00
raise Failed ( f " { self . type_str } Error: template attribute is not a dictionary " )
2023-02-23 04:00:41 +00:00
elif " name " not in original_variables :
2023-03-04 20:20:52 +00:00
raise Failed ( f " { self . type_str } Error: template sub-attribute name is required " )
2023-02-23 04:00:41 +00:00
elif not original_variables [ " name " ] :
2023-03-04 20:20:52 +00:00
raise Failed ( f " { self . type_str } Error: template sub-attribute name cannot be blank " )
2023-02-23 04:00:41 +00:00
named_templates . append ( original_variables [ " name " ] )
logger . debug ( f " Templates Called: { ' , ' . join ( named_templates ) } " )
new_variables = { }
if " variables " in methods :
logger . debug ( " " )
logger . debug ( " Validating Method: variables " )
if not isinstance ( meta [ methods [ " variables " ] ] , dict ) :
2023-03-04 20:20:52 +00:00
raise Failed ( f " { self . type_str } Error: variables must be a dictionary (key: value pairs) " )
2023-02-23 04:00:41 +00:00
logger . trace ( meta [ methods [ " variables " ] ] )
new_variables = meta [ methods [ " variables " ] ]
name = meta [ methods [ " name " ] ] if " name " in methods else None
new_attributes = self . apply_template ( name , mapping_name , meta , meta [ methods [ " template " ] ] , new_variables )
for attr in new_attributes :
if attr . lower ( ) not in methods :
meta [ attr ] = new_attributes [ attr ]
methods [ attr . lower ( ) ] = attr
if " run_definition " in methods :
logger . debug ( " " )
logger . debug ( " Validating Method: run_definition " )
if meta [ methods [ " run_definition " ] ] is None :
raise NotScheduled ( " Skipped because run_definition has no value " )
logger . debug ( f " Value: { meta [ methods [ ' run_definition ' ] ] } " )
valid_options = [ " true " , " false " ] + plex . library_types
for library_type in util . get_list ( meta [ methods [ " run_definition " ] ] , lower = True ) :
if library_type not in valid_options :
2023-03-04 20:20:52 +00:00
raise Failed ( f " { self . type_str } Error: { library_type } is invalid. Options: true, false, { ' , ' . join ( plex . library_types ) } " )
2023-02-23 04:00:41 +00:00
elif library_type == " false " :
raise NotScheduled ( f " Skipped because run_definition is false " )
elif library_type != " true " and self . library and library_type != self . library . Plex . type :
raise NotScheduled ( f " Skipped because run_definition library_type: { library_type } doesn ' t match " )
2023-04-27 21:15:41 +00:00
match_data = None
match_methods = { }
if " match " in methods :
logger . debug ( " " )
logger . debug ( " Validating Method: match " )
match_data = meta [ methods [ " match " ] ]
match_methods = { mm . lower ( ) : mm for mm in match_data }
2023-02-23 16:56:38 +00:00
mapping_id = None
2023-04-27 21:15:41 +00:00
item = [ ]
if ( " mapping_id " in match_methods or " mapping_id " in methods ) and not self . library . is_music :
2023-02-23 16:56:38 +00:00
logger . debug ( " " )
logger . debug ( " Validating Method: mapping_id " )
2023-04-27 21:15:41 +00:00
value = match_data [ match_methods [ " mapping_id " ] ] if " mapping_id " in match_methods else meta [ methods [ " mapping_id " ] ]
if not value :
2023-03-04 20:20:52 +00:00
raise Failed ( f " { self . type_str } Error: mapping_id attribute is blank " )
2023-04-27 21:15:41 +00:00
logger . debug ( f " Value: { value } " )
mapping_id = value
2023-02-23 16:56:38 +00:00
2023-04-27 21:15:41 +00:00
if mapping_id is None and ( isinstance ( mapping_name , int ) or mapping_name . startswith ( " tt " ) ) and not self . library . is_music :
2023-02-23 16:56:38 +00:00
mapping_id = mapping_name
2023-04-27 21:15:41 +00:00
if mapping_id is not None :
2023-02-23 16:56:38 +00:00
if str ( mapping_id ) . startswith ( " tt " ) :
id_type = " IMDb "
else :
id_type = " TMDb " if self . library . is_movie else " TVDb "
2023-02-23 18:57:00 +00:00
logger . info ( " " )
2023-02-23 16:56:38 +00:00
logger . info ( f " { id_type } ID Mapping: { mapping_id } " )
2023-02-23 19:13:25 +00:00
if self . library . is_movie and mapping_id in self . library . movie_map :
2023-08-17 13:48:05 +00:00
item . extend ( [ self . library . fetch_item ( i ) for i in self . library . movie_map [ mapping_id ] ] )
2023-02-23 19:13:25 +00:00
elif self . library . is_show and mapping_id in self . library . show_map :
2023-08-17 13:48:05 +00:00
item . extend ( [ self . library . fetch_item ( i ) for i in self . library . show_map [ mapping_id ] ] )
2023-02-23 19:13:25 +00:00
elif mapping_id in self . library . imdb_map :
2023-08-17 13:48:05 +00:00
item . extend ( [ self . library . fetch_item ( i ) for i in self . library . imdb_map [ mapping_id ] ] )
2023-02-23 16:56:38 +00:00
else :
2023-03-04 20:20:52 +00:00
logger . error ( f " { self . type_str } Error: { id_type } ID not mapped " )
2023-02-23 16:56:38 +00:00
continue
2023-02-23 04:00:41 +00:00
2023-02-23 16:56:38 +00:00
blank_edition = False
2023-02-23 18:57:00 +00:00
edition_titles = [ ]
edition_contains = [ ]
if self . library . is_movie :
2023-04-27 21:15:41 +00:00
if " blank_edition " in match_methods or " blank_edition " in methods :
2023-02-23 19:51:14 +00:00
logger . debug ( " " )
logger . debug ( " Validating Method: blank_edition " )
2023-04-27 21:15:41 +00:00
value = match_data [ match_methods [ " blank_edition " ] ] if " blank_edition " in match_methods else meta [ methods [ " blank_edition " ] ]
logger . debug ( f " Value: { value } " )
blank_edition = util . parse ( self . type_str , " blank_edition " , value , datatype = " bool " , default = False )
if " edition " in match_methods or " edition_filter " in methods :
2023-02-23 19:51:14 +00:00
logger . debug ( " " )
logger . debug ( " Validating Method: edition_filter " )
2023-04-27 21:15:41 +00:00
value = match_data [ match_methods [ " edition " ] ] if " edition " in match_methods else meta [ methods [ " edition_filter " ] ]
logger . debug ( f " Value: { value } " )
edition_titles = util . parse ( self . type_str , " edition " , value , datatype = " strlist " )
if " edition_contains " in match_methods or " edition_contains " in methods :
2023-02-23 19:51:14 +00:00
logger . debug ( " " )
logger . debug ( " Validating Method: edition_contains " )
2023-04-27 21:15:41 +00:00
value = match_data [ match_methods [ " edition_contains " ] ] if " edition_contains " in match_methods else meta [ methods [ " edition_contains " ] ]
logger . debug ( f " Value: { value } " )
edition_contains = util . parse ( self . type_str , " edition_contains " , value , datatype = " strlist " )
2022-05-09 20:40:39 +00:00
2023-02-22 22:01:29 +00:00
if not item :
2023-04-27 21:15:41 +00:00
titles = [ ]
if " title " in match_methods :
logger . debug ( " " )
logger . debug ( " Validating Method: title " )
value = match_data [ match_methods [ " title " ] ]
if not value :
raise Failed ( f " { self . type_str } Error: title attribute is blank " )
titles . extend ( util . parse ( self . type_str , " title " , value , datatype = " strlist " ) )
if not titles :
2023-04-28 03:14:07 +00:00
titles . append ( str ( mapping_name ) )
2023-04-27 21:15:41 +00:00
if " alt_title " in methods :
logger . debug ( " " )
logger . debug ( " Validating Method: alt_title " )
value = meta [ methods [ " alt_title " ] ]
if not value :
raise Failed ( f " { self . type_str } Error: alt_title attribute is blank " )
titles . append ( value )
2023-02-23 04:00:41 +00:00
year = None
2023-04-27 21:15:41 +00:00
if " year " in match_methods or " year " in methods :
logger . debug ( " " )
logger . debug ( " Validating Method: year " )
value = match_data [ match_methods [ " year " ] ] if " year " in match_methods else meta [ methods [ " year " ] ]
2023-04-30 16:48:25 +00:00
if not value :
2023-03-04 20:20:52 +00:00
raise Failed ( f " { self . type_str } Error: year attribute is blank " )
2023-04-27 21:15:41 +00:00
logger . debug ( f " Value: { value } " )
year = util . parse ( self . type_str , " year " , value , datatype = " int " , minimum = 1800 , maximum = next_year )
for title in titles :
temp_items = self . library . search_item ( title , year = year )
item . extend ( temp_items )
2023-02-23 04:00:41 +00:00
if not item :
2023-04-27 21:15:41 +00:00
logger . error ( f " Skipping { mapping_name } : Item not found " )
2023-02-23 04:00:41 +00:00
continue
2023-04-27 21:15:41 +00:00
2023-02-23 04:00:41 +00:00
if not isinstance ( item , list ) :
item = [ item ]
2023-02-23 16:56:38 +00:00
if blank_edition or edition_titles or edition_contains :
new_item = [ ]
2023-02-23 18:57:00 +00:00
logger . trace ( " " )
logger . trace ( " Edition Filtering: " )
2023-02-23 21:05:56 +00:00
if not self . library . plex_pass :
logger . warning ( " Plex Warning: Plex Pass is required to use the Edition Field scanning filenames instead " )
2023-02-23 16:56:38 +00:00
for i in item :
2023-02-27 18:37:23 +00:00
i = self . library . reload ( i )
2023-02-23 19:51:14 +00:00
if self . library . plex_pass :
check = i . editionTitle if i . editionTitle else " "
else :
values = [ loc for loc in i . locations if loc ]
if not values :
2023-04-28 03:43:26 +00:00
raise Failed ( f " Plex Error: No Filepaths found for { i . title } " )
2023-02-23 19:51:14 +00:00
res = re . search ( r ' (?i)[ \ [ { ]edition-([^} \ ]]*) ' , values [ 0 ] )
check = res . group ( 1 ) if res else " "
2023-02-23 18:57:00 +00:00
if blank_edition and not check :
logger . trace ( f " Found { i . title } with no Edition " )
2023-02-23 16:56:38 +00:00
new_item . append ( i )
2023-02-23 18:57:00 +00:00
elif edition_titles and check in edition_titles :
logger . trace ( f " Found { i . title } with Edition: { check } " )
new_item . append ( i )
else :
found = False
if edition_contains :
for ec in edition_contains :
if ec in check :
found = True
logger . trace ( f " Found { i . title } with Edition: { check } containing { ec } " )
new_item . append ( i )
break
if not found :
if check :
logger . trace ( f " { i . title } with Edition: { check } ignored " )
else :
logger . trace ( f " { i . title } with no Edition ignored " )
item = new_item
2023-02-23 04:00:41 +00:00
for i in item :
try :
2023-02-23 18:57:00 +00:00
logger . info ( " " )
2023-02-23 04:00:41 +00:00
logger . separator ( f " Updating { i . title } " , space = False , border = False )
2023-02-23 18:57:00 +00:00
logger . info ( " " )
2023-12-13 16:40:46 +00:00
self . update_metadata_item ( i , mapping_name , meta , methods )
2023-02-23 04:00:41 +00:00
except Failed as e :
logger . error ( e )
except NotScheduled as e :
logger . info ( e )
except Failed as e :
logger . error ( e )
2021-05-07 05:06:40 +00:00
2023-12-13 16:40:46 +00:00
def update_metadata_item ( self , item , mapping_name , meta , methods ) :
2021-05-07 05:06:40 +00:00
2022-05-09 20:53:23 +00:00
updated = False
2021-05-07 05:06:40 +00:00
2022-05-09 20:53:23 +00:00
def add_edit ( name , current_item , group = None , alias = None , key = None , value = None , var_type = " str " ) :
nonlocal updated
if value or name in alias :
if value or group [ alias [ name ] ] :
if key is None : key = name
if value is None : value = group [ alias [ name ] ]
try :
current = str ( getattr ( current_item , key , " " ) )
final_value = None
if var_type == " date " :
2023-12-07 19:49:32 +00:00
try :
final_value = util . validate_date ( value , return_as = " % Y- % m- %d " )
except Failed as ei :
raise Failed ( f " { self . type_str } Error: { name } { ei } " )
2022-05-09 20:53:23 +00:00
current = current [ : - 9 ]
elif var_type == " float " :
try :
value = float ( str ( value ) )
if 0 < = value < = 10 :
final_value = value
except ValueError :
pass
if final_value is None :
2023-03-04 20:20:52 +00:00
raise Failed ( f " { self . type_str } Error: { name } attribute must be a number between 0 and 10 " )
2022-05-09 20:53:23 +00:00
elif var_type == " int " :
try :
final_value = int ( str ( value ) )
except ValueError :
pass
if final_value is None :
2023-03-04 20:20:52 +00:00
raise Failed ( f " { self . type_str } Error: { name } attribute must be an integer " )
2021-05-07 05:06:40 +00:00
else :
2022-05-09 20:53:23 +00:00
final_value = value
if current != str ( final_value ) :
if key == " title " :
current_item . editTitle ( final_value )
else :
current_item . editField ( key , final_value )
2023-12-26 22:07:38 +00:00
logger . info ( f " Metadata: { name } updated to { final_value } " )
2022-05-09 20:53:23 +00:00
updated = True
except Failed as ee :
logger . error ( ee )
else :
2023-03-04 20:20:52 +00:00
logger . error ( f " { self . type_str } Error: { name } attribute is blank " )
2021-05-07 05:06:40 +00:00
2022-05-09 20:53:23 +00:00
def finish_edit ( current_item , description ) :
nonlocal updated
if updated :
try :
2023-12-26 22:07:38 +00:00
logger . info ( f " { description } Metadata Update Successful " )
2022-05-09 20:53:23 +00:00
except BadRequest :
2023-12-26 22:07:38 +00:00
logger . error ( f " { description } Metadata Update Failed " )
2022-03-13 19:39:34 +00:00
2022-05-09 20:53:23 +00:00
tmdb_item = None
tmdb_is_movie = None
if not self . library . is_music and ( " tmdb_show " in methods or " tmdb_id " in methods ) and " tmdb_movie " in methods :
2023-03-04 20:20:52 +00:00
logger . error ( f " { self . type_str } Error: Cannot use tmdb_movie and tmdb_show when editing the same metadata item " )
2021-05-07 05:06:40 +00:00
2022-05-09 20:53:23 +00:00
if not self . library . is_music and " tmdb_show " in methods or " tmdb_id " in methods or " tmdb_movie " in methods :
try :
if " tmdb_show " in methods or " tmdb_id " in methods :
data = meta [ methods [ " tmdb_show " if " tmdb_show " in methods else " tmdb_id " ] ]
if data is None :
2023-03-04 20:20:52 +00:00
logger . error ( f " { self . type_str } Error: tmdb_show attribute is blank " )
2022-05-09 20:53:23 +00:00
else :
tmdb_is_movie = False
tmdb_item = self . config . TMDb . get_show ( util . regex_first_int ( data , " Show " ) )
elif " tmdb_movie " in methods :
if meta [ methods [ " tmdb_movie " ] ] is None :
2023-03-04 20:20:52 +00:00
logger . error ( f " { self . type_str } Error: tmdb_movie attribute is blank " )
2022-05-09 20:40:39 +00:00
else :
2022-05-09 20:53:23 +00:00
tmdb_is_movie = True
tmdb_item = self . config . TMDb . get_movie ( util . regex_first_int ( meta [ methods [ " tmdb_movie " ] ] , " Movie " ) )
except Failed as e :
logger . error ( e )
2021-05-07 05:06:40 +00:00
2022-05-09 20:53:23 +00:00
originally_available = None
original_title = None
rating = None
studio = None
tagline = None
summary = None
genres = [ ]
if tmdb_item :
originally_available = datetime . strftime ( tmdb_item . release_date if tmdb_is_movie else tmdb_item . first_air_date , " % Y- % m- %d " )
2021-05-07 05:06:40 +00:00
2022-05-09 20:53:23 +00:00
if tmdb_item . original_title != tmdb_item . title :
original_title = tmdb_item . original_title
rating = tmdb_item . vote_average
studio = tmdb_item . studio
tagline = tmdb_item . tagline if len ( tmdb_item . tagline ) > 0 else None
summary = tmdb_item . overview
genres = tmdb_item . genres
2021-05-07 05:06:40 +00:00
2023-12-13 16:40:46 +00:00
add_edit ( " title " , item , meta , methods )
2022-05-09 20:53:23 +00:00
add_edit ( " sort_title " , item , meta , methods , key = " titleSort " )
2022-09-20 02:46:52 +00:00
if self . library . is_movie :
2023-02-23 19:51:14 +00:00
if " edition " in methods and not self . library . plex_pass :
logger . error ( " Plex Error: Plex Pass is Required to edit Edition " )
else :
add_edit ( " edition " , item , meta , methods , key = " editionTitle " )
2022-05-09 20:53:23 +00:00
add_edit ( " user_rating " , item , meta , methods , key = " userRating " , var_type = " float " )
if not self . library . is_music :
add_edit ( " originally_available " , item , meta , methods , key = " originallyAvailableAt " , value = originally_available , var_type = " date " )
add_edit ( " critic_rating " , item , meta , methods , value = rating , key = " rating " , var_type = " float " )
add_edit ( " audience_rating " , item , meta , methods , key = " audienceRating " , var_type = " float " )
add_edit ( " content_rating " , item , meta , methods , key = " contentRating " )
add_edit ( " original_title " , item , meta , methods , key = " originalTitle " , value = original_title )
add_edit ( " studio " , item , meta , methods , value = studio )
add_edit ( " tagline " , item , meta , methods , value = tagline )
add_edit ( " summary " , item , meta , methods , value = summary )
for tag_edit in util . tags_to_edit [ self . library . type ] :
if self . edit_tags ( tag_edit , item , meta , methods , extra = genres if tag_edit == " genre " else None ) :
updated = True
finish_edit ( item , f " { self . library . type } : { mapping_name } " )
if self . library . type in util . advance_tags_to_edit :
advance_edits = { }
2023-02-06 20:34:40 +00:00
prefs = None
2022-05-09 20:53:23 +00:00
for advance_edit in util . advance_tags_to_edit [ self . library . type ] :
if advance_edit in methods :
if advance_edit in [ " metadata_language " , " use_original_title " ] and self . library . agent not in plex . new_plex_agents :
2023-03-04 20:20:52 +00:00
logger . error ( f " { self . type_str } Error: { advance_edit } attribute only works for with the New Plex Movie Agent and New Plex TV Agent " )
2022-05-09 20:53:23 +00:00
elif meta [ methods [ advance_edit ] ] :
ad_key , options = plex . item_advance_keys [ f " item_ { advance_edit } " ]
method_data = str ( meta [ methods [ advance_edit ] ] ) . lower ( )
2023-02-06 20:34:40 +00:00
if prefs is None :
prefs = [ p . id for p in item . preferences ( ) ]
2022-05-09 20:53:23 +00:00
if method_data not in options :
2023-03-04 20:20:52 +00:00
logger . error ( f " { self . type_str } Error: { meta [ methods [ advance_edit ] ] } { advance_edit } attribute invalid " )
2022-05-09 20:53:23 +00:00
elif ad_key in prefs and getattr ( item , ad_key ) != options [ method_data ] :
advance_edits [ ad_key ] = options [ method_data ]
2023-12-26 22:07:38 +00:00
logger . info ( f " Metadata: { advance_edit } updated to { method_data } " )
2022-05-09 20:53:23 +00:00
else :
2023-03-04 20:20:52 +00:00
logger . error ( f " { self . type_str } Error: { advance_edit } attribute is blank " )
2022-05-09 20:53:23 +00:00
if advance_edits :
if self . library . edit_advance ( item , advance_edits ) :
updated = True
2023-12-26 22:07:38 +00:00
logger . info ( f " { mapping_name } Advanced Metadata Update Successful " )
2021-12-29 15:51:22 +00:00
else :
2023-12-26 22:07:38 +00:00
logger . error ( f " { mapping_name } Advanced Metadata Update Failed " )
2021-12-29 15:51:22 +00:00
2023-03-21 07:12:50 +00:00
style_data = None
if " style_data " in methods :
style_data = meta [ methods [ " style_data " ] ]
logger . trace ( f " Style Data: { style_data } " )
asset_location , folder_name , ups = self . library . item_images ( item , meta , methods , initial = True , asset_directory = self . asset_directory + self . library . asset_directory if self . asset_directory else None , style_data = style_data )
2022-11-14 17:08:07 +00:00
if ups :
updated = True
2024-01-10 18:56:26 +00:00
if " f1_season " not in methods :
logger . info ( f " { self . library . type } : { mapping_name } Metadata Update { ' Complete ' if updated else ' Not Needed ' } " )
2021-05-07 05:06:40 +00:00
2023-03-04 20:20:52 +00:00
update_seasons = self . update_seasons
2023-02-23 04:00:41 +00:00
if " update_seasons " in methods and self . library . is_show :
logger . debug ( " " )
logger . debug ( " Validating Method: update_seasons " )
2023-02-24 15:52:23 +00:00
if not meta [ methods [ " update_seasons " ] ] :
2023-03-04 20:20:52 +00:00
logger . warning ( f " { self . type_str } Warning: update_seasons has no value and season updates will be performed " )
2023-02-24 15:52:23 +00:00
else :
logger . debug ( f " Value: { meta [ methods [ ' update_seasons ' ] ] } " )
for library_type in util . get_list ( meta [ methods [ " run_definition " ] ] , lower = True ) :
if library_type not in [ " true " , " false " ] :
2023-03-04 20:20:52 +00:00
raise Failed ( f " { self . type_str } Error: { library_type } is invalid. Options: true or false " )
2023-02-24 15:52:23 +00:00
elif library_type == " false " :
update_seasons = False
2023-02-23 04:00:41 +00:00
2023-03-04 20:20:52 +00:00
update_episodes = self . update_episodes
2023-02-23 04:00:41 +00:00
if " update_episodes " in methods and self . library . is_show :
logger . debug ( " " )
logger . debug ( " Validating Method: update_episodes " )
2023-02-24 15:52:23 +00:00
if not meta [ methods [ " update_episodes " ] ] :
2023-03-04 20:20:52 +00:00
logger . warning ( f " { self . type_str } Warning: update_episodes has no value and episode updates will be performed " )
2023-02-24 15:52:23 +00:00
else :
logger . debug ( f " Value: { meta [ methods [ ' update_episodes ' ] ] } " )
for library_type in util . get_list ( meta [ methods [ " run_definition " ] ] , lower = True ) :
if library_type not in [ " true " , " false " ] :
2023-03-04 20:20:52 +00:00
raise Failed ( f " { self . type_str } Error: { library_type } is invalid. Options: true or false " )
2023-02-24 15:52:23 +00:00
elif library_type == " false " :
update_episodes = False
2023-02-23 04:00:41 +00:00
if " seasons " in methods and self . library . is_show and ( update_seasons or update_episodes ) :
2022-05-09 20:53:23 +00:00
if not meta [ methods [ " seasons " ] ] :
2023-03-04 20:20:52 +00:00
logger . error ( f " { self . type_str } Error: seasons attribute is blank " )
2022-05-09 20:53:23 +00:00
elif not isinstance ( meta [ methods [ " seasons " ] ] , dict ) :
2023-03-04 20:20:52 +00:00
logger . error ( f " { self . type_str } Error: seasons attribute must be a dictionary " )
2022-05-09 20:53:23 +00:00
else :
seasons = { }
for season in item . seasons ( ) :
seasons [ season . title ] = season
seasons [ int ( season . index ) ] = season
for season_id , season_dict in meta [ methods [ " seasons " ] ] . items ( ) :
updated = False
logger . info ( " " )
logger . info ( f " Updating season { season_id } of { mapping_name } ... " )
if season_id in seasons :
season = seasons [ season_id ]
else :
2023-03-04 20:20:52 +00:00
logger . error ( f " { self . type_str } Error: Season: { season_id } not found " )
2022-05-09 20:53:23 +00:00
continue
season_methods = { sm . lower ( ) : sm for sm in season_dict }
2023-03-21 07:12:50 +00:00
season_style_data = None
2023-02-23 04:00:41 +00:00
if update_seasons :
add_edit ( " title " , season , season_dict , season_methods )
add_edit ( " summary " , season , season_dict , season_methods )
add_edit ( " user_rating " , season , season_dict , season_methods , key = " userRating " , var_type = " float " )
if self . edit_tags ( " label " , season , season_dict , season_methods ) :
updated = True
finish_edit ( season , f " Season: { season_id } " )
2023-03-21 07:12:50 +00:00
if style_data and " seasons " in style_data and style_data [ " seasons " ] and season_id in style_data [ " seasons " ] :
season_style_data = style_data [ " seasons " ] [ season_id ]
2023-02-23 04:00:41 +00:00
_ , _ , ups = self . library . item_images ( season , season_dict , season_methods , asset_location = asset_location ,
title = f " { item . title } Season { season . seasonNumber } " ,
2023-02-24 15:52:23 +00:00
image_name = f " Season { ' 0 ' if season . seasonNumber < 10 else ' ' } { season . seasonNumber } " ,
2023-03-21 07:12:50 +00:00
folder_name = folder_name , style_data = season_style_data )
2024-03-26 16:15:46 +00:00
advance_edits = { }
prefs = None
for advance_edit in util . advance_tags_to_edit [ " Season " ] :
if advance_edit in season_methods :
if season_dict [ season_methods [ advance_edit ] ] :
ad_key , options = plex . item_advance_keys [ f " item_ { advance_edit } " ]
method_data = str ( season_dict [ season_methods [ advance_edit ] ] ) . lower ( )
if prefs is None :
prefs = [ p . id for p in season . preferences ( ) ]
if method_data not in options :
logger . error ( f " { self . type_str } Error: { meta [ methods [ advance_edit ] ] } { advance_edit } attribute invalid " )
elif ad_key in prefs and getattr ( season , ad_key ) != options [ method_data ] :
advance_edits [ ad_key ] = options [ method_data ]
logger . info ( f " Metadata: { advance_edit } updated to { method_data } " )
else :
logger . error ( f " { self . type_str } Error: { advance_edit } attribute is blank " )
if advance_edits :
if self . library . edit_advance ( season , advance_edits ) :
updated = True
logger . info ( " Advanced Metadata Update Successful " )
else :
logger . error ( " Advanced Metadata Update Failed " )
2023-02-23 04:00:41 +00:00
if ups :
updated = True
2023-12-26 22:07:38 +00:00
logger . info ( f " Season { season_id } of { mapping_name } Metadata Update { ' Complete ' if updated else ' Not Needed ' } " )
2023-02-23 04:00:41 +00:00
if " episodes " in season_methods and update_episodes and self . library . is_show :
2022-05-09 20:53:23 +00:00
if not season_dict [ season_methods [ " episodes " ] ] :
2023-03-04 20:20:52 +00:00
logger . error ( f " { self . type_str } Error: episodes attribute is blank " )
2022-05-09 20:53:23 +00:00
elif not isinstance ( season_dict [ season_methods [ " episodes " ] ] , dict ) :
2023-03-04 20:20:52 +00:00
logger . error ( f " { self . type_str } Error: episodes attribute must be a dictionary " )
2022-05-09 20:53:23 +00:00
else :
episodes = { }
for episode in season . episodes ( ) :
episodes [ episode . title ] = episode
2023-03-14 19:31:01 +00:00
if episode . index :
episodes [ int ( episode . index ) ] = episode
2023-03-15 13:19:08 +00:00
elif episode . originallyAvailableAt :
available = episode . originallyAvailableAt
episodes [ f " { available . month : 02 } / { available . day : 02 } " ] = episode
episodes [ f " { available . month } / { available . day } " ] = episode
episodes [ f " { available . month : 02 } - { available . day : 02 } " ] = episode
episodes [ f " { available . month } - { available . day } " ] = episode
2023-02-24 15:52:23 +00:00
for episode_id , episode_dict in season_dict [ season_methods [ " episodes " ] ] . items ( ) :
2022-05-09 20:53:23 +00:00
updated = False
logger . info ( " " )
2023-02-24 15:52:23 +00:00
logger . info ( f " Updating episode { episode_id } in { season_id } of { mapping_name } ... " )
if episode_id in episodes :
episode = episodes [ episode_id ]
2022-05-09 20:53:23 +00:00
else :
2023-03-04 20:20:52 +00:00
logger . error ( f " { self . type_str } Error: Episode { episode_id } in Season { season_id } not found " )
2022-05-09 20:53:23 +00:00
continue
episode_methods = { em . lower ( ) : em for em in episode_dict }
add_edit ( " title " , episode , episode_dict , episode_methods )
add_edit ( " sort_title " , episode , episode_dict , episode_methods , key = " titleSort " )
2022-07-26 18:30:40 +00:00
add_edit ( " content_rating " , episode , episode_dict , episode_methods , key = " contentRating " )
2022-05-09 20:53:23 +00:00
add_edit ( " critic_rating " , episode , episode_dict , episode_methods , key = " rating " , var_type = " float " )
add_edit ( " audience_rating " , episode , episode_dict , episode_methods , key = " audienceRating " , var_type = " float " )
add_edit ( " user_rating " , episode , episode_dict , episode_methods , key = " userRating " , var_type = " float " )
add_edit ( " originally_available " , episode , episode_dict , episode_methods , key = " originallyAvailableAt " , var_type = " date " )
add_edit ( " summary " , episode , episode_dict , episode_methods )
for tag_edit in [ " director " , " writer " , " label " ] :
if self . edit_tags ( tag_edit , episode , episode_dict , episode_methods ) :
updated = True
2023-02-24 15:52:23 +00:00
finish_edit ( episode , f " Episode: { episode_id } in Season: { season_id } " )
2023-03-21 07:12:50 +00:00
episode_style_data = None
if season_style_data and " episodes " in season_style_data and season_style_data [ " episodes " ] and episode_id in season_style_data [ " episodes " ] :
episode_style_data = season_style_data [ " episodes " ] [ episode_id ]
2022-11-14 17:08:07 +00:00
_ , _ , ups = self . library . item_images ( episode , episode_dict , episode_methods , asset_location = asset_location ,
title = f " { item . title } { episode . seasonEpisode . upper ( ) } " ,
2023-02-24 15:52:23 +00:00
image_name = episode . seasonEpisode . upper ( ) , folder_name = folder_name ,
2023-03-21 07:12:50 +00:00
style_data = episode_style_data )
2022-11-14 17:08:07 +00:00
if ups :
updated = True
2023-12-26 22:07:38 +00:00
logger . info ( f " Episode { episode_id } in Season { season_id } of { mapping_name } Metadata Update { ' Complete ' if updated else ' Not Needed ' } " )
2022-01-11 01:23:56 +00:00
2023-02-23 04:00:41 +00:00
if " episodes " in methods and update_episodes and self . library . is_show :
2022-05-09 20:53:23 +00:00
if not meta [ methods [ " episodes " ] ] :
2023-03-04 20:20:52 +00:00
logger . error ( f " { self . type_str } Error: episodes attribute is blank " )
2022-05-09 20:53:23 +00:00
elif not isinstance ( meta [ methods [ " episodes " ] ] , dict ) :
2023-03-04 20:20:52 +00:00
logger . error ( f " { self . type_str } Error: episodes attribute must be a dictionary " )
2022-05-09 20:53:23 +00:00
else :
for episode_str , episode_dict in meta [ methods [ " episodes " ] ] . items ( ) :
updated = False
logger . info ( " " )
match = re . search ( " [Ss] \\ d+[Ee] \\ d+ " , episode_str )
if not match :
2023-03-04 20:20:52 +00:00
logger . error ( f " { self . type_str } Error: episode { episode_str } invalid must have S##E## format " )
2022-05-09 20:53:23 +00:00
continue
output = match . group ( 0 ) [ 1 : ] . split ( " E " if " E " in match . group ( 0 ) else " e " )
season_id = int ( output [ 0 ] )
episode_id = int ( output [ 1 ] )
logger . info ( f " Updating episode S { season_id } E { episode_id } of { mapping_name } ... " )
try :
episode = item . episode ( season = season_id , episode = episode_id )
except NotFound :
2023-03-04 20:20:52 +00:00
logger . error ( f " { self . type_str } Error: episode { episode_id } of season { season_id } not found " )
2022-05-09 20:53:23 +00:00
continue
episode_methods = { em . lower ( ) : em for em in episode_dict }
add_edit ( " title " , episode , episode_dict , episode_methods )
add_edit ( " sort_title " , episode , episode_dict , episode_methods , key = " titleSort " )
2022-07-26 18:30:40 +00:00
add_edit ( " content_rating " , episode , episode_dict , episode_methods , key = " contentRating " )
2022-05-09 20:53:23 +00:00
add_edit ( " critic_rating " , episode , episode_dict , episode_methods , key = " rating " , var_type = " float " )
add_edit ( " audience_rating " , episode , episode_dict , episode_methods , key = " audienceRating " , var_type = " float " )
add_edit ( " user_rating " , episode , episode_dict , episode_methods , key = " userRating " , var_type = " float " )
add_edit ( " originally_available " , episode , episode_dict , episode_methods , key = " originallyAvailableAt " , var_type = " date " )
add_edit ( " summary " , episode , episode_dict , episode_methods )
for tag_edit in [ " director " , " writer " , " label " ] :
if self . edit_tags ( tag_edit , episode , episode_dict , episode_methods ) :
updated = True
finish_edit ( episode , f " Episode: { episode_str } in Season: { season_id } " )
2022-11-14 17:08:07 +00:00
_ , _ , ups = self . library . item_images ( episode , episode_dict , episode_methods , asset_location = asset_location ,
title = f " { item . title } { episode . seasonEpisode . upper ( ) } " ,
image_name = episode . seasonEpisode . upper ( ) , folder_name = folder_name )
if ups :
updated = True
2023-12-26 22:07:38 +00:00
logger . info ( f " Episode S { season_id } E { episode_id } of { mapping_name } Metadata Update { ' Complete ' if updated else ' Not Needed ' } " )
2021-12-21 21:33:37 +00:00
2022-05-09 20:53:23 +00:00
if " albums " in methods and self . library . is_music :
if not meta [ methods [ " albums " ] ] :
2023-03-04 20:20:52 +00:00
logger . error ( f " { self . type_str } Error: albums attribute is blank " )
2022-05-09 20:53:23 +00:00
elif not isinstance ( meta [ methods [ " albums " ] ] , dict ) :
2023-03-04 20:20:52 +00:00
logger . error ( f " { self . type_str } Error: albums attribute must be a dictionary " )
2022-05-09 20:53:23 +00:00
else :
albums = { album . title : album for album in item . albums ( ) }
for album_name , album_dict in meta [ methods [ " albums " ] ] . items ( ) :
updated = False
title = None
album_methods = { am . lower ( ) : am for am in album_dict }
logger . info ( " " )
logger . info ( f " Updating album { album_name } of { mapping_name } ... " )
if album_name in albums :
album = albums [ album_name ]
elif " alt_title " in album_methods and album_dict [ album_methods [ " alt_title " ] ] and album_dict [ album_methods [ " alt_title " ] ] in albums :
album = albums [ album_dict [ album_methods [ " alt_title " ] ] ]
title = album_name
2022-03-22 17:46:25 +00:00
else :
2023-03-04 20:20:52 +00:00
logger . error ( f " { self . type_str } Error: Album: { album_name } not found " )
2022-03-21 22:18:51 +00:00
continue
2022-05-09 20:53:23 +00:00
add_edit ( " title " , album , album_dict , album_methods , value = title )
add_edit ( " sort_title " , album , album_dict , album_methods , key = " titleSort " )
add_edit ( " critic_rating " , album , album_dict , album_methods , key = " rating " , var_type = " float " )
add_edit ( " user_rating " , album , album_dict , album_methods , key = " userRating " , var_type = " float " )
add_edit ( " originally_available " , album , album_dict , album_methods , key = " originallyAvailableAt " , var_type = " date " )
add_edit ( " record_label " , album , album_dict , album_methods , key = " studio " )
add_edit ( " summary " , album , album_dict , album_methods )
for tag_edit in [ " genre " , " style " , " mood " , " collection " , " label " ] :
if self . edit_tags ( tag_edit , album , album_dict , album_methods ) :
updated = True
2022-12-13 21:12:59 +00:00
if not title :
title = album . title
2022-05-09 20:53:23 +00:00
finish_edit ( album , f " Album: { title } " )
2022-11-14 17:08:07 +00:00
_ , _ , ups = self . library . item_images ( album , album_dict , album_methods , asset_location = asset_location ,
title = f " { item . title } Album { album . title } " , image_name = album . title , folder_name = folder_name )
if ups :
updated = True
2023-12-26 22:07:38 +00:00
logger . info ( f " Album: { title } of { mapping_name } Metadata Update { ' Complete ' if updated else ' Not Needed ' } " )
2022-05-09 20:53:23 +00:00
if " tracks " in album_methods :
if not album_dict [ album_methods [ " tracks " ] ] :
2023-03-04 20:20:52 +00:00
logger . error ( f " { self . type_str } Error: tracks attribute is blank " )
2022-05-09 20:53:23 +00:00
elif not isinstance ( album_dict [ album_methods [ " tracks " ] ] , dict ) :
2023-03-04 20:20:52 +00:00
logger . error ( f " { self . type_str } Error: tracks attribute must be a dictionary " )
2022-05-09 20:53:23 +00:00
else :
tracks = { }
for track in album . tracks ( ) :
tracks [ track . title ] = track
tracks [ int ( track . index ) ] = track
for track_num , track_dict in album_dict [ album_methods [ " tracks " ] ] . items ( ) :
updated = False
title = None
track_methods = { tm . lower ( ) : tm for tm in track_dict }
logger . info ( " " )
logger . info ( f " Updating track { track_num } on { album_name } of { mapping_name } ... " )
if track_num in tracks :
track = tracks [ track_num ]
elif " alt_title " in track_methods and track_dict [ track_methods [ " alt_title " ] ] and track_dict [ track_methods [ " alt_title " ] ] in tracks :
track = tracks [ track_dict [ track_methods [ " alt_title " ] ] ]
title = track_num
else :
2023-03-04 20:20:52 +00:00
logger . error ( f " { self . type_str } Error: Track: { track_num } not found " )
2022-05-09 20:53:23 +00:00
continue
add_edit ( " title " , track , track_dict , track_methods , value = title )
add_edit ( " user_rating " , track , track_dict , track_methods , key = " userRating " , var_type = " float " )
add_edit ( " track " , track , track_dict , track_methods , key = " index " , var_type = " int " )
add_edit ( " disc " , track , track_dict , track_methods , key = " parentIndex " , var_type = " int " )
add_edit ( " original_artist " , track , track_dict , track_methods , key = " originalTitle " )
for tag_edit in [ " mood " , " collection " , " label " ] :
if self . edit_tags ( tag_edit , track , track_dict , track_methods ) :
updated = True
2022-12-13 21:12:59 +00:00
if not title :
title = track . title
2022-05-09 20:53:23 +00:00
finish_edit ( track , f " Track: { title } " )
2023-12-26 22:07:38 +00:00
logger . info ( f " Track: { track_num } on Album: { title } of { mapping_name } Metadata Update { ' Complete ' if updated else ' Not Needed ' } " )
2022-05-09 20:53:23 +00:00
if " f1_season " in methods and self . library . is_show :
f1_season = None
current_year = datetime . now ( ) . year
if meta [ methods [ " f1_season " ] ] is None :
2023-03-04 20:20:52 +00:00
raise Failed ( f " { self . type_str } Error: f1_season attribute is blank " )
2022-05-09 20:53:23 +00:00
try :
year_value = int ( str ( meta [ methods [ " f1_season " ] ] ) )
if 1950 < = year_value < = current_year :
f1_season = year_value
except ValueError :
pass
if f1_season is None :
2023-03-04 20:20:52 +00:00
raise Failed ( f " { self . type_str } Error: f1_season attribute must be an integer between 1950 and { current_year } " )
2022-05-09 20:53:23 +00:00
round_prefix = False
if " round_prefix " in methods :
if meta [ methods [ " round_prefix " ] ] is True :
round_prefix = True
else :
2023-03-04 20:20:52 +00:00
logger . error ( f " { self . type_str } Error: round_prefix must be true to do anything " )
2022-05-09 20:53:23 +00:00
shorten_gp = False
if " shorten_gp " in methods :
if meta [ methods [ " shorten_gp " ] ] is True :
shorten_gp = True
else :
2023-03-04 20:20:52 +00:00
logger . error ( f " { self . type_str } Error: shorten_gp must be true to do anything " )
2022-05-09 20:53:23 +00:00
f1_language = None
if " f1_language " in methods :
if str ( meta [ methods [ " f1_language " ] ] ) . lower ( ) in ergast . translations :
f1_language = str ( meta [ methods [ " f1_language " ] ] ) . lower ( )
else :
2024-04-22 14:20:12 +00:00
logger . error ( f " { self . type_str } Error: f1_language must be a language code Kometa has a translation for. Options: { ergast . translations } " )
2024-01-10 18:56:26 +00:00
logger . info ( f " Setting { item . title } of { self . type_str } to F1 Season { f1_season } " )
2022-05-09 20:53:23 +00:00
races = self . config . Ergast . get_races ( f1_season , f1_language )
race_lookup = { r . round : r for r in races }
2024-01-10 18:56:26 +00:00
logger . trace ( race_lookup )
2022-05-09 20:53:23 +00:00
for season in item . seasons ( ) :
if not season . seasonNumber :
continue
sprint_weekend = False
for episode in season . episodes ( ) :
if " sprint " in episode . locations [ 0 ] . lower ( ) :
sprint_weekend = True
break
if season . seasonNumber in race_lookup :
race = race_lookup [ season . seasonNumber ]
title = race . format_name ( round_prefix , shorten_gp )
updated = False
add_edit ( " title " , season , value = title )
finish_edit ( season , f " Season: { title } " )
2022-11-14 17:08:07 +00:00
_ , _ , ups = self . library . item_images ( season , { } , { } , asset_location = asset_location , title = title ,
image_name = f " Season { ' 0 ' if season . seasonNumber < 10 else ' ' } { season . seasonNumber } " , folder_name = folder_name )
if ups :
updated = True
2023-12-26 22:07:38 +00:00
logger . info ( f " Race { season . seasonNumber } of F1 Season { f1_season } : Metadata Update { ' Complete ' if updated else ' Not Needed ' } " )
2022-03-21 22:18:51 +00:00
for episode in season . episodes ( ) :
2022-05-09 20:53:23 +00:00
if len ( episode . locations ) > 0 :
ep_title , session_date = race . session_info ( episode . locations [ 0 ] , sprint_weekend )
add_edit ( " title " , episode , value = ep_title )
add_edit ( " originally_available " , episode , key = " originallyAvailableAt " , var_type = " date " , value = session_date )
finish_edit ( episode , f " Season: { season . seasonNumber } Episode: { episode . episodeNumber } " )
2022-11-14 17:08:07 +00:00
_ , _ , ups = self . library . item_images ( episode , { } , { } , asset_location = asset_location , title = ep_title ,
image_name = episode . seasonEpisode . upper ( ) , folder_name = folder_name )
if ups :
updated = True
2023-12-26 22:07:38 +00:00
logger . info ( f " Session { episode . title } : Metadata Update { ' Complete ' if updated else ' Not Needed ' } " )
2022-05-09 20:53:23 +00:00
else :
logger . warning ( f " Ergast Error: No Round: { season . seasonNumber } for Season { f1_season } " )
2022-03-21 22:18:51 +00:00
2021-12-21 21:33:37 +00:00
class PlaylistFile ( DataFile ) :
2022-04-21 05:35:07 +00:00
def __init__ ( self , config , file_type , path , temp_vars , asset_directory ) :
2024-02-28 20:45:11 +00:00
super ( ) . __init__ ( config , file_type , path , temp_vars , asset_directory , " Playlist File " )
2021-12-21 21:33:37 +00:00
self . data_type = " Playlist "
2022-04-10 01:45:33 +00:00
data = self . load_file ( self . type , self . path )
2021-12-21 21:33:37 +00:00
self . playlists = get_dict ( " playlists " , data , self . config . playlist_names )
self . templates = get_dict ( " templates " , data )
2022-04-10 04:33:59 +00:00
self . external_templates ( data )
2021-12-21 21:33:37 +00:00
if not self . playlists :
raise Failed ( " YAML Error: playlists attribute is required " )
2023-12-12 22:03:15 +00:00
logger . info ( " Playlist File Loaded Successfully " )
2022-04-18 18:16:39 +00:00
class OverlayFile ( DataFile ) :
2022-10-28 23:32:48 +00:00
def __init__ ( self , config , library , file_type , path , temp_vars , asset_directory , queue_current ) :
2024-02-28 20:45:11 +00:00
self . file_num = len ( library . overlay_files )
super ( ) . __init__ ( config , file_type , path , temp_vars , asset_directory , f " Overlay File { self . file_num } " )
2022-04-18 18:16:39 +00:00
self . library = library
self . data_type = " Overlay "
2022-09-27 06:19:29 +00:00
data = self . load_file ( self . type , self . path , overlay = True )
2022-06-23 15:31:54 +00:00
self . overlays = get_dict ( " overlays " , data )
2022-04-18 18:16:39 +00:00
self . templates = get_dict ( " templates " , data )
2022-10-28 20:48:15 +00:00
queues = get_dict ( " queues " , data )
2022-10-05 07:00:01 +00:00
self . queues = { }
2022-10-28 23:32:48 +00:00
self . queue_names = { }
2022-10-28 20:48:15 +00:00
position = temp_vars [ " position " ] if " position " in temp_vars and temp_vars [ " position " ] else None
2022-11-04 15:56:20 +00:00
overlay_limit = util . parse ( " Config " , " overlay_limit " , temp_vars [ " overlay_limit " ] , datatype = " int " , default = 0 , minimum = 0 ) if " overlay_limit " in temp_vars else None
2022-10-05 07:00:01 +00:00
for queue_name , queue in queues . items ( ) :
2022-10-28 20:48:15 +00:00
queue_position = temp_vars [ f " position_ { queue_name } " ] if f " position_ { queue_name } " in temp_vars and temp_vars [ f " position_ { queue_name } " ] else position
2022-10-27 06:39:30 +00:00
initial_queue = None
2022-10-30 07:23:29 +00:00
defaults = { " horizontal_align " : None , " vertical_align " : None , " horizontal_offset " : None , " vertical_offset " : None }
if isinstance ( queue , dict ) and " default " in queue and queue [ " default " ] and isinstance ( queue [ " default " ] , dict ) :
for k , v in queue [ " default " ] . items ( ) :
if k == " position " :
if not queue_position :
queue_position = v
2022-11-04 15:56:20 +00:00
elif k == " overlay_limit " :
if overlay_limit is None :
overlay_limit = util . parse ( " Config " , " overlay_limit " , v , datatype = " int " , default = 0 , minimum = 0 )
2022-11-07 20:00:29 +00:00
elif k == " conditionals " :
if not v :
raise Failed ( f " Queue Error: default sub-attribute conditionals is blank " )
if not isinstance ( v , dict ) :
raise Failed ( f " Queue Error: default sub-attribute conditionals is not a dictionary " )
for con_key , con_value in v . items ( ) :
if not isinstance ( con_value , dict ) :
raise Failed ( f " Queue Error: conditional { con_key } is not a dictionary " )
if " default " not in con_value :
raise Failed ( f " Queue Error: default sub-attribute required for conditional { con_key } " )
if " conditions " not in con_value :
raise Failed ( f " Queue Error: conditions sub-attribute required for conditional { con_key } " )
conditions = con_value [ " conditions " ]
if isinstance ( conditions , dict ) :
conditions = [ conditions ]
if not isinstance ( conditions , list ) :
raise Failed ( f " { self . data_type } Error: conditions sub-attribute must be a list or dictionary " )
condition_found = False
for i , condition in enumerate ( conditions , 1 ) :
if not isinstance ( condition , dict ) :
raise Failed ( f " { self . data_type } Error: each condition must be a dictionary " )
if " value " not in condition :
raise Failed ( f " { self . data_type } Error: each condition must have a result value " )
condition_passed = True
for var_key , var_value in condition . items ( ) :
if var_key == " value " :
continue
if var_key . endswith ( " .exists " ) :
var_value = util . parse ( self . data_type , var_key , var_value , datatype = " bool " , default = False )
if ( not var_value and var_key [ : - 7 ] in temp_vars and temp_vars [ var_key [ : - 7 ] ] ) or ( var_value and ( var_key [ : - 7 ] not in temp_vars or not temp_vars [ var_key [ : - 7 ] ] ) ) :
logger . debug ( f " Condition { i } Failed: { var_key } : { ' true does not exist ' if var_value else ' false exists ' } " )
condition_passed = False
elif var_key . endswith ( " .not " ) :
if ( isinstance ( var_value , list ) and temp_vars [ var_key ] in var_value ) or \
( not isinstance ( var_value , list ) and str ( temp_vars [ var_key ] ) == str ( var_value ) ) :
if isinstance ( var_value , list ) :
logger . debug ( f ' Condition { i } Failed: { var_key } " { temp_vars [ var_key ] } " in { var_value } ' )
else :
logger . debug ( f ' Condition { i } Failed: { var_key } " { temp_vars [ var_key ] } " is " { var_value } " ' )
condition_passed = False
elif var_key in temp_vars :
if ( isinstance ( var_value , list ) and temp_vars [ var_key ] not in var_value ) or \
( not isinstance ( var_value , list ) and str ( temp_vars [ var_key ] ) != str ( var_value ) ) :
if isinstance ( var_value , list ) :
logger . debug ( f ' Condition { i } Failed: { var_key } " { temp_vars [ var_key ] } " not in { var_value } ' )
else :
logger . debug ( f ' Condition { i } Failed: { var_key } " { temp_vars [ var_key ] } " is not " { var_value } " ' )
condition_passed = False
else :
logger . debug ( f " Condition { i } Failed: { var_key } is not a variable provided or a default variable " )
condition_passed = False
if condition_passed is False :
break
if condition_passed :
condition_found = True
defaults [ con_key ] = condition [ " value " ]
break
if not condition_found :
defaults [ con_key ] = con_value [ " default " ]
2022-10-30 07:23:29 +00:00
else :
defaults [ k ] = v
2022-10-28 20:48:15 +00:00
if queue_position and isinstance ( queue_position , list ) :
initial_queue = queue_position
elif isinstance ( queue , list ) :
2022-10-27 06:39:30 +00:00
initial_queue = queue
2022-10-17 06:48:08 +00:00
elif isinstance ( queue , dict ) :
2022-10-30 07:23:29 +00:00
if queue_position :
pos_str = str ( queue_position )
for x in range ( 4 ) :
dict_to_use = temp_vars if x < 2 else defaults
for k , v in dict_to_use . items ( ) :
if f " << { k } >> " in pos_str :
pos_str = pos_str . replace ( f " << { k } >> " , str ( v ) )
if pos_str in queue :
initial_queue = queue [ pos_str ]
if not initial_queue :
initial_queue = next ( ( v for k , v in queue . items ( ) if k != " default " ) , None )
2022-10-27 06:39:30 +00:00
if not isinstance ( initial_queue , list ) :
raise Failed ( f " Config Error: queue { queue_name } must be a list " )
final_queue = [ ]
for pos in initial_queue :
if not pos :
pos = { }
2022-10-30 07:23:29 +00:00
defaults [ " horizontal_align " ] = pos [ " horizontal_align " ] if " horizontal_align " in pos else defaults [ " horizontal_align " ]
defaults [ " vertical_align " ] = pos [ " vertical_align " ] if " vertical_align " in pos else defaults [ " vertical_align " ]
defaults [ " horizontal_offset " ] = pos [ " horizontal_offset " ] if " horizontal_offset " in pos else defaults [ " horizontal_offset " ]
defaults [ " vertical_offset " ] = pos [ " vertical_offset " ] if " vertical_offset " in pos else defaults [ " vertical_offset " ]
2022-10-27 06:39:30 +00:00
new_pos = {
2022-10-30 07:23:29 +00:00
" horizontal_align " : defaults [ " horizontal_align " ] , " vertical_align " : defaults [ " vertical_align " ] ,
" horizontal_offset " : defaults [ " horizontal_offset " ] , " vertical_offset " : defaults [ " vertical_offset " ]
2022-10-27 06:39:30 +00:00
}
for pk , pv in new_pos . items ( ) :
if pv is None :
raise Failed ( f " Config Error: queue missing { pv } attribute " )
2022-10-28 23:32:48 +00:00
final_queue . append ( util . parse_cords ( new_pos , f " { queue_name } queue " , required = True ) )
2022-11-04 15:56:20 +00:00
if overlay_limit and len ( final_queue ) > = overlay_limit :
break
2022-10-28 23:32:48 +00:00
self . queues [ queue_current ] = final_queue
self . queue_names [ queue_name ] = queue_current
queue_current + = 1
2022-10-17 06:48:08 +00:00
self . external_templates ( data , overlay = True )
2022-04-18 18:16:39 +00:00
if not self . overlays :
raise Failed ( " YAML Error: overlays attribute is required " )
logger . info ( f " Overlay File Loaded Successfully " )