""" General file system methods. .. moduleauthor:: Jaisen Mathai """ from builtins import object import filecmp import hashlib import logging import os import re import shutil import time from datetime import datetime, timedelta from elodie import geolocation from elodie import log from elodie.config import load_config from elodie import constants from elodie.localstorage import Db from elodie.media.media import get_media_class, get_all_subclasses from elodie.plugins.plugins import Plugins from elodie.summary import Summary class FileSystem(object): """A class for interacting with the file system.""" def __init__(self, mode='copy', dry_run=False, exclude_regex_list=set(), logger=logging.getLogger(), day_begins=0, filter_by_ext=(), keep_folders=None, max_deep=None): # The default folder path is along the lines of 2017-06-17_01-04-14-dsc_1234-some-title.jpg self.default_file_name_definition = { 'date': '%Y-%m-%d_%H-%M-%S', 'name': '%date-%original_name-%title.%extension', } # The default folder path is along the lines of 2015-01-Jan/Chicago self.default_folder_path_definition = { 'date': '%Y-%m-%b', 'location': '%city', 'full_path': '%date/%album|%location|"{}"'.format( geolocation.__DEFAULT_LOCATION__ ), } self.cached_file_name_definition = None self.cached_folder_path_definition = None # Python3 treats the regex \s differently than Python2. # It captures some additional characters like the unicode checkmark \u2713. # See build failures in Python3 here. # https://travis-ci.org/jmathai/elodie/builds/483012902 self.whitespace_regex = '[ \t\n\r\f\v]+' self.dry_run = dry_run self.exclude_regex_list = exclude_regex_list self.mode = mode self.logger = logger self.summary = Summary() self.day_begins = day_begins self.filter_by_ext = filter_by_ext self.keep_folders = keep_folders self.max_deep = max_deep # Instantiate a plugins object self.plugins = Plugins() def create_directory(self, directory_path): """Create a directory if it does not already exist. :param str directory_name: A fully qualified path of the to create. :returns: bool """ try: if os.path.exists(directory_path): return True else: if not self.dry_run: os.makedirs(directory_path) self.logger.info(f'Create {directory_path}') return True except OSError: # OSError is thrown for cases like no permission pass return False def delete_directory_if_empty(self, directory_path): """Delete a directory only if it's empty. Instead of checking first using `len([name for name in os.listdir(directory_path)]) == 0`, we catch the OSError exception. :param str directory_name: A fully qualified path of the directory to delete. """ try: os.rmdir(directory_path) return True except OSError: pass return False def walklevel(self, src_path, maxlevel=None): """ Walk into input directory recursively until desired maxlevel source: https://stackoverflow.com/questions/229186/os-walk-without-digging-into-directories-below """ src_path = src_path.rstrip(os.path.sep) if not os.path.isdir(src_path): return None num_sep = src_path.count(os.path.sep) for root, dirs, files in os.walk(src_path): level = root.count(os.path.sep) - num_sep yield root, dirs, files, level if maxlevel is not None and level >= maxlevel: del dirs[:] def get_all_files(self, path, extensions=False, exclude_regex_list=set()): """Recursively get all files which match a path and extension. :param str path string: Path to start recursive file listing :param tuple(str) extensions: File extensions to include (whitelist) :returns: generator """ if self.filter_by_ext != () and not extensions: # Filtering files by extensions. if '%media' in self.filter_by_ext: extensions = set() subclasses = get_all_subclasses() for cls in subclasses: extensions.update(cls.extensions) else: extensions = self.filter_by_ext # Create a list of compiled regular expressions to match against the file path compiled_regex_list = [re.compile(regex) for regex in exclude_regex_list] for dirname, dirnames, filenames in os.walk(path): if dirname == os.path.join(path, '.elodie'): continue for filename in filenames: # If file extension is in `extensions` # And if file path is not in exclude regexes # Then append to the list filename_path = os.path.join(dirname, filename) if ( extensions == False or os.path.splitext(filename)[1][1:].lower() in extensions and not self.should_exclude(filename_path, compiled_regex_list, False) ): yield filename_path def get_current_directory(self): """Get the current working directory. :returns: str """ return os.getcwd() def check_for_early_morning_photos(self, date): """check for early hour photos to be grouped with previous day""" if date.hour < self.day_begins: self.logger.info('moving this photo to the previous day for\ classification purposes (day_begins=' + str(self.day_begins) + ')') date = date - timedelta(hours=date.hour+1) # push it to the day before for classificiation purposes return date def get_file_name(self, metadata): """Generate file name for a photo or video using its metadata. Originally we hardcoded the file name to include an ISO date format. We use an ISO8601-like format for the file name prefix. Instead of colons as the separator for hours, minutes and seconds we use a hyphen. https://en.wikipedia.org/wiki/ISO_8601#General_principles PR #225 made the file name customizable and fixed issues #107 #110 #111. https://github.com/jmathai/elodie/pull/225 :param media: A Photo or Video instance :type media: :class:`~elodie.media.photo.Photo` or :class:`~elodie.media.video.Video` :returns: str or None for non-photo or non-videos """ if(metadata is None): return None # Get the name template and definition. # Name template is in the form %date-%original_name-%title.%extension # Definition is in the form # [ # [('date', '%Y-%m-%d_%H-%M-%S')], # [('original_name', '')], [('title', '')], // contains a fallback # [('extension', '')] # ] name_template, definition = self.get_file_name_definition() name = name_template for parts in definition: this_value = None for this_part in parts: part, mask = this_part if part in ('date', 'day', 'month', 'year'): date = self.get_date_taken(metadata) if date is not None: this_value = date.strftime(mask) else: this_value='' break elif part in ('location', 'city', 'state', 'country'): place_name = geolocation.place_name( metadata['latitude'], metadata['longitude'], db ) location_parts = re.findall('(%[^%]+)', mask) this_value = self.parse_mask_for_location( mask, location_parts, place_name, ) break elif part in ('album', 'extension', 'title'): if metadata[part]: this_value = re.sub(self.whitespace_regex, '-', metadata[part].strip()) break elif part in ('original_name'): # First we check if we have metadata['original_name']. # We have to do this for backwards compatibility because # we original did not store this back into EXIF. if metadata[part]: this_value = os.path.splitext(metadata['original_name'])[0] else: # We didn't always store original_name so this is # for backwards compatability. # We want to remove the hardcoded date prefix we used # to add to the name. # This helps when re-running the program on file # which were already processed. this_value = re.sub( '^\d{4}-\d{2}-\d{2}_\d{2}-\d{2}-\d{2}-', '', metadata['base_name'] ) if(len(this_value) == 0): this_value = metadata['base_name'] # Lastly we want to sanitize the name this_value = re.sub(self.whitespace_regex, '-', this_value.strip()) elif part.startswith('"') and part.endswith('"'): this_value = part[1:-1] break # Here we replace the placeholder with it's corresponding value. # Check if this_value was not set so that the placeholder # can be removed completely. # For example, %title- will be replaced with '' # Else replace the placeholder (i.e. %title) with the value. if this_value is None: name = re.sub( #'[^a-z_]+%{}'.format(part), '[^a-zA-Z0-9_]+%{}'.format(part), '', name, ) else: name = re.sub( '%{}'.format(part), this_value, name, ) config = load_config(constants.CONFIG_FILE) if('File' in config and 'capitalization' in config['File'] and config['File']['capitalization'] == 'upper'): return name.upper() else: return name.lower() def get_file_name_definition(self): """Returns a list of folder definitions. Each element in the list represents a folder. Fallback folders are supported and are nested lists. Return values take the following form. [ ('date', '%Y-%m-%d'), [ ('location', '%city'), ('album', ''), ('"Unknown Location", '') ] ] :returns: list """ # If we've done this already then return it immediately without # incurring any extra work if self.cached_file_name_definition is not None: return self.cached_file_name_definition config = load_config(constants.CONFIG_FILE) # If File is in the config we assume name and its # corresponding values are also present config_file = self.default_file_name_definition if('File' in config): config_file = config['File'] # Find all subpatterns of name that map to the components of the file's # name. # I.e. %date-%original_name-%title.%extension => ['date', 'original_name', 'title', 'extension'] #noqa path_parts = re.findall( '(\%[a-z_]+)', config_file['name'] ) if not path_parts or len(path_parts) == 0: return (config_file['name'], self.default_file_name_definition) self.cached_file_name_definition = [] for part in path_parts: if part in config_file: part = part[1:] self.cached_file_name_definition.append( [(part, config_file[part])] ) else: this_part = [] for p in part.split('|'): p = p[1:] this_part.append( (p, config_file[p] if p in config_file else '') ) self.cached_file_name_definition.append(this_part) self.cached_file_name_definition = (config_file['name'], self.cached_file_name_definition) return self.cached_file_name_definition def get_folder_path_definition(self): """Returns a list of folder definitions. Each element in the list represents a folder. Fallback folders are supported and are nested lists. Return values take the following form. [ ('date', '%Y-%m-%d'), [ ('location', '%city'), ('album', ''), ('"Unknown Location", '') ] ] :returns: list """ # If we've done this already then return it immediately without # incurring any extra work if self.cached_folder_path_definition is not None: return self.cached_folder_path_definition config = load_config(constants.CONFIG_FILE) # If Directory is in the config we assume full_path and its # corresponding values (date, location) are also present config_directory = self.default_folder_path_definition if 'Directory' in config: if 'full_path' in config['Directory']: config_directory = config['Directory'] # Find all subpatterns of full_path that map to directories. # I.e. %foo/%bar => ['foo', 'bar'] # I.e. %foo/%bar|%example|"something" => ['foo', 'bar|example|"something"'] path_parts = re.findall( '(\%[^/]+)', config_directory['full_path'] ) if not path_parts or len(path_parts) == 0: return self.default_folder_path_definition self.cached_folder_path_definition = [] for part in path_parts: part = part.replace('%', '') if part in config_directory: self.cached_folder_path_definition.append( [(part, config_directory[part])] ) else: this_part = [] for p in part.split('|'): this_part.append( (p, config_directory[p] if p in config_directory else '') ) self.cached_folder_path_definition.append(this_part) return self.cached_folder_path_definition def get_folder_path(self, metadata, db, path_parts=None): """Given a media's metadata this function returns the folder path as a string. :param dict metadata: Metadata dictionary. :returns: str """ if path_parts is None: path_parts = self.get_folder_path_definition() path = [] for path_part in path_parts: # We support fallback values so that # 'album|city|"Unknown Location" # %album|%city|"Unknown Location" results in # My Album - when an album exists # Sunnyvale - when no album exists but a city exists # Unknown Location - when neither an album nor location exist for this_part in path_part: part, mask = this_part this_path = self.get_dynamic_path(part, mask, metadata, db) if this_path: path.append(this_path.strip()) # We break as soon as we have a value to append # Else we continue for fallbacks break return os.path.join(*path) def get_date_from_string(self, string, user_regex=None): # If missing datetime from EXIF data check if filename is in datetime format. # For this use a user provided regex if possible. # Otherwise assume a filename such as IMG_20160915_123456.jpg as default. if user_regex is not None: matches = re.findall(user_regex, string) else: regex = { # regex to match date format type %Y%m%d, %y%m%d, %d%m%Y, # etc... 'a': re.compile( r'.*[_-]?(?P\d{4})[_-]?(?P\d{2})[_-]?(?P\d{2})[_-]?(?P\d{2})[_-]?(?P\d{2})[_-]?(?P\d{2})'), 'b': re.compile ( '[-_./](?P\d{4})[-_.]?(?P\d{2})[-_.]?(?P\d{2})[-_./]'), # not very accurate 'c': re.compile ( '[-_./](?P\d{2})[-_.]?(?P\d{2})[-_.]?(?P\d{2})[-_./]'), 'd': re.compile ( '[-_./](?P\d{2})[-_.](?P\d{2})[-_.](?P\d{4})[-_./]') } matches = [] for i, rx in regex.items(): match = re.findall(rx, string) if match != []: if i == 'c': match = [('20'+match[0][0],match[0][1],match[0][2])] elif i == 'd': # reorder items match = [(match[0][2],match[0][1],match[0][0])] # matches = match + matches if len(match) != 1: # The time string is not uniq continue matches.append((match[0], rx)) # We want only the first match for the moment break # check if there is only one result if len(set(matches)) == 1: try: # Convert str to int date_object = tuple(map(int, matches[0][0])) time = False if len(date_object) > 3: time = True date = datetime(*date_object) except (KeyError, ValueError): return None return date return None def get_date_taken(self, metadata): ''' Get the date taken from metadata or filename :returns: datetime or None. ''' if metadata is None: return None basename = metadata['base_name'] date_original = metadata['date_original'] if metadata['original_name'] is not None: date_filename = self.get_date_from_string(metadata['original_name']) else: date_filename = self.get_date_from_string(basename) date_created = metadata['date_created'] if metadata['date_original'] is not None: if (date_filename is not None and date_filename != date_original): log.warn(f"{basename} time mark is different from {date_original}") # TODO ask for keep date taken, filename time, or neither return metadata['date_original'] elif True: if date_filename is not None: if date_created is not None and date_filename > date_created: log.warn(f"{basename} time mark is more recent than {date_created}") return date_filename if True: if date_created is not None: # TODO warm and ask for confirmation return date_created elif metadata['date_modified'] is not None: return metadata['date_modified'] def get_dynamic_path(self, part, mask, metadata, db): """Parse a specific folder's name given a mask and metadata. :param part: Name of the part as defined in the path (i.e. date from %date) :param mask: Mask representing the template for the path (i.e. %city %state :param metadata: Metadata dictionary. :returns: str """ # Each part has its own custom logic and we evaluate a single part and return # the evaluated string. if part in ('date'): # If Directory is in the config we assume full_path and its # corresponding values (date, location) are also present config_directory = self.default_folder_path_definition config = load_config(constants.CONFIG_FILE) if('Directory' in config): if 'full_path' in config['Directory']: config_directory = config['Directory'] # Get date mask from config mask = '' if 'date' in config_directory: mask = config_directory['date'] if part in ('custom'): custom_parts = re.findall('(%[a-z_]+)', mask) folder = mask for i in custom_parts: folder = folder.replace( i, self.get_dynamic_path(i[1:], i, metadata, db) ) return folder elif part in ('date', 'day', 'month', 'year'): date = self.get_date_taken(metadata) # early morning photos can be grouped with previous day date = self.check_for_early_morning_photos(date) if date is not None: return date.strftime(mask) else: return '' elif part in ('location', 'city', 'state', 'country'): place_name = geolocation.place_name( metadata['latitude'], metadata['longitude'], db ) location_parts = re.findall('(%[^%]+)', mask) parsed_folder_name = self.parse_mask_for_location( mask, location_parts, place_name, ) return parsed_folder_name elif part in ('folder'): return os.path.basename(metadata['directory_path']) elif part in ('album', 'camera_make', 'camera_model'): if metadata[part]: return metadata[part] elif part.startswith('"') and part.endswith('"'): # Fallback string return part[1:-1] return '' def parse_mask_for_location(self, mask, location_parts, place_name): """Takes a mask for a location and interpolates the actual place names. Given these parameters here are the outputs. mask=%city location_parts=[('%city','%city','city')] place_name={'city': u'Sunnyvale'} output=Sunnyvale mask=%city-%state location_parts=[('%city-','%city','city'), ('%state','%state','state')] place_name={'city': u'Sunnyvale', 'state': u'California'} output=Sunnyvale-California mask=%country location_parts=[('%country','%country','country')] place_name={'default': u'Sunnyvale', 'city': u'Sunnyvale'} output=Sunnyvale :param str mask: The location mask in the form of %city-%state, etc :param list location_parts: A list of tuples in the form of [('%city-', '%city', 'city'), ('%state', '%state', 'state')] :param dict place_name: A dictionary of place keywords and names like {'default': u'California', 'state': u'California'} :returns: str """ found = False folder_name = mask for loc_part in location_parts: # We assume the search returns a tuple of length 2. # If not then it's a bad mask in config.ini. # loc_part = '%country-random' # component_full = '%country-random' # component = '%country' # key = 'country component_full, component, key = re.search( '((%([a-z]+))[^%]*)', loc_part ).groups() if(key in place_name): found = True replace_target = component replace_with = place_name[key] else: replace_target = component_full replace_with = '' folder_name = folder_name.replace( replace_target, replace_with, ) if(not found and folder_name == ''): folder_name = place_name['default'] return folder_name def process_checksum(self, _file, db, allow_duplicate): checksum = db.checksum(_file) if(checksum is None): log.info('Could not get checksum for %s.' % _file) return None # If duplicates are not allowed then we check if we've seen this file # before via checksum. We also check that the file exists at the # location we believe it to be. # If we find a checksum match but the file doesn't exist where we # believe it to be then we write a debug log and proceed to import. checksum_file = db.get_hash(checksum) # BUG: inconsistency if file removed manually without update db if(allow_duplicate is False and checksum_file is not None): if(os.path.isfile(checksum_file)): log.info('%s already at %s.' % ( _file, checksum_file )) return None else: log.info('%s matched checksum but file not found at %s.' % ( # noqa _file, checksum_file )) return checksum def checksum(self, file_path, blocksize=65536): """Create a hash value for the given file. See http://stackoverflow.com/a/3431835/1318758. :param str file_path: Path to the file to create a hash for. :param int blocksize: Read blocks of this size from the file when creating the hash. :returns: str or None """ hasher = hashlib.sha256() with open(file_path, 'rb') as f: buf = f.read(blocksize) while len(buf) > 0: hasher.update(buf) buf = f.read(blocksize) return hasher.hexdigest() return None def checkcomp(self, src_path, dest_path): """Check file. """ src_checksum = self.checksum(src_path) if self.dry_run: return src_checksum dest_checksum = self.checksum(dest_path) if dest_checksum != src_checksum: self.logger.info(f'Source checksum and destination checksum are not the same') return False return src_checksum def sort_file(self, src_path, dest_path, remove_duplicates=True): '''Copy or move file to dest_path.''' mode = self.mode dry_run = self.dry_run # check for collisions if(src_path == dest_path): self.logger.info(f'File {dest_path} already sorted') return True if os.path.isfile(dest_path): self.logger.info(f'File {dest_path} already exist') if remove_duplicates: if filecmp.cmp(src_path, dest_path): self.logger.info(f'File in source and destination are identical. Duplicate will be ignored.') if(mode == 'move'): if not dry_run: shutil.remove(src_path) self.logger.info(f'remove: {src_path}') return True else: # name is same, but file is different self.logger.info(f'File in source and destination are different.') return False else: return False else: if(mode == 'move'): if not dry_run: # Move the processed file into the destination directory shutil.move(src_path, dest_path) self.logger.info(f'move: {src_path} -> {dest_path}') elif mode == 'copy': if not dry_run: shutil.copy2(src_path, dest_path) self.logger.info(f'copy: {src_path} -> {dest_path}') return True return False def check_file(self, src_path, dest_path, db): # Check if file remain the same checksum = self.checkcomp(src_path, dest_path) has_errors = False if checksum: if not self.dry_run: db.add_hash(checksum, dest_path) db.update_hash_db() if dest_path: self.logger.info(f'{src_path} -> {dest_path}') self.summary.append((src_path, dest_path)) else: self.logger.error(f'Files {src_path} and {dest_path} are not identical') # sys.exit(1) self.summary.append((src_path, False)) has_errors = True return self.summary, has_errors def get_files_in_path(self, path, extensions=False): """Recursively get files which match a path and extension. :param str path string: Path to start recursive file listing :param tuple(str) extensions: File extensions to include (whitelist) :returns: file_path, subdirs """ if self.filter_by_ext != () and not extensions: # Filtering files by extensions. if '%media' in self.filter_by_ext: extensions = set() subclasses = get_all_subclasses() for cls in subclasses: extensions.update(cls.extensions) else: extensions = self.filter_by_ext file_list = set() if os.path.isfile(path): if not self.should_exclude(path, self.exclude_regex_list, True): file_list.add((path, '')) # Create a list of compiled regular expressions to match against the file path compiled_regex_list = [re.compile(regex) for regex in self.exclude_regex_list] subdirs = '' for dirname, dirnames, filenames, level in self.walklevel(path, self.max_deep): if dirname == os.path.join(path, '.elodie'): continue if self.keep_folders is not None: if level < self.keep_folders: subdirs = '' else: subdirs = os.path.join(subdirs, os.path.basename(dirname)) for filename in filenames: # If file extension is in `extensions` # And if file path is not in exclude regexes # Then append to the list filename_path = os.path.join(dirname, filename) if ( extensions == False or os.path.splitext(filename)[1][1:].lower() in extensions and not self.should_exclude(filename_path, compiled_regex_list, False) ): file_list.add((filename_path, subdirs)) return file_list def sort_files(self, paths, destination, db, remove_duplicates=False, ignore_tags=set()): has_errors = False for path in paths: # some error checking if not os.path.exists(path): self.logger.error(f'Directory {path} does not exist') path = os.path.expanduser(path) conflict_file_list = set() for src_path, subdirs in self.get_files_in_path(path): # Process files media = get_media_class(src_path, ignore_tags) if media: metadata = media.get_metadata() # Get the destination path according to metadata directory_name = self.get_folder_path(metadata, db) file_name = self.get_file_name(metadata) else: # Keep same directory structure directory_name = os.path.dirname(os.path.relpath(src_path, path)) file_name = os.path.basename(src_path) dest_directory = os.path.join(destination, directory_name) dest_path = os.path.join(dest_directory, subdirs, file_name) self.create_directory(dest_directory) result = self.sort_file(src_path, dest_path, remove_duplicates) if result: self.summary, has_errors = self.check_file(src_path, dest_path, db) else: # There is conflict files conflict_file_list.add((src_path, dest_path)) for src_path, dest_path in conflict_file_list: # Try to sort the file result = self.sort_file(src_path, dest_path, remove_duplicates) if result: conflict_file_list.remove((src_path, dest_path)) else: n = 1 while not result: # Add appendix to the name pre, ext = os.path.splitext(dest_path) dest_path = pre + '_' + str(n) + ext result = self.sort_file(src_path, dest_path, remove_duplicates) if n > 100: self.logger.error(f'{self.mode}: to many append for {dest_path}...') break self.logger.info(f'Same name already exists...renaming to: {dest_path}') if result: self.summary, has_errors = self.check_file(src_path, dest_path, db) else: self.summary.append((src_path, False)) has_errors = True return self.summary, has_errors def process_file(self, _file, destination, db, media, album_from_folder, mode, **kwargs): allow_duplicate = False if('allowDuplicate' in kwargs): allow_duplicate = kwargs['allowDuplicate'] stat_info_original = os.stat(_file) metadata = media.get_metadata(album_from_folder) if(not media.is_valid()): print('%s is not a valid media file. Skipping...' % _file) return checksum = self.process_checksum(_file, db, allow_duplicate) if(checksum is None): log.info('Original checksum returned None for %s. Skipping...' % _file) return # Run `before()` for every loaded plugin and if any of them raise an exception # then we skip importing the file and log a message. plugins_run_before_status = self.plugins.run_all_before(_file, destination) if(plugins_run_before_status == False): log.warn('At least one plugin pre-run failed for %s' % _file) return directory_name = self.get_folder_path(metadata, db) dest_directory = os.path.join(destination, directory_name) file_name = self.get_file_name(metadata) dest_path = os.path.join(dest_directory, file_name) # If source and destination are identical then # we should not write the file. gh-210 # TODO Allow this for update? if(_file == dest_path): print('Final source and destination path should not be identical') return self.create_directory(dest_directory) if(mode == 'move'): stat = os.stat(_file) # Move the processed file into the destination directory shutil.move(_file, dest_path) elif mode == 'copy': shutil.copy2(_file, dest_path) if mode != 'dry_run': # Set the utime based on what the original file contained # before we made any changes. # Then set the utime on the destination file based on metadata. date_taken = self.get_date_taken(metadata) self.set_utime_from_metadata(date_taken, dest_path) media.set_original_name(dest_path) if album_from_folder: media.set_album_from_folder(dest_path) # get checksum of dest file dest_checksum = db.checksum(dest_path) db.add_hash(dest_checksum, dest_path) db.update_hash_db() # Run `after()` for every loaded plugin and if any of them raise an exception # then we skip importing the file and log a message. plugins_run_after_status = self.plugins.run_all_after(_file, destination, dest_path, metadata) if(plugins_run_after_status == False): log.warn('At least one plugin pre-run failed for %s' % _file) return return dest_path def set_utime_from_metadata(self, date_taken, file_path): """ Set the modification time on the file based on the file name. """ # Initialize date taken to what's returned from the metadata function. os.utime(file_path, (int(datetime.now().timestamp()), int(date_taken.timestamp()))) # If the folder and file name follow a time format of # YYYY-MM-DD_HH-MM-SS-IMG_0001.JPG then we override the date_taken def should_exclude(self, path, regex_list=set(), needs_compiled=False): if(len(regex_list) == 0): return False if(needs_compiled): compiled_list = [] for regex in regex_list: compiled_list.append(re.compile(regex)) regex_list = compiled_list return any(regex.search(path) for regex in regex_list)