mirror of
https://github.com/clinton-hall/nzbToMedia.git
synced 2025-07-06 21:21:12 -07:00
Refactor common utils to utils.common
This commit is contained in:
parent
844c1d15e9
commit
aa057e65d5
2 changed files with 116 additions and 120 deletions
|
@ -1,15 +1,9 @@
|
||||||
# coding=utf-8
|
# coding=utf-8
|
||||||
|
|
||||||
from __future__ import print_function, unicode_literals
|
|
||||||
|
|
||||||
import os
|
|
||||||
|
|
||||||
import requests
|
import requests
|
||||||
from six import text_type
|
|
||||||
|
|
||||||
import core
|
|
||||||
from core import logger
|
|
||||||
from core.utils import shutil_custom
|
from core.utils import shutil_custom
|
||||||
|
from core.utils.common import clean_dir, flatten, get_dirs, process_dir
|
||||||
from core.utils.download_info import get_download_info, update_download_info_status
|
from core.utils.download_info import get_download_info, update_download_info_status
|
||||||
from core.utils.encoding import char_replace, convert_to_ascii
|
from core.utils.encoding import char_replace, convert_to_ascii
|
||||||
from core.utils.files import (
|
from core.utils.files import (
|
||||||
|
@ -24,7 +18,7 @@ from core.utils.files import (
|
||||||
from core.utils.identification import category_search, find_imdbid
|
from core.utils.identification import category_search, find_imdbid
|
||||||
from core.utils.links import copy_link, replace_links
|
from core.utils.links import copy_link, replace_links
|
||||||
from core.utils.naming import clean_file_name, is_sample, sanitize_name
|
from core.utils.naming import clean_file_name, is_sample, sanitize_name
|
||||||
from core.utils.network import find_download, test_connection, wake_on_lan, wake_up, server_responding
|
from core.utils.network import find_download, server_responding, test_connection, wake_on_lan, wake_up
|
||||||
from core.utils.notifications import plex_update
|
from core.utils.notifications import plex_update
|
||||||
from core.utils.nzbs import get_nzoid, report_nzb
|
from core.utils.nzbs import get_nzoid, report_nzb
|
||||||
from core.utils.parsers import (
|
from core.utils.parsers import (
|
||||||
|
@ -52,117 +46,5 @@ from core.utils.processes import RunningProcess
|
||||||
from core.utils.subtitles import import_subs
|
from core.utils.subtitles import import_subs
|
||||||
from core.utils.torrents import create_torrent_class, pause_torrent, remove_torrent, resume_torrent
|
from core.utils.torrents import create_torrent_class, pause_torrent, remove_torrent, resume_torrent
|
||||||
|
|
||||||
try:
|
|
||||||
import jaraco
|
|
||||||
except ImportError:
|
|
||||||
if os.name == 'nt':
|
|
||||||
raise
|
|
||||||
|
|
||||||
requests.packages.urllib3.disable_warnings()
|
requests.packages.urllib3.disable_warnings()
|
||||||
|
|
||||||
shutil_custom.monkey_patch()
|
shutil_custom.monkey_patch()
|
||||||
|
|
||||||
|
|
||||||
def flatten(output_destination):
|
|
||||||
return flatten_dir(output_destination, list_media_files(output_destination))
|
|
||||||
|
|
||||||
|
|
||||||
def clean_dir(path, section, subsection):
|
|
||||||
cfg = dict(core.CFG[section][subsection])
|
|
||||||
min_size = int(cfg.get('minSize', 0))
|
|
||||||
delete_ignored = int(cfg.get('delete_ignored', 0))
|
|
||||||
try:
|
|
||||||
files = list_media_files(path, min_size=min_size, delete_ignored=delete_ignored)
|
|
||||||
except Exception:
|
|
||||||
files = []
|
|
||||||
return clean_directory(path, files)
|
|
||||||
|
|
||||||
|
|
||||||
def process_dir(path, link):
|
|
||||||
folders = []
|
|
||||||
|
|
||||||
logger.info('Searching {0} for mediafiles to post-process ...'.format(path))
|
|
||||||
dir_contents = os.listdir(text_type(path))
|
|
||||||
|
|
||||||
# search for single files and move them into their own folder for post-processing
|
|
||||||
|
|
||||||
# Generate list of sync files
|
|
||||||
sync_files = (
|
|
||||||
item for item in dir_contents
|
|
||||||
if os.path.splitext(item)[1] in ['.!sync', '.bts']
|
|
||||||
)
|
|
||||||
|
|
||||||
# Generate a list of file paths
|
|
||||||
filepaths = (
|
|
||||||
os.path.join(path, item) for item in dir_contents
|
|
||||||
if item not in ['Thumbs.db', 'thumbs.db']
|
|
||||||
)
|
|
||||||
|
|
||||||
# Generate a list of media files
|
|
||||||
mediafiles = (
|
|
||||||
item for item in filepaths
|
|
||||||
if os.path.isfile(item)
|
|
||||||
)
|
|
||||||
|
|
||||||
if any(sync_files):
|
|
||||||
logger.info('')
|
|
||||||
else:
|
|
||||||
for mediafile in mediafiles:
|
|
||||||
try:
|
|
||||||
move_file(mediafile, path, link)
|
|
||||||
except Exception as e:
|
|
||||||
logger.error('Failed to move {0} to its own directory: {1}'.format(os.path.split(mediafile)[1], e))
|
|
||||||
|
|
||||||
# removeEmptyFolders(path, removeRoot=False)
|
|
||||||
|
|
||||||
# Generate all path contents
|
|
||||||
path_contents = (
|
|
||||||
os.path.join(path, item)
|
|
||||||
for item in os.listdir(text_type(path))
|
|
||||||
)
|
|
||||||
|
|
||||||
# Generate all directories from path contents
|
|
||||||
directories = (
|
|
||||||
path for path in path_contents
|
|
||||||
if os.path.isdir(path)
|
|
||||||
)
|
|
||||||
|
|
||||||
for directory in directories:
|
|
||||||
dir_contents = os.listdir(directory)
|
|
||||||
sync_files = (
|
|
||||||
item for item in dir_contents
|
|
||||||
if os.path.splitext(item)[1] in ['.!sync', '.bts']
|
|
||||||
)
|
|
||||||
if not any(dir_contents) or any(sync_files):
|
|
||||||
continue
|
|
||||||
folders.append(directory)
|
|
||||||
|
|
||||||
return folders
|
|
||||||
|
|
||||||
|
|
||||||
def get_dirs(section, subsection, link='hard'):
|
|
||||||
to_return = []
|
|
||||||
|
|
||||||
watch_directory = core.CFG[section][subsection]['watch_dir']
|
|
||||||
directory = os.path.join(watch_directory, subsection)
|
|
||||||
|
|
||||||
if not os.path.exists(directory):
|
|
||||||
directory = watch_directory
|
|
||||||
|
|
||||||
try:
|
|
||||||
to_return.extend(process_dir(directory, link))
|
|
||||||
except Exception as e:
|
|
||||||
logger.error('Failed to add directories from {0} for post-processing: {1}'.format(watch_directory, e))
|
|
||||||
|
|
||||||
if core.USELINK == 'move':
|
|
||||||
try:
|
|
||||||
output_directory = os.path.join(core.OUTPUTDIRECTORY, subsection)
|
|
||||||
if os.path.exists(output_directory):
|
|
||||||
to_return.extend(process_dir(output_directory, link))
|
|
||||||
except Exception as e:
|
|
||||||
logger.error('Failed to add directories from {0} for post-processing: {1}'.format(core.OUTPUTDIRECTORY, e))
|
|
||||||
|
|
||||||
if not to_return:
|
|
||||||
logger.debug('No directories identified in {0}:{1} for post-processing'.format(section, subsection))
|
|
||||||
|
|
||||||
return list(set(to_return))
|
|
||||||
|
|
114
core/utils/common.py
Normal file
114
core/utils/common.py
Normal file
|
@ -0,0 +1,114 @@
|
||||||
|
|
||||||
|
import os.path
|
||||||
|
|
||||||
|
from six import text_type
|
||||||
|
|
||||||
|
import core
|
||||||
|
from core import logger
|
||||||
|
from core.utils.files import list_media_files, move_file
|
||||||
|
from core.utils.paths import clean_directory, flatten_dir
|
||||||
|
|
||||||
|
|
||||||
|
def flatten(output_destination):
|
||||||
|
return flatten_dir(output_destination, list_media_files(output_destination))
|
||||||
|
|
||||||
|
|
||||||
|
def clean_dir(path, section, subsection):
|
||||||
|
cfg = dict(core.CFG[section][subsection])
|
||||||
|
min_size = int(cfg.get('minSize', 0))
|
||||||
|
delete_ignored = int(cfg.get('delete_ignored', 0))
|
||||||
|
try:
|
||||||
|
files = list_media_files(path, min_size=min_size, delete_ignored=delete_ignored)
|
||||||
|
except Exception:
|
||||||
|
files = []
|
||||||
|
return clean_directory(path, files)
|
||||||
|
|
||||||
|
|
||||||
|
def process_dir(path, link):
|
||||||
|
folders = []
|
||||||
|
|
||||||
|
logger.info('Searching {0} for mediafiles to post-process ...'.format(path))
|
||||||
|
dir_contents = os.listdir(text_type(path))
|
||||||
|
|
||||||
|
# search for single files and move them into their own folder for post-processing
|
||||||
|
|
||||||
|
# Generate list of sync files
|
||||||
|
sync_files = (
|
||||||
|
item for item in dir_contents
|
||||||
|
if os.path.splitext(item)[1] in ['.!sync', '.bts']
|
||||||
|
)
|
||||||
|
|
||||||
|
# Generate a list of file paths
|
||||||
|
filepaths = (
|
||||||
|
os.path.join(path, item) for item in dir_contents
|
||||||
|
if item not in ['Thumbs.db', 'thumbs.db']
|
||||||
|
)
|
||||||
|
|
||||||
|
# Generate a list of media files
|
||||||
|
mediafiles = (
|
||||||
|
item for item in filepaths
|
||||||
|
if os.path.isfile(item)
|
||||||
|
)
|
||||||
|
|
||||||
|
if any(sync_files):
|
||||||
|
logger.info('')
|
||||||
|
else:
|
||||||
|
for mediafile in mediafiles:
|
||||||
|
try:
|
||||||
|
move_file(mediafile, path, link)
|
||||||
|
except Exception as e:
|
||||||
|
logger.error('Failed to move {0} to its own directory: {1}'.format(os.path.split(mediafile)[1], e))
|
||||||
|
|
||||||
|
# removeEmptyFolders(path, removeRoot=False)
|
||||||
|
|
||||||
|
# Generate all path contents
|
||||||
|
path_contents = (
|
||||||
|
os.path.join(path, item)
|
||||||
|
for item in os.listdir(text_type(path))
|
||||||
|
)
|
||||||
|
|
||||||
|
# Generate all directories from path contents
|
||||||
|
directories = (
|
||||||
|
path for path in path_contents
|
||||||
|
if os.path.isdir(path)
|
||||||
|
)
|
||||||
|
|
||||||
|
for directory in directories:
|
||||||
|
dir_contents = os.listdir(directory)
|
||||||
|
sync_files = (
|
||||||
|
item for item in dir_contents
|
||||||
|
if os.path.splitext(item)[1] in ['.!sync', '.bts']
|
||||||
|
)
|
||||||
|
if not any(dir_contents) or any(sync_files):
|
||||||
|
continue
|
||||||
|
folders.append(directory)
|
||||||
|
|
||||||
|
return folders
|
||||||
|
|
||||||
|
|
||||||
|
def get_dirs(section, subsection, link='hard'):
|
||||||
|
to_return = []
|
||||||
|
|
||||||
|
watch_directory = core.CFG[section][subsection]['watch_dir']
|
||||||
|
directory = os.path.join(watch_directory, subsection)
|
||||||
|
|
||||||
|
if not os.path.exists(directory):
|
||||||
|
directory = watch_directory
|
||||||
|
|
||||||
|
try:
|
||||||
|
to_return.extend(process_dir(directory, link))
|
||||||
|
except Exception as e:
|
||||||
|
logger.error('Failed to add directories from {0} for post-processing: {1}'.format(watch_directory, e))
|
||||||
|
|
||||||
|
if core.USELINK == 'move':
|
||||||
|
try:
|
||||||
|
output_directory = os.path.join(core.OUTPUTDIRECTORY, subsection)
|
||||||
|
if os.path.exists(output_directory):
|
||||||
|
to_return.extend(process_dir(output_directory, link))
|
||||||
|
except Exception as e:
|
||||||
|
logger.error('Failed to add directories from {0} for post-processing: {1}'.format(core.OUTPUTDIRECTORY, e))
|
||||||
|
|
||||||
|
if not to_return:
|
||||||
|
logger.debug('No directories identified in {0}:{1} for post-processing'.format(section, subsection))
|
||||||
|
|
||||||
|
return list(set(to_return))
|
Loading…
Add table
Add a link
Reference in a new issue