reworked scheduler to pass server to instance to remove duplication

This commit is contained in:
Nicholas St. Germain 2018-12-01 21:31:58 -06:00
parent bf1db64b82
commit 7be718751b
6 changed files with 270 additions and 291 deletions

View file

@ -116,7 +116,7 @@ class TautulliServer(NamedTuple):
id: int = None id: int = None
url: str = None url: str = None
fallback_ip: str = None fallback_ip: str = None
apikey: str = None api_key: str = None
verify_ssl: bool = None verify_ssl: bool = None
get_activity: bool = False get_activity: bool = False
get_activity_run_seconds: int = 30 get_activity_run_seconds: int = 30
@ -319,7 +319,6 @@ class TautulliStream(NamedTuple):
transcode_progress: int = None transcode_progress: int = None
subtitle_language: str = None subtitle_language: str = None
stream_subtitle_container: str = None stream_subtitle_container: str = None
_cache_time: int = None
def geoip_download(): def geoip_download():

View file

@ -7,11 +7,11 @@ from Varken.helpers import Movie, Queue
class RadarrAPI(object): class RadarrAPI(object):
def __init__(self, servers, influx_server): def __init__(self, server, influx_server):
self.now = datetime.now(timezone.utc).astimezone().isoformat() self.now = datetime.now(timezone.utc).astimezone().isoformat()
self.influx = InfluxDBClient(influx_server.url, influx_server.port, influx_server.username, self.influx = InfluxDBClient(influx_server.url, influx_server.port, influx_server.username,
influx_server.password, 'plex2') influx_server.password, 'plex2')
self.servers = servers self.server = server
# Create session to reduce server web thread load, and globally define pageSize for all requests # Create session to reduce server web thread load, and globally define pageSize for all requests
self.session = requests.Session() self.session = requests.Session()
@ -20,87 +20,85 @@ class RadarrAPI(object):
self.influx.write_points(payload) self.influx.write_points(payload)
@logging @logging
def get_missing(self, notimplemented): def get_missing(self):
endpoint = '/api/movie' endpoint = '/api/movie'
self.now = datetime.now(timezone.utc).astimezone().isoformat() self.now = datetime.now(timezone.utc).astimezone().isoformat()
influx_payload = [] influx_payload = []
for server in self.servers: missing = []
missing = [] headers = {'X-Api-Key': self.server.api_key}
headers = {'X-Api-Key': server.api_key} get = self.session.get(self.server.url + endpoint, headers=headers, verify=self.server.verify_ssl).json()
get = self.session.get(server.url + endpoint, headers=headers, verify=server.verify_ssl).json() movies = [Movie(**movie) for movie in get]
movies = [Movie(**movie) for movie in get]
for movie in movies: for movie in movies:
if server.get_missing: if self.server.get_missing:
if not movie.downloaded and movie.isAvailable: if not movie.downloaded and movie.isAvailable:
ma = True ma = True
else: else:
ma = False ma = False
movie_name = '{} ({})'.format(movie.title, movie.year) movie_name = '{} ({})'.format(movie.title, movie.year)
missing.append((movie_name, ma, movie.tmdbId)) missing.append((movie_name, ma, movie.tmdbId))
for title, ma, mid in missing: for title, ma, mid in missing:
influx_payload.append( influx_payload.append(
{ {
"measurement": "Radarr", "measurement": "Radarr",
"tags": { "tags": {
"Missing": True, "Missing": True,
"Missing_Available": ma, "Missing_Available": ma,
"tmdbId": mid, "tmdbId": mid,
"server": server.id "server": self.server.id
}, },
"time": self.now, "time": self.now,
"fields": { "fields": {
"name": title "name": title
}
} }
) }
)
self.influx_push(influx_payload) self.influx_push(influx_payload)
@logging @logging
def get_queue(self, notimplemented): def get_queue(self):
endpoint = '/api/queue' endpoint = '/api/queue'
self.now = datetime.now(timezone.utc).astimezone().isoformat() self.now = datetime.now(timezone.utc).astimezone().isoformat()
influx_payload = [] influx_payload = []
for server in self.servers: queue = []
queue = [] headers = {'X-Api-Key': self.server.api_key}
headers = {'X-Api-Key': server.api_key} get = self.session.get(self.server.url + endpoint, headers=headers, verify=self.server.verify_ssl).json()
get = self.session.get(server.url + endpoint, headers=headers, verify=server.verify_ssl).json() for movie in get:
for movie in get: movie['movie'] = Movie(**movie['movie'])
movie['movie'] = Movie(**movie['movie']) download_queue = [Queue(**movie) for movie in get]
download_queue = [Queue(**movie) for movie in get]
for queue_item in download_queue: for queue_item in download_queue:
name = '{} ({})'.format(queue_item.movie.title, queue_item.movie.year) name = '{} ({})'.format(queue_item.movie.title, queue_item.movie.year)
if queue_item.protocol.upper() == 'USENET': if queue_item.protocol.upper() == 'USENET':
protocol_id = 1 protocol_id = 1
else: else:
protocol_id = 0 protocol_id = 0
queue.append((name, queue_item.quality['quality']['name'], queue_item.protocol.upper(), queue.append((name, queue_item.quality['quality']['name'], queue_item.protocol.upper(),
protocol_id, queue_item.id)) protocol_id, queue_item.id))
for movie, quality, protocol, protocol_id, qid in queue: for movie, quality, protocol, protocol_id, qid in queue:
influx_payload.append( influx_payload.append(
{ {
"measurement": "Radarr", "measurement": "Radarr",
"tags": { "tags": {
"type": "Queue", "type": "Queue",
"tmdbId": qid, "tmdbId": qid,
"server": server.id "server": self.server.id
}, },
"time": self.now, "time": self.now,
"fields": { "fields": {
"name": movie, "name": movie,
"quality": quality, "quality": quality,
"protocol": protocol, "protocol": protocol,
"protocol_id": protocol_id "protocol_id": protocol_id
}
} }
) }
)
self.influx_push(influx_payload) self.influx_push(influx_payload)

View file

@ -7,147 +7,141 @@ from Varken.helpers import TVShow, Queue
class SonarrAPI(object): class SonarrAPI(object):
def __init__(self, servers, influx_server): def __init__(self, server, influx_server):
# Set Time of initialization # Set Time of initialization
self.now = datetime.now(timezone.utc).astimezone().isoformat() self.now = datetime.now(timezone.utc).astimezone().isoformat()
self.today = str(date.today()) self.today = str(date.today())
self.influx = InfluxDBClient(influx_server.url, influx_server.port, influx_server.username, self.influx = InfluxDBClient(influx_server.url, influx_server.port, influx_server.username,
influx_server.password, 'plex') influx_server.password, 'plex')
self.servers = servers self.server = server
# Create session to reduce server web thread load, and globally define pageSize for all requests # Create session to reduce server web thread load, and globally define pageSize for all requests
self.session = requests.Session() self.session = requests.Session()
self.session.params = {'pageSize': 1000} self.session.params = {'pageSize': 1000}
@logging @logging
def get_missing(self, days_past): def get_missing(self):
endpoint = '/api/calendar' endpoint = '/api/calendar'
last_days = str(date.today() + timedelta(days=-days_past)) last_days = str(date.today() + timedelta(days=-self.server.missing_days))
self.now = datetime.now(timezone.utc).astimezone().isoformat() self.now = datetime.now(timezone.utc).astimezone().isoformat()
params = {'start': last_days, 'end': self.today} params = {'start': last_days, 'end': self.today}
influx_payload = [] influx_payload = []
missing = []
headers = {'X-Api-Key': self.server.api_key}
for server in self.servers: get = self.session.get(self.server.url + endpoint, params=params, headers=headers,
missing = [] verify=self.server.verify_ssl).json()
headers = {'X-Api-Key': server.api_key} # Iteratively create a list of TVShow Objects from response json
tv_shows = [TVShow(**show) for show in get]
get = self.session.get(server.url + endpoint, params=params, headers=headers, # Add show to missing list if file does not exist
verify=server.verify_ssl).json() for show in tv_shows:
# Iteratively create a list of TVShow Objects from response json if not show.hasFile:
tv_shows = [TVShow(**show) for show in get] sxe = 'S{:0>2}E{:0>2}'.format(show.seasonNumber, show.episodeNumber)
missing.append((show.series['title'], sxe, show.airDate, show.title, show.id))
# Add show to missing list if file does not exist for series_title, sxe, air_date, episode_title, sonarr_id in missing:
for show in tv_shows: influx_payload.append(
if not show.hasFile: {
sxe = 'S{:0>2}E{:0>2}'.format(show.seasonNumber, show.episodeNumber) "measurement": "Sonarr",
missing.append((show.series['title'], sxe, show.airDate, show.title, show.id)) "tags": {
"type": "Missing",
for series_title, sxe, air_date, episode_title, sonarr_id in missing: "sonarrId": sonarr_id,
influx_payload.append( "server": self.server.id
{ },
"measurement": "Sonarr", "time": self.now,
"tags": { "fields": {
"type": "Missing", "name": series_title,
"sonarrId": sonarr_id, "epname": episode_title,
"server": server.id "sxe": sxe,
}, "airs": air_date
"time": self.now,
"fields": {
"name": series_title,
"epname": episode_title,
"sxe": sxe,
"airs": air_date
}
} }
) }
)
self.influx_push(influx_payload) self.influx_push(influx_payload)
@logging @logging
def get_future(self, future_days): def get_future(self):
endpoint = '/api/calendar/' endpoint = '/api/calendar/'
self.now = datetime.now(timezone.utc).astimezone().isoformat() self.now = datetime.now(timezone.utc).astimezone().isoformat()
future = str(date.today() + timedelta(days=future_days)) future = str(date.today() + timedelta(days=self.server.future_days))
influx_payload = [] influx_payload = []
air_days = []
headers = {'X-Api-Key': self.server.api_key}
params = {'start': self.today, 'end': future}
for server in self.servers: get = self.session.get(self.server.url + endpoint, params=params, headers=headers,
air_days = [] verify=self.server.verify_ssl).json()
tv_shows = [TVShow(**show) for show in get]
headers = {'X-Api-Key': server.api_key} for show in tv_shows:
params = {'start': self.today, 'end': future} sxe = 'S{:0>2}E{:0>2}'.format(show.seasonNumber, show.episodeNumber)
air_days.append((show.series['title'], show.hasFile, sxe, show.title, show.airDate, show.id))
get = self.session.get(server.url + endpoint, params=params, headers=headers, for series_title, dl_status, sxe, episode_title, air_date, sonarr_id in air_days:
verify=server.verify_ssl).json() influx_payload.append(
tv_shows = [TVShow(**show) for show in get] {
"measurement": "Sonarr",
for show in tv_shows: "tags": {
sxe = 'S{:0>2}E{:0>2}'.format(show.seasonNumber, show.episodeNumber) "type": "Future",
air_days.append((show.series['title'], show.hasFile, sxe, show.title, show.airDate, show.id)) "sonarrId": sonarr_id,
"server": self.server.id
for series_title, dl_status, sxe, episode_title, air_date, sonarr_id in air_days: },
influx_payload.append( "time": self.now,
{ "fields": {
"measurement": "Sonarr", "name": series_title,
"tags": { "epname": episode_title,
"type": "Future", "sxe": sxe,
"sonarrId": sonarr_id, "airs": air_date,
"server": server.id "downloaded": dl_status
},
"time": self.now,
"fields": {
"name": series_title,
"epname": episode_title,
"sxe": sxe,
"airs": air_date,
"downloaded": dl_status
}
} }
) }
)
self.influx_push(influx_payload) self.influx_push(influx_payload)
@logging @logging
def get_queue(self, notimplemented): def get_queue(self):
influx_payload = [] influx_payload = []
endpoint = '/api/queue' endpoint = '/api/queue'
self.now = datetime.now(timezone.utc).astimezone().isoformat() self.now = datetime.now(timezone.utc).astimezone().isoformat()
queue = []
headers = {'X-Api-Key': self.server.api_key}
for server in self.servers: get = self.session.get(self.server.url + endpoint, headers=headers, verify=self.server.verify_ssl).json()
queue = [] download_queue = [Queue(**show) for show in get]
headers = {'X-Api-Key': server.api_key}
get = self.session.get(server.url + endpoint, headers=headers, verify=server.verify_ssl).json() for show in download_queue:
download_queue = [Queue(**show) for show in get] sxe = 'S{:0>2}E{:0>2}'.format(show.episode['seasonNumber'], show.episode['episodeNumber'])
if show.protocol.upper() == 'USENET':
protocol_id = 1
else:
protocol_id = 0
for show in download_queue: queue.append((show.series['title'], show.episode['title'], show.protocol.upper(),
sxe = 'S{:0>2}E{:0>2}'.format(show.episode['seasonNumber'], show.episode['episodeNumber']) protocol_id, sxe, show.id))
if show.protocol.upper() == 'USENET':
protocol_id = 1
else:
protocol_id = 0
queue.append((show.series['title'], show.episode['title'], show.protocol.upper(), for series_title, episode_title, protocol, protocol_id, sxe, sonarr_id in queue:
protocol_id, sxe, show.id)) influx_payload.append(
{
"measurement": "Sonarr",
"tags": {
"type": "Queue",
"sonarrId": sonarr_id,
"server": self.server.id
for series_title, episode_title, protocol, protocol_id, sxe, sonarr_id in queue: },
influx_payload.append( "time": self.now,
{ "fields": {
"measurement": "Sonarr", "name": series_title,
"tags": { "epname": episode_title,
"type": "Queue", "sxe": sxe,
"sonarrId": sonarr_id, "protocol": protocol,
"server": server.id "protocol_id": protocol_id
},
"time": self.now,
"fields": {
"name": series_title,
"epname": episode_title,
"sxe": sxe,
"protocol": protocol,
"protocol_id": protocol_id
}
} }
) }
)
self.influx_push(influx_payload) self.influx_push(influx_payload)

View file

@ -7,12 +7,12 @@ from Varken.logger import logging
class TautulliAPI(object): class TautulliAPI(object):
def __init__(self, servers, influx_server): def __init__(self, server, influx_server):
# Set Time of initialization # Set Time of initialization
self.now = datetime.now(timezone.utc).astimezone().isoformat() self.now = datetime.now(timezone.utc).astimezone().isoformat()
self.influx = InfluxDBClient(influx_server.url, influx_server.port, influx_server.username, self.influx = InfluxDBClient(influx_server.url, influx_server.port, influx_server.username,
influx_server.password, 'plex2') influx_server.password, 'plex2')
self.servers = servers self.server = server
self.session = requests.Session() self.session = requests.Session()
self.endpoint = '/api/v2' self.endpoint = '/api/v2'
@ -21,128 +21,124 @@ class TautulliAPI(object):
self.influx.write_points(payload) self.influx.write_points(payload)
@logging @logging
def get_activity(self, notimplemented): def get_activity(self):
self.now = datetime.now(timezone.utc).astimezone().isoformat() self.now = datetime.now(timezone.utc).astimezone().isoformat()
params = {'cmd': 'get_activity'} params = {'cmd': 'get_activity'}
influx_payload = [] influx_payload = []
params['apikey'] = self.server.api_key
g = self.session.get(self.server.url + self.endpoint, params=params, verify=self.server.verify_ssl)
get = g.json()['response']['data']
for server in self.servers: influx_payload.append(
params['apikey'] = server.apikey {
g = self.session.get(server.url + self.endpoint, params=params, verify=server.verify_ssl) "measurement": "Tautulli",
get = g.json()['response']['data'] "tags": {
"type": "current_stream_stats",
"server": self.server.id
},
"time": self.now,
"fields": {
"stream_count": int(get['stream_count']),
"total_bandwidth": int(get['total_bandwidth']),
"wan_bandwidth": int(get['wan_bandwidth']),
"lan_bandwidth": int(get['lan_bandwidth']),
"transcode_streams": int(get['stream_count_transcode']),
"direct_play_streams": int(get['stream_count_direct_play']),
"direct_streams": int(get['stream_count_direct_stream'])
}
}
)
self.influx_push(influx_payload)
@logging
def get_sessions(self):
self.now = datetime.now(timezone.utc).astimezone().isoformat()
params = {'cmd': 'get_activity'}
influx_payload = []
params['apikey'] = self.server.api_key
g = self.session.get(self.server.url + self.endpoint, params=params, verify=self.server.verify_ssl)
get = g.json()['response']['data']['sessions']
sessions = [TautulliStream(**session) for session in get]
for session in sessions:
try:
geodata = geo_lookup(session.ip_address_public)
except (ValueError, AddressNotFoundError):
if self.server.fallback_ip:
geodata = geo_lookup(self.server.fallback_ip)
else:
my_ip = requests.get('http://ip.42.pl/raw').text
geodata = geo_lookup(my_ip)
if not all([geodata.location.latitude, geodata.location.longitude]):
latitude = 37.234332396
longitude = -115.80666344
else:
latitude = geodata.location.latitude
longitude = geodata.location.longitude
decision = session.transcode_decision
if decision == 'copy':
decision = 'direct stream'
video_decision = session.stream_video_decision
if video_decision == 'copy':
video_decision = 'direct stream'
elif video_decision == '':
video_decision = 'Music'
quality = session.stream_video_resolution
if not quality:
quality = session.container.upper()
elif quality in ('SD', 'sd', '4k'):
quality = session.stream_video_resolution.upper()
else:
quality = session.stream_video_resolution + 'p'
player_state = session.state.lower()
if player_state == 'playing':
player_state = 0
elif player_state == 'paused':
player_state = 1
elif player_state == 'buffering':
player_state = 3
influx_payload.append( influx_payload.append(
{ {
"measurement": "Tautulli", "measurement": "Tautulli",
"tags": { "tags": {
"type": "current_stream_stats", "type": "Session",
"server": server.id "session_id": session.session_id,
"name": session.friendly_name,
"title": session.full_title,
"platform": session.platform,
"product_version": session.product_version,
"quality": quality,
"video_decision": video_decision.title(),
"transcode_decision": decision.title(),
"media_type": session.media_type.title(),
"audio_codec": session.audio_codec.upper(),
"audio_profile": session.audio_profile.upper(),
"stream_audio_codec": session.stream_audio_codec.upper(),
"quality_profile": session.quality_profile,
"progress_percent": session.progress_percent,
"region_code": geodata.subdivisions.most_specific.iso_code,
"location": geodata.city.name,
"full_location": '{} - {}'.format(geodata.subdivisions.most_specific.name,
geodata.city.name),
"latitude": latitude,
"longitude": longitude,
"player_state": player_state,
"device_type": session.platform,
"server": self.server.id
}, },
"time": self.now, "time": self.now,
"fields": { "fields": {
"stream_count": int(get['stream_count']), "session_id": session.session_id,
"total_bandwidth": int(get['total_bandwidth']), "session_key": session.session_key
"wan_bandwidth": int(get['wan_bandwidth']),
"lan_bandwidth": int(get['lan_bandwidth']),
"transcode_streams": int(get['stream_count_transcode']),
"direct_play_streams": int(get['stream_count_direct_play']),
"direct_streams": int(get['stream_count_direct_stream'])
} }
} }
) )
self.influx_push(influx_payload) self.influx_push(influx_payload)
@logging
def get_sessions(self, notimplemented):
self.now = datetime.now(timezone.utc).astimezone().isoformat()
params = {'cmd': 'get_activity'}
influx_payload = []
for server in self.servers:
params['apikey'] = server.apikey
g = self.session.get(server.url + self.endpoint, params=params, verify=server.verify_ssl)
get = g.json()['response']['data']['sessions']
sessions = [TautulliStream(**session) for session in get]
for session in sessions:
try:
geodata = geo_lookup(session.ip_address_public)
except (ValueError, AddressNotFoundError):
if server.fallback_ip:
geodata = geo_lookup(server.fallback_ip)
else:
my_ip = requests.get('http://ip.42.pl/raw').text
geodata = geo_lookup(my_ip)
if not all([geodata.location.latitude, geodata.location.longitude]):
latitude = 37.234332396
longitude = -115.80666344
else:
latitude = geodata.location.latitude
longitude = geodata.location.longitude
decision = session.transcode_decision
if decision == 'copy':
decision = 'direct stream'
video_decision = session.stream_video_decision
if video_decision == 'copy':
video_decision = 'direct stream'
elif video_decision == '':
video_decision = 'Music'
quality = session.stream_video_resolution
if not quality:
quality = session.container.upper()
elif quality in ('SD', 'sd', '4k'):
quality = session.stream_video_resolution.upper()
else:
quality = session.stream_video_resolution + 'p'
player_state = session.state.lower()
if player_state == 'playing':
player_state = 0
elif player_state == 'paused':
player_state = 1
elif player_state == 'buffering':
player_state = 3
influx_payload.append(
{
"measurement": "Tautulli",
"tags": {
"type": "Session",
"session_id": session.session_id,
"name": session.friendly_name,
"title": session.full_title,
"platform": session.platform,
"product_version": session.product_version,
"quality": quality,
"video_decision": video_decision.title(),
"transcode_decision": decision.title(),
"media_type": session.media_type.title(),
"audio_codec": session.audio_codec.upper(),
"audio_profile": session.audio_profile.upper(),
"stream_audio_codec": session.stream_audio_codec.upper(),
"quality_profile": session.quality_profile,
"progress_percent": session.progress_percent,
"region_code": geodata.subdivisions.most_specific.iso_code,
"location": geodata.city.name,
"full_location": '{} - {}'.format(geodata.subdivisions.most_specific.name,
geodata.city.name),
"latitude": latitude,
"longitude": longitude,
"player_state": player_state,
"device_type": session.platform,
"server": server.id
},
"time": self.now,
"fields": {
"session_id": session.session_id,
"session_key": session.session_key
}
}
)
self.influx_push(influx_payload)

View file

@ -61,7 +61,6 @@ verify_ssl = true
queue = true queue = true
queue_run_seconds = 300 queue_run_seconds = 300
get_missing = true get_missing = true
get_missing_available = true
get_missing_run_seconds = 300 get_missing_run_seconds = 300
[radarr-2] [radarr-2]

View file

@ -8,8 +8,8 @@ from Varken.tautulli import TautulliAPI
from Varken.radarr import RadarrAPI from Varken.radarr import RadarrAPI
def threaded(job, days=None): def threaded(job):
thread = threading.Thread(target=job, args=([days])) thread = threading.Thread(target=job)
thread.start() thread.start()
@ -17,38 +17,31 @@ if __name__ == "__main__":
CONFIG = INIParser() CONFIG = INIParser()
if CONFIG.sonarr_enabled: if CONFIG.sonarr_enabled:
SONARR = SonarrAPI(CONFIG.sonarr_servers, CONFIG.influx_server)
for server in CONFIG.sonarr_servers: for server in CONFIG.sonarr_servers:
SONARR = SonarrAPI(server, CONFIG.influx_server)
if server.queue: if server.queue:
schedule.every(server.queue_run_seconds).seconds.do(threaded, SONARR.get_queue) schedule.every(server.queue_run_seconds).seconds.do(threaded, SONARR.get_queue)
if server.missing_days > 0: if server.missing_days > 0:
schedule.every(server.missing_days_run_seconds).seconds.do(threaded, SONARR.get_missing, schedule.every(server.missing_days_run_seconds).seconds.do(threaded, SONARR.get_missing)
server.missing_days)
if server.future_days > 0: if server.future_days > 0:
schedule.every(server.future_days_run_seconds).seconds.do(threaded, SONARR.get_future, schedule.every(server.future_days_run_seconds).seconds.do(threaded, SONARR.get_future)
server.future_days)
if CONFIG.tautulli_enabled: if CONFIG.tautulli_enabled:
TAUTULLI = TautulliAPI(CONFIG.tautulli_servers, CONFIG.influx_server)
for server in CONFIG.tautulli_servers: for server in CONFIG.tautulli_servers:
TAUTULLI = TautulliAPI(server, CONFIG.influx_server)
if server.get_activity: if server.get_activity:
schedule.every(server.get_activity_run_seconds).seconds.do(threaded, TAUTULLI.get_activity) schedule.every(server.get_activity_run_seconds).seconds.do(threaded, TAUTULLI.get_activity)
if server.get_sessions: if server.get_sessions:
schedule.every(server.get_sessions_run_seconds).seconds.do(threaded, TAUTULLI.get_sessions) schedule.every(server.get_sessions_run_seconds).seconds.do(threaded, TAUTULLI.get_sessions)
if CONFIG.radarr_enabled: if CONFIG.radarr_enabled:
RADARR = RadarrAPI(CONFIG.radarr_servers, CONFIG.influx_server)
for server in CONFIG.radarr_servers: for server in CONFIG.radarr_servers:
if any([server.get_missing, server.get_missing_available]): RADARR = RadarrAPI(server, CONFIG.influx_server)
if server.get_missing:
schedule.every(server.get_missing_run_seconds).seconds.do(threaded, RADARR.get_missing) schedule.every(server.get_missing_run_seconds).seconds.do(threaded, RADARR.get_missing)
if server.queue: if server.queue:
schedule.every(server.queue_run_seconds).seconds.do(threaded, RADARR.get_queue) schedule.every(server.queue_run_seconds).seconds.do(threaded, RADARR.get_queue)
while True: while True:
schedule.run_pending() schedule.run_pending()
sleep(1) sleep(1)