Skip to content

Commit

Permalink
refactor and test the callback receiver as a base for a task dispatcher
Browse files Browse the repository at this point in the history
  • Loading branch information
ryanpetrello committed Oct 11, 2018
1 parent 8ad4643 commit da74f1d
Show file tree
Hide file tree
Showing 7 changed files with 370 additions and 231 deletions.
Empty file added awx/main/dispatch/__init__.py
Empty file.
97 changes: 97 additions & 0 deletions awx/main/dispatch/pool.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,97 @@
import errno
import logging
import os
import signal
import traceback

from multiprocessing import Process
from multiprocessing import Queue as MPQueue
from Queue import Full as QueueFull

from django.conf import settings
from django.db import connection as django_connection
from django.core.cache import cache as django_cache

logger = logging.getLogger('awx.main.dispatch')


def signame(sig):
return dict(
(k, v) for v, k in signal.__dict__.items()
if v.startswith('SIG') and not v.startswith('SIG_')
)[sig]


class WorkerPool(object):

def __init__(self, min_workers=None, queue_size=None):
self.min_workers = min_workers or settings.JOB_EVENT_WORKERS
self.queue_size = queue_size or settings.JOB_EVENT_MAX_QUEUE_SIZE

# self.workers tracks the state of worker running worker processes:
# [
# (total_messages_consumed, multiprocessing.Queue, multiprocessing.Process),
# (total_messages_consumed, multiprocessing.Queue, multiprocessing.Process),
# (total_messages_consumed, multiprocessing.Queue, multiprocessing.Process),
# (total_messages_consumed, multiprocessing.Queue, multiprocessing.Process)
# ]
self.workers = []

def __len__(self):
return len(self.workers)

def init_workers(self, target, *target_args):
def shutdown_handler(active_workers):
def _handler(signum, frame):
logger.debug('received shutdown {}'.format(signame(signum)))
try:
for active_worker in active_workers:
logger.debug('terminating worker')
signal.signal(signum, signal.SIG_DFL)
os.kill(os.getpid(), signum) # Rethrow signal, this time without catching it
except Exception:
logger.exception('error in shutdown_handler')
return _handler

django_connection.close()
django_cache.close()
for idx in range(self.min_workers):
queue_actual = MPQueue(self.queue_size)
w = Process(target=target, args=(queue_actual, idx,) + target_args)
w.start()
logger.debug('started {}[{}]'.format(target.im_self.__class__.__name__, idx))
self.workers.append([0, queue_actual, w])

signal.signal(signal.SIGINT, shutdown_handler([p[2] for p in self.workers]))
signal.signal(signal.SIGTERM, shutdown_handler([p[2] for p in self.workers]))

def write(self, preferred_queue, body):
queue_order = sorted(range(self.min_workers), cmp=lambda x, y: -1 if x==preferred_queue else 0)
write_attempt_order = []
for queue_actual in queue_order:
try:
worker_actual = self.workers[queue_actual]
worker_actual[1].put(body, block=True, timeout=5)
logger.debug('delivered to Worker[{}] qsize {}'.format(
queue_actual, worker_actual[1].qsize()
))
worker_actual[0] += 1
return queue_actual
except QueueFull:
pass
except Exception:
tb = traceback.format_exc()
logger.warn("could not write to queue %s" % preferred_queue)
logger.warn("detail: {}".format(tb))
write_attempt_order.append(preferred_queue)
logger.warn("could not write payload to any queue, attempted order: {}".format(write_attempt_order))
return None

def stop(self):
for worker in self.workers:
messages, queue, process = worker
try:
os.kill(process.pid, signal.SIGTERM)
except OSError as e:
if e.errno != errno.ESRCH:
raise
173 changes: 173 additions & 0 deletions awx/main/dispatch/worker.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,173 @@
# Copyright (c) 2018 Ansible by Red Hat
# All Rights Reserved.

import logging
import os
import signal
import time
import traceback
from uuid import UUID
from Queue import Empty as QueueEmpty

from kombu.mixins import ConsumerMixin
from django.conf import settings
from django.db import DatabaseError, OperationalError, connection as django_connection
from django.db.utils import InterfaceError, InternalError

from awx.main.models import (JobEvent, AdHocCommandEvent, ProjectUpdateEvent,
InventoryUpdateEvent, SystemJobEvent, UnifiedJob)
from awx.main.consumers import emit_channel_notification
from awx.main.dispatch.pool import WorkerPool

logger = logging.getLogger('awx.main.dispatch')


class WorkerSignalHandler:

def __init__(self):
self.kill_now = False
signal.signal(signal.SIGINT, self.exit_gracefully)
signal.signal(signal.SIGTERM, self.exit_gracefully)

def exit_gracefully(self, *args, **kwargs):
self.kill_now = True


class AWXConsumer(ConsumerMixin):

def __init__(self, connection, worker, queues=[]):
self.connection = connection
self.total_messages = 0
self.queues = queues
self.pool = WorkerPool()
self.pool.init_workers(worker.work_loop)

def get_consumers(self, Consumer, channel):
return [Consumer(queues=self.queues, accept=['json'],
callbacks=[self.process_task])]

def process_task(self, body, message):
if "uuid" in body and body['uuid']:
try:
queue = UUID(body['uuid']).int % len(self.pool)
except Exception:
queue = self.total_messages % len(self.pool)
else:
queue = self.total_messages % len(self.pool)
self.pool.write(queue, body)
self.total_messages += 1
message.ack()


class BaseWorker(object):

def work_loop(self, queue, idx, *args):
signal_handler = WorkerSignalHandler()
while not signal_handler.kill_now:
try:
body = queue.get(block=True, timeout=1)
except QueueEmpty:
continue
except Exception as e:
logger.error("Exception on worker, restarting: " + str(e))
continue
self.perform_work(body, *args)

def perform_work(self, body):
raise NotImplemented()


class CallbackBrokerWorker(BaseWorker):

MAX_RETRIES = 2

def perform_work(self, body):
try:
event_map = {
'job_id': JobEvent,
'ad_hoc_command_id': AdHocCommandEvent,
'project_update_id': ProjectUpdateEvent,
'inventory_update_id': InventoryUpdateEvent,
'system_job_id': SystemJobEvent,
}

if not any([key in body for key in event_map]):
raise Exception('Payload does not have a job identifier')
if settings.DEBUG:
from pygments import highlight
from pygments.lexers import PythonLexer
from pygments.formatters import Terminal256Formatter
from pprint import pformat
logger.info('Body: {}'.format(
highlight(pformat(body, width=160), PythonLexer(), Terminal256Formatter(style='friendly'))
)[:1024 * 4])

def _save_event_data():
for key, cls in event_map.items():
if key in body:
cls.create_from_data(**body)

job_identifier = 'unknown job'
for key in event_map.keys():
if key in body:
job_identifier = body[key]
break

if body.get('event') == 'EOF':
try:
final_counter = body.get('final_counter', 0)
logger.info('Event processing is finished for Job {}, sending notifications'.format(job_identifier))
# EOF events are sent when stdout for the running task is
# closed. don't actually persist them to the database; we
# just use them to report `summary` websocket events as an
# approximation for when a job is "done"
emit_channel_notification(
'jobs-summary',
dict(group_name='jobs', unified_job_id=job_identifier, final_counter=final_counter)
)
# Additionally, when we've processed all events, we should
# have all the data we need to send out success/failure
# notification templates
uj = UnifiedJob.objects.get(pk=job_identifier)
if hasattr(uj, 'send_notification_templates'):
retries = 0
while retries < 5:
if uj.finished:
uj.send_notification_templates('succeeded' if uj.status == 'successful' else 'failed')
break
else:
# wait a few seconds to avoid a race where the
# events are persisted _before_ the UJ.status
# changes from running -> successful
retries += 1
time.sleep(1)
uj = UnifiedJob.objects.get(pk=job_identifier)
except Exception:
logger.exception('Worker failed to emit notifications: Job {}'.format(job_identifier))
return

retries = 0
while retries <= self.MAX_RETRIES:
try:
_save_event_data()
break
except (OperationalError, InterfaceError, InternalError) as e:
if retries >= self.MAX_RETRIES:
logger.exception('Worker could not re-establish database connectivity, shutting down gracefully: Job {}'.format(job_identifier))
os.kill(os.getppid(), signal.SIGINT)
return
delay = 60 * retries
logger.exception('Database Error Saving Job Event, retry #{i} in {delay} seconds:'.format(
i=retries + 1,
delay=delay
))
django_connection.close()
time.sleep(delay)
retries += 1
except DatabaseError as e:
logger.exception('Database Error Saving Job Event for Job {}'.format(job_identifier))
break
except Exception as exc:
tb = traceback.format_exc()
logger.error('Callback Task Processor Raised Exception: %r', exc)
logger.error('Detail: {}'.format(tb))
Loading

0 comments on commit da74f1d

Please sign in to comment.