2020-04-07 14:14:01 +02:00
|
|
|
From 9b47e4497ca16ec3671dd9da838895820ea9b0af Mon Sep 17 00:00:00 2001
|
2019-04-12 11:57:21 +02:00
|
|
|
From: Mihai Dinca <mdinca@suse.de>
|
|
|
|
Date: Fri, 16 Nov 2018 17:05:29 +0100
|
|
|
|
Subject: [PATCH] Async batch implementation
|
|
|
|
|
|
|
|
Add find_job checks
|
|
|
|
|
|
|
|
Check if should close on all events
|
|
|
|
|
|
|
|
Make batch_delay a request parameter
|
|
|
|
|
|
|
|
Allow multiple event handlers
|
|
|
|
|
|
|
|
Use config value for gather_job_timeout when not in payload
|
|
|
|
|
|
|
|
Add async batch unittests
|
|
|
|
|
|
|
|
Allow metadata to pass
|
|
|
|
|
|
|
|
Pass metadata only to batch jobs
|
|
|
|
|
|
|
|
Add the metadata to the start/done events
|
|
|
|
|
|
|
|
Pass only metadata not all **kwargs
|
|
|
|
|
|
|
|
Add separate batch presence_ping timeout
|
|
|
|
---
|
|
|
|
salt/auth/__init__.py | 4 +-
|
2020-04-07 14:14:01 +02:00
|
|
|
salt/cli/batch.py | 91 ++++++++--
|
|
|
|
salt/cli/batch_async.py | 227 ++++++++++++++++++++++++
|
|
|
|
salt/client/__init__.py | 41 +----
|
|
|
|
salt/master.py | 25 +++
|
2019-04-12 11:57:21 +02:00
|
|
|
salt/netapi/__init__.py | 3 +-
|
2020-04-07 14:14:01 +02:00
|
|
|
salt/transport/ipc.py | 13 +-
|
2019-11-28 16:41:55 +01:00
|
|
|
salt/utils/event.py | 8 +-
|
2020-04-07 14:14:01 +02:00
|
|
|
tests/unit/cli/test_batch_async.py | 351 +++++++++++++++++++++++++++++++++++++
|
|
|
|
9 files changed, 703 insertions(+), 60 deletions(-)
|
2019-04-12 11:57:21 +02:00
|
|
|
create mode 100644 salt/cli/batch_async.py
|
|
|
|
create mode 100644 tests/unit/cli/test_batch_async.py
|
|
|
|
|
|
|
|
diff --git a/salt/auth/__init__.py b/salt/auth/__init__.py
|
2020-04-07 14:14:01 +02:00
|
|
|
index ecbd1c808c..ba9dbad509 100644
|
2019-04-12 11:57:21 +02:00
|
|
|
--- a/salt/auth/__init__.py
|
|
|
|
+++ b/salt/auth/__init__.py
|
2019-11-28 16:41:55 +01:00
|
|
|
@@ -52,7 +52,9 @@ AUTH_INTERNAL_KEYWORDS = frozenset([
|
2019-04-12 11:57:21 +02:00
|
|
|
'metadata',
|
|
|
|
'print_event',
|
|
|
|
'raw',
|
|
|
|
- 'yield_pub_data'
|
|
|
|
+ 'yield_pub_data',
|
|
|
|
+ 'batch',
|
|
|
|
+ 'batch_delay'
|
|
|
|
])
|
|
|
|
|
|
|
|
|
|
|
|
diff --git a/salt/cli/batch.py b/salt/cli/batch.py
|
2019-11-28 16:41:55 +01:00
|
|
|
index 723944b60a..36e66da1af 100644
|
2019-04-12 11:57:21 +02:00
|
|
|
--- a/salt/cli/batch.py
|
|
|
|
+++ b/salt/cli/batch.py
|
|
|
|
@@ -26,6 +26,79 @@ import logging
|
|
|
|
log = logging.getLogger(__name__)
|
|
|
|
|
|
|
|
|
|
|
|
+def get_bnum(opts, minions, quiet):
|
|
|
|
+ '''
|
|
|
|
+ Return the active number of minions to maintain
|
|
|
|
+ '''
|
|
|
|
+ partition = lambda x: float(x) / 100.0 * len(minions)
|
|
|
|
+ try:
|
2019-11-28 16:41:55 +01:00
|
|
|
+ if isinstance(opts['batch'], six.string_types) and '%' in opts['batch']:
|
2019-04-12 11:57:21 +02:00
|
|
|
+ res = partition(float(opts['batch'].strip('%')))
|
|
|
|
+ if res < 1:
|
|
|
|
+ return int(math.ceil(res))
|
|
|
|
+ else:
|
|
|
|
+ return int(res)
|
|
|
|
+ else:
|
|
|
|
+ return int(opts['batch'])
|
|
|
|
+ except ValueError:
|
|
|
|
+ if not quiet:
|
|
|
|
+ salt.utils.stringutils.print_cli('Invalid batch data sent: {0}\nData must be in the '
|
|
|
|
+ 'form of %10, 10% or 3'.format(opts['batch']))
|
|
|
|
+
|
|
|
|
+
|
|
|
|
+def batch_get_opts(
|
|
|
|
+ tgt,
|
|
|
|
+ fun,
|
|
|
|
+ batch,
|
|
|
|
+ parent_opts,
|
|
|
|
+ arg=(),
|
|
|
|
+ tgt_type='glob',
|
|
|
|
+ ret='',
|
|
|
|
+ kwarg=None,
|
|
|
|
+ **kwargs):
|
|
|
|
+ # We need to re-import salt.utils.args here
|
|
|
|
+ # even though it has already been imported.
|
|
|
|
+ # when cmd_batch is called via the NetAPI
|
|
|
|
+ # the module is unavailable.
|
|
|
|
+ import salt.utils.args
|
|
|
|
+
|
|
|
|
+ arg = salt.utils.args.condition_input(arg, kwarg)
|
|
|
|
+ opts = {'tgt': tgt,
|
|
|
|
+ 'fun': fun,
|
|
|
|
+ 'arg': arg,
|
|
|
|
+ 'tgt_type': tgt_type,
|
|
|
|
+ 'ret': ret,
|
|
|
|
+ 'batch': batch,
|
2019-11-28 16:41:55 +01:00
|
|
|
+ 'failhard': kwargs.get('failhard', parent_opts.get('failhard', False)),
|
2019-04-12 11:57:21 +02:00
|
|
|
+ 'raw': kwargs.get('raw', False)}
|
|
|
|
+
|
|
|
|
+ if 'timeout' in kwargs:
|
|
|
|
+ opts['timeout'] = kwargs['timeout']
|
|
|
|
+ if 'gather_job_timeout' in kwargs:
|
|
|
|
+ opts['gather_job_timeout'] = kwargs['gather_job_timeout']
|
|
|
|
+ if 'batch_wait' in kwargs:
|
|
|
|
+ opts['batch_wait'] = int(kwargs['batch_wait'])
|
|
|
|
+
|
|
|
|
+ for key, val in six.iteritems(parent_opts):
|
|
|
|
+ if key not in opts:
|
|
|
|
+ opts[key] = val
|
|
|
|
+
|
|
|
|
+ return opts
|
|
|
|
+
|
|
|
|
+
|
|
|
|
+def batch_get_eauth(kwargs):
|
|
|
|
+ eauth = {}
|
|
|
|
+ if 'eauth' in kwargs:
|
|
|
|
+ eauth['eauth'] = kwargs.pop('eauth')
|
|
|
|
+ if 'username' in kwargs:
|
|
|
|
+ eauth['username'] = kwargs.pop('username')
|
|
|
|
+ if 'password' in kwargs:
|
|
|
|
+ eauth['password'] = kwargs.pop('password')
|
|
|
|
+ if 'token' in kwargs:
|
|
|
|
+ eauth['token'] = kwargs.pop('token')
|
|
|
|
+ return eauth
|
|
|
|
+
|
|
|
|
+
|
|
|
|
class Batch(object):
|
|
|
|
'''
|
|
|
|
Manage the execution of batch runs
|
|
|
|
@@ -80,23 +153,7 @@ class Batch(object):
|
|
|
|
return (list(fret), ping_gen, nret.difference(fret))
|
|
|
|
|
|
|
|
def get_bnum(self):
|
|
|
|
- '''
|
|
|
|
- Return the active number of minions to maintain
|
|
|
|
- '''
|
|
|
|
- partition = lambda x: float(x) / 100.0 * len(self.minions)
|
|
|
|
- try:
|
2019-11-28 16:41:55 +01:00
|
|
|
- if isinstance(self.opts['batch'], six.string_types) and '%' in self.opts['batch']:
|
2019-04-12 11:57:21 +02:00
|
|
|
- res = partition(float(self.opts['batch'].strip('%')))
|
|
|
|
- if res < 1:
|
|
|
|
- return int(math.ceil(res))
|
|
|
|
- else:
|
|
|
|
- return int(res)
|
|
|
|
- else:
|
|
|
|
- return int(self.opts['batch'])
|
|
|
|
- except ValueError:
|
|
|
|
- if not self.quiet:
|
|
|
|
- salt.utils.stringutils.print_cli('Invalid batch data sent: {0}\nData must be in the '
|
|
|
|
- 'form of %10, 10% or 3'.format(self.opts['batch']))
|
|
|
|
+ return get_bnum(self.opts, self.minions, self.quiet)
|
|
|
|
|
|
|
|
def __update_wait(self, wait):
|
|
|
|
now = datetime.now()
|
|
|
|
diff --git a/salt/cli/batch_async.py b/salt/cli/batch_async.py
|
|
|
|
new file mode 100644
|
|
|
|
index 0000000000..3160d46d8b
|
|
|
|
--- /dev/null
|
|
|
|
+++ b/salt/cli/batch_async.py
|
|
|
|
@@ -0,0 +1,227 @@
|
|
|
|
+# -*- coding: utf-8 -*-
|
|
|
|
+'''
|
|
|
|
+Execute a job on the targeted minions by using a moving window of fixed size `batch`.
|
|
|
|
+'''
|
|
|
|
+
|
|
|
|
+# Import python libs
|
|
|
|
+from __future__ import absolute_import, print_function, unicode_literals
|
|
|
|
+import tornado
|
|
|
|
+
|
|
|
|
+# Import salt libs
|
|
|
|
+import salt.client
|
|
|
|
+
|
|
|
|
+# pylint: enable=import-error,no-name-in-module,redefined-builtin
|
|
|
|
+import logging
|
|
|
|
+import fnmatch
|
|
|
|
+
|
|
|
|
+log = logging.getLogger(__name__)
|
|
|
|
+
|
|
|
|
+from salt.cli.batch import get_bnum, batch_get_opts, batch_get_eauth
|
|
|
|
+
|
|
|
|
+
|
|
|
|
+class BatchAsync(object):
|
|
|
|
+ '''
|
|
|
|
+ Run a job on the targeted minions by using a moving window of fixed size `batch`.
|
|
|
|
+
|
|
|
|
+ ``BatchAsync`` is used to execute a job on the targeted minions by keeping
|
|
|
|
+ the number of concurrent running minions to the size of `batch` parameter.
|
|
|
|
+
|
|
|
|
+ The control parameters are:
|
|
|
|
+ - batch: number/percentage of concurrent running minions
|
|
|
|
+ - batch_delay: minimum wait time between batches
|
|
|
|
+ - batch_presence_ping_timeout: time to wait for presence pings before starting the batch
|
|
|
|
+ - gather_job_timeout: `find_job` timeout
|
|
|
|
+ - timeout: time to wait before firing a `find_job`
|
|
|
|
+
|
|
|
|
+ When the batch stars, a `start` event is fired:
|
|
|
|
+ - tag: salt/batch/<batch-jid>/start
|
|
|
|
+ - data: {
|
|
|
|
+ "available_minions": self.minions,
|
|
|
|
+ "down_minions": self.down_minions
|
|
|
|
+ }
|
|
|
|
+
|
|
|
|
+ When the batch ends, an `done` event is fired:
|
|
|
|
+ - tag: salt/batch/<batch-jid>/done
|
|
|
|
+ - data: {
|
|
|
|
+ "available_minions": self.minions,
|
|
|
|
+ "down_minions": self.down_minions,
|
|
|
|
+ "done_minions": self.done_minions,
|
|
|
|
+ "timedout_minions": self.timedout_minions
|
|
|
|
+ }
|
|
|
|
+ '''
|
|
|
|
+ def __init__(self, parent_opts, jid_gen, clear_load):
|
|
|
|
+ ioloop = tornado.ioloop.IOLoop.current()
|
|
|
|
+ self.local = salt.client.get_local_client(parent_opts['conf_file'])
|
|
|
|
+ if 'gather_job_timeout' in clear_load['kwargs']:
|
|
|
|
+ clear_load['gather_job_timeout'] = clear_load['kwargs'].pop('gather_job_timeout')
|
|
|
|
+ else:
|
|
|
|
+ clear_load['gather_job_timeout'] = self.local.opts['gather_job_timeout']
|
|
|
|
+ self.batch_presence_ping_timeout = clear_load['kwargs'].get('batch_presence_ping_timeout', None)
|
|
|
|
+ self.batch_delay = clear_load['kwargs'].get('batch_delay', 1)
|
|
|
|
+ self.opts = batch_get_opts(
|
|
|
|
+ clear_load.pop('tgt'),
|
|
|
|
+ clear_load.pop('fun'),
|
|
|
|
+ clear_load['kwargs'].pop('batch'),
|
|
|
|
+ self.local.opts,
|
|
|
|
+ **clear_load)
|
|
|
|
+ self.eauth = batch_get_eauth(clear_load['kwargs'])
|
|
|
|
+ self.metadata = clear_load['kwargs'].get('metadata', {})
|
|
|
|
+ self.minions = set()
|
|
|
|
+ self.down_minions = set()
|
|
|
|
+ self.timedout_minions = set()
|
|
|
|
+ self.done_minions = set()
|
|
|
|
+ self.active = set()
|
|
|
|
+ self.initialized = False
|
|
|
|
+ self.ping_jid = jid_gen()
|
|
|
|
+ self.batch_jid = jid_gen()
|
|
|
|
+ self.find_job_jid = jid_gen()
|
|
|
|
+ self.find_job_returned = set()
|
|
|
|
+ self.event = salt.utils.event.get_event(
|
|
|
|
+ 'master',
|
|
|
|
+ self.opts['sock_dir'],
|
|
|
|
+ self.opts['transport'],
|
|
|
|
+ opts=self.opts,
|
|
|
|
+ listen=True,
|
|
|
|
+ io_loop=ioloop,
|
|
|
|
+ keep_loop=True)
|
|
|
|
+
|
|
|
|
+ def __set_event_handler(self):
|
|
|
|
+ ping_return_pattern = 'salt/job/{0}/ret/*'.format(self.ping_jid)
|
|
|
|
+ batch_return_pattern = 'salt/job/{0}/ret/*'.format(self.batch_jid)
|
|
|
|
+ find_job_return_pattern = 'salt/job/{0}/ret/*'.format(self.find_job_jid)
|
|
|
|
+ self.event.subscribe(ping_return_pattern, match_type='glob')
|
|
|
|
+ self.event.subscribe(batch_return_pattern, match_type='glob')
|
|
|
|
+ self.event.subscribe(find_job_return_pattern, match_type='glob')
|
|
|
|
+ self.event.patterns = {
|
|
|
|
+ (ping_return_pattern, 'ping_return'),
|
|
|
|
+ (batch_return_pattern, 'batch_run'),
|
|
|
|
+ (find_job_return_pattern, 'find_job_return')
|
|
|
|
+ }
|
|
|
|
+ self.event.set_event_handler(self.__event_handler)
|
|
|
|
+
|
|
|
|
+ def __event_handler(self, raw):
|
|
|
|
+ if not self.event:
|
|
|
|
+ return
|
|
|
|
+ mtag, data = self.event.unpack(raw, self.event.serial)
|
|
|
|
+ for (pattern, op) in self.event.patterns:
|
|
|
|
+ if fnmatch.fnmatch(mtag, pattern):
|
|
|
|
+ minion = data['id']
|
|
|
|
+ if op == 'ping_return':
|
|
|
|
+ self.minions.add(minion)
|
|
|
|
+ self.down_minions.remove(minion)
|
|
|
|
+ if not self.down_minions:
|
|
|
|
+ self.event.io_loop.spawn_callback(self.start_batch)
|
|
|
|
+ elif op == 'find_job_return':
|
|
|
|
+ self.find_job_returned.add(minion)
|
|
|
|
+ elif op == 'batch_run':
|
|
|
|
+ if minion in self.active:
|
|
|
|
+ self.active.remove(minion)
|
|
|
|
+ self.done_minions.add(minion)
|
|
|
|
+ # call later so that we maybe gather more returns
|
|
|
|
+ self.event.io_loop.call_later(self.batch_delay, self.schedule_next)
|
|
|
|
+
|
|
|
|
+ if self.initialized and self.done_minions == self.minions.difference(self.timedout_minions):
|
|
|
|
+ self.end_batch()
|
|
|
|
+
|
|
|
|
+ def _get_next(self):
|
|
|
|
+ to_run = self.minions.difference(
|
|
|
|
+ self.done_minions).difference(
|
|
|
|
+ self.active).difference(
|
|
|
|
+ self.timedout_minions)
|
|
|
|
+ next_batch_size = min(
|
|
|
|
+ len(to_run), # partial batch (all left)
|
|
|
|
+ self.batch_size - len(self.active) # full batch or available slots
|
|
|
|
+ )
|
|
|
|
+ return set(list(to_run)[:next_batch_size])
|
|
|
|
+
|
|
|
|
+ @tornado.gen.coroutine
|
|
|
|
+ def check_find_job(self, minions):
|
|
|
|
+ did_not_return = minions.difference(self.find_job_returned)
|
|
|
|
+ if did_not_return:
|
|
|
|
+ for minion in did_not_return:
|
|
|
|
+ if minion in self.find_job_returned:
|
|
|
|
+ self.find_job_returned.remove(minion)
|
|
|
|
+ if minion in self.active:
|
|
|
|
+ self.active.remove(minion)
|
|
|
|
+ self.timedout_minions.add(minion)
|
|
|
|
+ running = minions.difference(did_not_return).difference(self.done_minions).difference(self.timedout_minions)
|
|
|
|
+ if running:
|
|
|
|
+ self.event.io_loop.add_callback(self.find_job, running)
|
|
|
|
+
|
|
|
|
+ @tornado.gen.coroutine
|
|
|
|
+ def find_job(self, minions):
|
|
|
|
+ not_done = minions.difference(self.done_minions)
|
|
|
|
+ ping_return = yield self.local.run_job_async(
|
|
|
|
+ not_done,
|
|
|
|
+ 'saltutil.find_job',
|
|
|
|
+ [self.batch_jid],
|
|
|
|
+ 'list',
|
|
|
|
+ gather_job_timeout=self.opts['gather_job_timeout'],
|
|
|
|
+ jid=self.find_job_jid,
|
|
|
|
+ **self.eauth)
|
|
|
|
+ self.event.io_loop.call_later(
|
|
|
|
+ self.opts['gather_job_timeout'],
|
|
|
|
+ self.check_find_job,
|
|
|
|
+ not_done)
|
|
|
|
+
|
|
|
|
+ @tornado.gen.coroutine
|
|
|
|
+ def start(self):
|
|
|
|
+ self.__set_event_handler()
|
|
|
|
+ #start batching even if not all minions respond to ping
|
|
|
|
+ self.event.io_loop.call_later(
|
|
|
|
+ self.batch_presence_ping_timeout or self.opts['gather_job_timeout'],
|
|
|
|
+ self.start_batch)
|
|
|
|
+ ping_return = yield self.local.run_job_async(
|
|
|
|
+ self.opts['tgt'],
|
|
|
|
+ 'test.ping',
|
|
|
|
+ [],
|
|
|
|
+ self.opts.get(
|
|
|
|
+ 'selected_target_option',
|
|
|
|
+ self.opts.get('tgt_type', 'glob')
|
|
|
|
+ ),
|
|
|
|
+ gather_job_timeout=self.opts['gather_job_timeout'],
|
|
|
|
+ jid=self.ping_jid,
|
|
|
|
+ metadata=self.metadata,
|
|
|
|
+ **self.eauth)
|
|
|
|
+ self.down_minions = set(ping_return['minions'])
|
|
|
|
+
|
|
|
|
+ @tornado.gen.coroutine
|
|
|
|
+ def start_batch(self):
|
|
|
|
+ if not self.initialized:
|
|
|
|
+ self.batch_size = get_bnum(self.opts, self.minions, True)
|
|
|
|
+ self.initialized = True
|
|
|
|
+ data = {
|
|
|
|
+ "available_minions": self.minions,
|
|
|
|
+ "down_minions": self.down_minions,
|
|
|
|
+ "metadata": self.metadata
|
|
|
|
+ }
|
|
|
|
+ self.event.fire_event(data, "salt/batch/{0}/start".format(self.batch_jid))
|
|
|
|
+ yield self.schedule_next()
|
|
|
|
+
|
|
|
|
+ def end_batch(self):
|
|
|
|
+ data = {
|
|
|
|
+ "available_minions": self.minions,
|
|
|
|
+ "down_minions": self.down_minions,
|
|
|
|
+ "done_minions": self.done_minions,
|
|
|
|
+ "timedout_minions": self.timedout_minions,
|
|
|
|
+ "metadata": self.metadata
|
|
|
|
+ }
|
|
|
|
+ self.event.fire_event(data, "salt/batch/{0}/done".format(self.batch_jid))
|
|
|
|
+ self.event.remove_event_handler(self.__event_handler)
|
|
|
|
+
|
|
|
|
+ @tornado.gen.coroutine
|
|
|
|
+ def schedule_next(self):
|
|
|
|
+ next_batch = self._get_next()
|
|
|
|
+ if next_batch:
|
|
|
|
+ yield self.local.run_job_async(
|
|
|
|
+ next_batch,
|
|
|
|
+ self.opts['fun'],
|
|
|
|
+ self.opts['arg'],
|
|
|
|
+ 'list',
|
|
|
|
+ raw=self.opts.get('raw', False),
|
|
|
|
+ ret=self.opts.get('return', ''),
|
|
|
|
+ gather_job_timeout=self.opts['gather_job_timeout'],
|
|
|
|
+ jid=self.batch_jid,
|
|
|
|
+ metadata=self.metadata)
|
|
|
|
+ self.event.io_loop.call_later(self.opts['timeout'], self.find_job, set(next_batch))
|
|
|
|
+ self.active = self.active.union(next_batch)
|
|
|
|
diff --git a/salt/client/__init__.py b/salt/client/__init__.py
|
2020-04-07 14:14:01 +02:00
|
|
|
index 08fbe191d4..3bbc7f9de7 100644
|
2019-04-12 11:57:21 +02:00
|
|
|
--- a/salt/client/__init__.py
|
|
|
|
+++ b/salt/client/__init__.py
|
2020-04-07 14:14:01 +02:00
|
|
|
@@ -529,45 +529,14 @@ class LocalClient(object):
|
2019-04-12 11:57:21 +02:00
|
|
|
{'dave': {...}}
|
|
|
|
{'stewart': {...}}
|
|
|
|
'''
|
|
|
|
- # We need to re-import salt.utils.args here
|
|
|
|
- # even though it has already been imported.
|
|
|
|
- # when cmd_batch is called via the NetAPI
|
|
|
|
- # the module is unavailable.
|
|
|
|
- import salt.utils.args
|
|
|
|
-
|
|
|
|
# Late import - not used anywhere else in this file
|
|
|
|
import salt.cli.batch
|
|
|
|
+ opts = salt.cli.batch.batch_get_opts(
|
|
|
|
+ tgt, fun, batch, self.opts,
|
|
|
|
+ arg=arg, tgt_type=tgt_type, ret=ret, kwarg=kwarg, **kwargs)
|
|
|
|
+
|
|
|
|
+ eauth = salt.cli.batch.batch_get_eauth(kwargs)
|
|
|
|
|
|
|
|
- arg = salt.utils.args.condition_input(arg, kwarg)
|
|
|
|
- opts = {'tgt': tgt,
|
|
|
|
- 'fun': fun,
|
|
|
|
- 'arg': arg,
|
|
|
|
- 'tgt_type': tgt_type,
|
|
|
|
- 'ret': ret,
|
|
|
|
- 'batch': batch,
|
2019-11-28 16:41:55 +01:00
|
|
|
- 'failhard': kwargs.get('failhard', self.opts.get('failhard', False)),
|
2019-04-12 11:57:21 +02:00
|
|
|
- 'raw': kwargs.get('raw', False)}
|
|
|
|
-
|
|
|
|
- if 'timeout' in kwargs:
|
|
|
|
- opts['timeout'] = kwargs['timeout']
|
|
|
|
- if 'gather_job_timeout' in kwargs:
|
|
|
|
- opts['gather_job_timeout'] = kwargs['gather_job_timeout']
|
|
|
|
- if 'batch_wait' in kwargs:
|
|
|
|
- opts['batch_wait'] = int(kwargs['batch_wait'])
|
|
|
|
-
|
|
|
|
- eauth = {}
|
|
|
|
- if 'eauth' in kwargs:
|
|
|
|
- eauth['eauth'] = kwargs.pop('eauth')
|
|
|
|
- if 'username' in kwargs:
|
|
|
|
- eauth['username'] = kwargs.pop('username')
|
|
|
|
- if 'password' in kwargs:
|
|
|
|
- eauth['password'] = kwargs.pop('password')
|
|
|
|
- if 'token' in kwargs:
|
|
|
|
- eauth['token'] = kwargs.pop('token')
|
|
|
|
-
|
|
|
|
- for key, val in six.iteritems(self.opts):
|
|
|
|
- if key not in opts:
|
|
|
|
- opts[key] = val
|
|
|
|
batch = salt.cli.batch.Batch(opts, eauth=eauth, quiet=True)
|
|
|
|
for ret in batch.run():
|
|
|
|
yield ret
|
|
|
|
diff --git a/salt/master.py b/salt/master.py
|
2020-04-07 14:14:01 +02:00
|
|
|
index fb2e0c35bf..5e2277ba76 100644
|
2019-04-12 11:57:21 +02:00
|
|
|
--- a/salt/master.py
|
|
|
|
+++ b/salt/master.py
|
2020-04-07 14:14:01 +02:00
|
|
|
@@ -31,6 +31,7 @@ import salt.ext.tornado.gen # pylint: disable=F0401
|
2019-04-12 11:57:21 +02:00
|
|
|
|
|
|
|
# Import salt libs
|
|
|
|
import salt.crypt
|
|
|
|
+import salt.cli.batch_async
|
|
|
|
import salt.client
|
|
|
|
import salt.client.ssh.client
|
|
|
|
import salt.exceptions
|
2019-11-28 16:41:55 +01:00
|
|
|
@@ -2040,6 +2041,27 @@ class ClearFuncs(object):
|
2019-04-12 11:57:21 +02:00
|
|
|
return False
|
|
|
|
return self.loadauth.get_tok(clear_load['token'])
|
|
|
|
|
|
|
|
+ def publish_batch(self, clear_load, minions, missing):
|
|
|
|
+ batch_load = {}
|
|
|
|
+ batch_load.update(clear_load)
|
|
|
|
+ import salt.cli.batch_async
|
|
|
|
+ batch = salt.cli.batch_async.BatchAsync(
|
|
|
|
+ self.local.opts,
|
|
|
|
+ functools.partial(self._prep_jid, clear_load, {}),
|
|
|
|
+ batch_load
|
|
|
|
+ )
|
|
|
|
+ ioloop = tornado.ioloop.IOLoop.current()
|
|
|
|
+ ioloop.add_callback(batch.start)
|
|
|
|
+
|
|
|
|
+ return {
|
|
|
|
+ 'enc': 'clear',
|
|
|
|
+ 'load': {
|
|
|
|
+ 'jid': batch.batch_jid,
|
|
|
|
+ 'minions': minions,
|
|
|
|
+ 'missing': missing
|
|
|
|
+ }
|
|
|
|
+ }
|
|
|
|
+
|
|
|
|
def publish(self, clear_load):
|
|
|
|
'''
|
|
|
|
This method sends out publications to the minions, it can only be used
|
2019-11-28 16:41:55 +01:00
|
|
|
@@ -2131,6 +2153,9 @@ class ClearFuncs(object):
|
2019-04-12 11:57:21 +02:00
|
|
|
'error': 'Master could not resolve minions for target {0}'.format(clear_load['tgt'])
|
|
|
|
}
|
|
|
|
}
|
|
|
|
+ if extra.get('batch', None):
|
|
|
|
+ return self.publish_batch(clear_load, minions, missing)
|
|
|
|
+
|
|
|
|
jid = self._prep_jid(clear_load, extra)
|
|
|
|
if jid is None:
|
|
|
|
return {'enc': 'clear',
|
|
|
|
diff --git a/salt/netapi/__init__.py b/salt/netapi/__init__.py
|
2020-04-07 14:14:01 +02:00
|
|
|
index 88d550f27f..31a24bb420 100644
|
2019-04-12 11:57:21 +02:00
|
|
|
--- a/salt/netapi/__init__.py
|
|
|
|
+++ b/salt/netapi/__init__.py
|
2020-04-07 14:14:01 +02:00
|
|
|
@@ -93,7 +93,8 @@ class NetapiClient(object):
|
2019-04-12 11:57:21 +02:00
|
|
|
:return: job ID
|
|
|
|
'''
|
|
|
|
local = salt.client.get_local_client(mopts=self.opts)
|
|
|
|
- return local.run_job(*args, **kwargs)
|
|
|
|
+ ret = local.run_job(*args, **kwargs)
|
|
|
|
+ return ret
|
|
|
|
|
|
|
|
def local(self, *args, **kwargs):
|
|
|
|
'''
|
|
|
|
diff --git a/salt/transport/ipc.py b/salt/transport/ipc.py
|
2020-04-07 14:14:01 +02:00
|
|
|
index 89fb31fb4f..d2b295a633 100644
|
2019-04-12 11:57:21 +02:00
|
|
|
--- a/salt/transport/ipc.py
|
|
|
|
+++ b/salt/transport/ipc.py
|
2020-04-07 14:14:01 +02:00
|
|
|
@@ -618,6 +618,7 @@ class IPCMessageSubscriber(IPCClient):
|
2019-11-28 16:41:55 +01:00
|
|
|
self._read_stream_future = None
|
|
|
|
self._saved_data = []
|
|
|
|
self._read_in_progress = Lock()
|
2019-04-12 11:57:21 +02:00
|
|
|
+ self.callbacks = set()
|
|
|
|
|
2020-04-07 14:14:01 +02:00
|
|
|
@salt.ext.tornado.gen.coroutine
|
2019-11-28 16:41:55 +01:00
|
|
|
def _read(self, timeout, callback=None):
|
2020-04-07 14:14:01 +02:00
|
|
|
@@ -692,8 +693,12 @@ class IPCMessageSubscriber(IPCClient):
|
2019-11-28 16:41:55 +01:00
|
|
|
return self._saved_data.pop(0)
|
|
|
|
return self.io_loop.run_sync(lambda: self._read(timeout))
|
2019-04-12 11:57:21 +02:00
|
|
|
|
2020-04-07 14:14:01 +02:00
|
|
|
- @salt.ext.tornado.gen.coroutine
|
|
|
|
- def read_async(self, callback):
|
2019-04-12 11:57:21 +02:00
|
|
|
+ def __run_callbacks(self, raw):
|
|
|
|
+ for callback in self.callbacks:
|
|
|
|
+ self.io_loop.spawn_callback(callback, raw)
|
|
|
|
+
|
2020-04-07 14:14:01 +02:00
|
|
|
+ @tornado.gen.coroutine
|
2019-04-12 11:57:21 +02:00
|
|
|
+ def read_async(self):
|
|
|
|
'''
|
|
|
|
Asynchronously read messages and invoke a callback when they are ready.
|
|
|
|
|
2020-04-07 14:14:01 +02:00
|
|
|
@@ -707,8 +712,8 @@ class IPCMessageSubscriber(IPCClient):
|
|
|
|
yield salt.ext.tornado.gen.sleep(1)
|
|
|
|
except Exception as exc: # pylint: disable=broad-except
|
2019-04-12 11:57:21 +02:00
|
|
|
log.error('Exception occurred while Subscriber connecting: %s', exc)
|
2020-04-07 14:14:01 +02:00
|
|
|
- yield salt.ext.tornado.gen.sleep(1)
|
2019-11-28 16:41:55 +01:00
|
|
|
- yield self._read(None, callback)
|
2020-04-07 14:14:01 +02:00
|
|
|
+ yield tornado.gen.sleep(1)
|
2019-11-28 16:41:55 +01:00
|
|
|
+ yield self._read(None, self.__run_callbacks)
|
2019-04-12 11:57:21 +02:00
|
|
|
|
|
|
|
def close(self):
|
|
|
|
'''
|
|
|
|
diff --git a/salt/utils/event.py b/salt/utils/event.py
|
2020-04-07 14:14:01 +02:00
|
|
|
index 0a3074a6f7..19e5db070d 100644
|
2019-04-12 11:57:21 +02:00
|
|
|
--- a/salt/utils/event.py
|
|
|
|
+++ b/salt/utils/event.py
|
2020-04-07 14:14:01 +02:00
|
|
|
@@ -857,6 +857,10 @@ class SaltEvent(object):
|
2019-04-12 11:57:21 +02:00
|
|
|
# Minion fired a bad retcode, fire an event
|
|
|
|
self._fire_ret_load_specific_fun(load)
|
|
|
|
|
|
|
|
+ def remove_event_handler(self, event_handler):
|
|
|
|
+ if event_handler in self.subscriber.callbacks:
|
|
|
|
+ self.subscriber.callbacks.remove(event_handler)
|
|
|
|
+
|
|
|
|
def set_event_handler(self, event_handler):
|
|
|
|
'''
|
|
|
|
Invoke the event_handler callback each time an event arrives.
|
2020-04-07 14:14:01 +02:00
|
|
|
@@ -865,8 +869,10 @@ class SaltEvent(object):
|
2019-04-12 11:57:21 +02:00
|
|
|
|
|
|
|
if not self.cpub:
|
|
|
|
self.connect_pub()
|
|
|
|
+
|
|
|
|
+ self.subscriber.callbacks.add(event_handler)
|
2019-11-28 16:41:55 +01:00
|
|
|
# This will handle reconnects
|
|
|
|
- return self.subscriber.read_async(event_handler)
|
|
|
|
+ return self.subscriber.read_async()
|
2019-04-12 11:57:21 +02:00
|
|
|
|
2020-04-07 14:14:01 +02:00
|
|
|
# pylint: disable=W1701
|
2019-04-12 11:57:21 +02:00
|
|
|
def __del__(self):
|
|
|
|
diff --git a/tests/unit/cli/test_batch_async.py b/tests/unit/cli/test_batch_async.py
|
|
|
|
new file mode 100644
|
|
|
|
index 0000000000..f65b6a06c3
|
|
|
|
--- /dev/null
|
|
|
|
+++ b/tests/unit/cli/test_batch_async.py
|
|
|
|
@@ -0,0 +1,351 @@
|
|
|
|
+# -*- coding: utf-8 -*-
|
|
|
|
+
|
|
|
|
+from __future__ import absolute_import
|
|
|
|
+
|
|
|
|
+# Import Salt Libs
|
|
|
|
+from salt.cli.batch_async import BatchAsync
|
|
|
|
+
|
|
|
|
+import tornado
|
|
|
|
+from tornado.testing import AsyncTestCase
|
|
|
|
+from tests.support.unit import skipIf, TestCase
|
|
|
|
+from tests.support.mock import MagicMock, patch, NO_MOCK, NO_MOCK_REASON
|
|
|
|
+
|
|
|
|
+
|
|
|
|
+@skipIf(NO_MOCK, NO_MOCK_REASON)
|
|
|
|
+class AsyncBatchTestCase(AsyncTestCase, TestCase):
|
|
|
|
+
|
|
|
|
+ def setUp(self):
|
|
|
|
+ self.io_loop = self.get_new_ioloop()
|
|
|
|
+ opts = {'batch': '1',
|
|
|
|
+ 'conf_file': {},
|
|
|
|
+ 'tgt': '*',
|
|
|
|
+ 'timeout': 5,
|
|
|
|
+ 'gather_job_timeout': 5,
|
|
|
|
+ 'batch_presence_ping_timeout': 1,
|
|
|
|
+ 'transport': None,
|
|
|
|
+ 'sock_dir': ''}
|
|
|
|
+
|
|
|
|
+ with patch('salt.client.get_local_client', MagicMock(return_value=MagicMock())):
|
|
|
|
+ with patch('salt.cli.batch_async.batch_get_opts',
|
|
|
|
+ MagicMock(return_value=opts)
|
|
|
|
+ ):
|
|
|
|
+ self.batch = BatchAsync(
|
|
|
|
+ opts,
|
|
|
|
+ MagicMock(side_effect=['1234', '1235', '1236']),
|
|
|
|
+ {
|
|
|
|
+ 'tgt': '',
|
|
|
|
+ 'fun': '',
|
|
|
|
+ 'kwargs': {
|
|
|
|
+ 'batch': '',
|
|
|
|
+ 'batch_presence_ping_timeout': 1
|
|
|
|
+ }
|
|
|
|
+ })
|
|
|
|
+
|
|
|
|
+ def test_ping_jid(self):
|
|
|
|
+ self.assertEqual(self.batch.ping_jid, '1234')
|
|
|
|
+
|
|
|
|
+ def test_batch_jid(self):
|
|
|
|
+ self.assertEqual(self.batch.batch_jid, '1235')
|
|
|
|
+
|
|
|
|
+ def test_find_job_jid(self):
|
|
|
|
+ self.assertEqual(self.batch.find_job_jid, '1236')
|
|
|
|
+
|
|
|
|
+ def test_batch_size(self):
|
|
|
|
+ '''
|
|
|
|
+ Tests passing batch value as a number
|
|
|
|
+ '''
|
|
|
|
+ self.batch.opts = {'batch': '2', 'timeout': 5}
|
|
|
|
+ self.batch.minions = set(['foo', 'bar'])
|
|
|
|
+ self.batch.start_batch()
|
|
|
|
+ self.assertEqual(self.batch.batch_size, 2)
|
|
|
|
+
|
|
|
|
+ @tornado.testing.gen_test
|
|
|
|
+ def test_batch_start_on_batch_presence_ping_timeout(self):
|
|
|
|
+ self.batch.event = MagicMock()
|
|
|
|
+ future = tornado.gen.Future()
|
|
|
|
+ future.set_result({'minions': ['foo', 'bar']})
|
|
|
|
+ self.batch.local.run_job_async.return_value = future
|
|
|
|
+ ret = self.batch.start()
|
|
|
|
+ # assert start_batch is called later with batch_presence_ping_timeout as param
|
|
|
|
+ self.assertEqual(
|
|
|
|
+ self.batch.event.io_loop.call_later.call_args[0],
|
|
|
|
+ (self.batch.batch_presence_ping_timeout, self.batch.start_batch))
|
|
|
|
+ # assert test.ping called
|
|
|
|
+ self.assertEqual(
|
|
|
|
+ self.batch.local.run_job_async.call_args[0],
|
|
|
|
+ ('*', 'test.ping', [], 'glob')
|
|
|
|
+ )
|
|
|
|
+ # assert down_minions == all minions matched by tgt
|
|
|
|
+ self.assertEqual(self.batch.down_minions, set(['foo', 'bar']))
|
|
|
|
+
|
|
|
|
+ @tornado.testing.gen_test
|
|
|
|
+ def test_batch_start_on_gather_job_timeout(self):
|
|
|
|
+ self.batch.event = MagicMock()
|
|
|
|
+ future = tornado.gen.Future()
|
|
|
|
+ future.set_result({'minions': ['foo', 'bar']})
|
|
|
|
+ self.batch.local.run_job_async.return_value = future
|
|
|
|
+ self.batch.batch_presence_ping_timeout = None
|
|
|
|
+ ret = self.batch.start()
|
|
|
|
+ # assert start_batch is called later with gather_job_timeout as param
|
|
|
|
+ self.assertEqual(
|
|
|
|
+ self.batch.event.io_loop.call_later.call_args[0],
|
|
|
|
+ (self.batch.opts['gather_job_timeout'], self.batch.start_batch))
|
|
|
|
+
|
|
|
|
+ def test_batch_fire_start_event(self):
|
|
|
|
+ self.batch.minions = set(['foo', 'bar'])
|
|
|
|
+ self.batch.opts = {'batch': '2', 'timeout': 5}
|
|
|
|
+ self.batch.event = MagicMock()
|
|
|
|
+ self.batch.metadata = {'mykey': 'myvalue'}
|
|
|
|
+ self.batch.start_batch()
|
|
|
|
+ self.assertEqual(
|
|
|
|
+ self.batch.event.fire_event.call_args[0],
|
|
|
|
+ (
|
|
|
|
+ {
|
|
|
|
+ 'available_minions': set(['foo', 'bar']),
|
|
|
|
+ 'down_minions': set(),
|
|
|
|
+ 'metadata': self.batch.metadata
|
|
|
|
+ },
|
|
|
|
+ "salt/batch/1235/start"
|
|
|
|
+ )
|
|
|
|
+ )
|
|
|
|
+
|
|
|
|
+ @tornado.testing.gen_test
|
|
|
|
+ def test_start_batch_calls_next(self):
|
|
|
|
+ self.batch.schedule_next = MagicMock(return_value=MagicMock())
|
|
|
|
+ self.batch.event = MagicMock()
|
|
|
|
+ future = tornado.gen.Future()
|
|
|
|
+ future.set_result(None)
|
|
|
|
+ self.batch.schedule_next = MagicMock(return_value=future)
|
|
|
|
+ self.batch.start_batch()
|
|
|
|
+ self.assertEqual(self.batch.initialized, True)
|
|
|
|
+ self.assertEqual(len(self.batch.schedule_next.mock_calls), 1)
|
|
|
|
+
|
|
|
|
+ def test_batch_fire_done_event(self):
|
|
|
|
+ self.batch.minions = set(['foo', 'bar'])
|
|
|
|
+ self.batch.event = MagicMock()
|
|
|
|
+ self.batch.metadata = {'mykey': 'myvalue'}
|
|
|
|
+ self.batch.end_batch()
|
|
|
|
+ self.assertEqual(
|
|
|
|
+ self.batch.event.fire_event.call_args[0],
|
|
|
|
+ (
|
|
|
|
+ {
|
|
|
|
+ 'available_minions': set(['foo', 'bar']),
|
|
|
|
+ 'done_minions': set(),
|
|
|
|
+ 'down_minions': set(),
|
|
|
|
+ 'timedout_minions': set(),
|
|
|
|
+ 'metadata': self.batch.metadata
|
|
|
|
+ },
|
|
|
|
+ "salt/batch/1235/done"
|
|
|
|
+ )
|
|
|
|
+ )
|
|
|
|
+ self.assertEqual(
|
|
|
|
+ len(self.batch.event.remove_event_handler.mock_calls), 1)
|
|
|
|
+
|
|
|
|
+ @tornado.testing.gen_test
|
|
|
|
+ def test_batch_next(self):
|
|
|
|
+ self.batch.event = MagicMock()
|
|
|
|
+ self.batch.opts['fun'] = 'my.fun'
|
|
|
|
+ self.batch.opts['arg'] = []
|
|
|
|
+ self.batch._get_next = MagicMock(return_value={'foo', 'bar'})
|
|
|
|
+ self.batch.batch_size = 2
|
|
|
|
+ future = tornado.gen.Future()
|
|
|
|
+ future.set_result({'minions': ['foo', 'bar']})
|
|
|
|
+ self.batch.local.run_job_async.return_value = future
|
|
|
|
+ ret = self.batch.schedule_next().result()
|
|
|
|
+ self.assertEqual(
|
|
|
|
+ self.batch.local.run_job_async.call_args[0],
|
|
|
|
+ ({'foo', 'bar'}, 'my.fun', [], 'list')
|
|
|
|
+ )
|
|
|
|
+ self.assertEqual(
|
|
|
|
+ self.batch.event.io_loop.call_later.call_args[0],
|
|
|
|
+ (self.batch.opts['timeout'], self.batch.find_job, {'foo', 'bar'})
|
|
|
|
+ )
|
|
|
|
+ self.assertEqual(self.batch.active, {'bar', 'foo'})
|
|
|
|
+
|
|
|
|
+ def test_next_batch(self):
|
|
|
|
+ self.batch.minions = {'foo', 'bar'}
|
|
|
|
+ self.batch.batch_size = 2
|
|
|
|
+ self.assertEqual(self.batch._get_next(), {'foo', 'bar'})
|
|
|
|
+
|
|
|
|
+ def test_next_batch_one_done(self):
|
|
|
|
+ self.batch.minions = {'foo', 'bar'}
|
|
|
|
+ self.batch.done_minions = {'bar'}
|
|
|
|
+ self.batch.batch_size = 2
|
|
|
|
+ self.assertEqual(self.batch._get_next(), {'foo'})
|
|
|
|
+
|
|
|
|
+ def test_next_batch_one_done_one_active(self):
|
|
|
|
+ self.batch.minions = {'foo', 'bar', 'baz'}
|
|
|
|
+ self.batch.done_minions = {'bar'}
|
|
|
|
+ self.batch.active = {'baz'}
|
|
|
|
+ self.batch.batch_size = 2
|
|
|
|
+ self.assertEqual(self.batch._get_next(), {'foo'})
|
|
|
|
+
|
|
|
|
+ def test_next_batch_one_done_one_active_one_timedout(self):
|
|
|
|
+ self.batch.minions = {'foo', 'bar', 'baz', 'faz'}
|
|
|
|
+ self.batch.done_minions = {'bar'}
|
|
|
|
+ self.batch.active = {'baz'}
|
|
|
|
+ self.batch.timedout_minions = {'faz'}
|
|
|
|
+ self.batch.batch_size = 2
|
|
|
|
+ self.assertEqual(self.batch._get_next(), {'foo'})
|
|
|
|
+
|
|
|
|
+ def test_next_batch_bigger_size(self):
|
|
|
|
+ self.batch.minions = {'foo', 'bar'}
|
|
|
|
+ self.batch.batch_size = 3
|
|
|
|
+ self.assertEqual(self.batch._get_next(), {'foo', 'bar'})
|
|
|
|
+
|
|
|
|
+ def test_next_batch_all_done(self):
|
|
|
|
+ self.batch.minions = {'foo', 'bar'}
|
|
|
|
+ self.batch.done_minions = {'foo', 'bar'}
|
|
|
|
+ self.batch.batch_size = 2
|
|
|
|
+ self.assertEqual(self.batch._get_next(), set())
|
|
|
|
+
|
|
|
|
+ def test_next_batch_all_active(self):
|
|
|
|
+ self.batch.minions = {'foo', 'bar'}
|
|
|
|
+ self.batch.active = {'foo', 'bar'}
|
|
|
|
+ self.batch.batch_size = 2
|
|
|
|
+ self.assertEqual(self.batch._get_next(), set())
|
|
|
|
+
|
|
|
|
+ def test_next_batch_all_timedout(self):
|
|
|
|
+ self.batch.minions = {'foo', 'bar'}
|
|
|
|
+ self.batch.timedout_minions = {'foo', 'bar'}
|
|
|
|
+ self.batch.batch_size = 2
|
|
|
|
+ self.assertEqual(self.batch._get_next(), set())
|
|
|
|
+
|
|
|
|
+ def test_batch__event_handler_ping_return(self):
|
|
|
|
+ self.batch.down_minions = {'foo'}
|
|
|
|
+ self.batch.event = MagicMock(
|
|
|
|
+ unpack=MagicMock(return_value=('salt/job/1234/ret/foo', {'id': 'foo'})))
|
|
|
|
+ self.batch.start()
|
|
|
|
+ self.assertEqual(self.batch.minions, set())
|
|
|
|
+ self.batch._BatchAsync__event_handler(MagicMock())
|
|
|
|
+ self.assertEqual(self.batch.minions, {'foo'})
|
|
|
|
+ self.assertEqual(self.batch.done_minions, set())
|
|
|
|
+
|
|
|
|
+ def test_batch__event_handler_call_start_batch_when_all_pings_return(self):
|
|
|
|
+ self.batch.down_minions = {'foo'}
|
|
|
|
+ self.batch.event = MagicMock(
|
|
|
|
+ unpack=MagicMock(return_value=('salt/job/1234/ret/foo', {'id': 'foo'})))
|
|
|
|
+ self.batch.start()
|
|
|
|
+ self.batch._BatchAsync__event_handler(MagicMock())
|
|
|
|
+ self.assertEqual(
|
|
|
|
+ self.batch.event.io_loop.spawn_callback.call_args[0],
|
|
|
|
+ (self.batch.start_batch,))
|
|
|
|
+
|
|
|
|
+ def test_batch__event_handler_not_call_start_batch_when_not_all_pings_return(self):
|
|
|
|
+ self.batch.down_minions = {'foo', 'bar'}
|
|
|
|
+ self.batch.event = MagicMock(
|
|
|
|
+ unpack=MagicMock(return_value=('salt/job/1234/ret/foo', {'id': 'foo'})))
|
|
|
|
+ self.batch.start()
|
|
|
|
+ self.batch._BatchAsync__event_handler(MagicMock())
|
|
|
|
+ self.assertEqual(
|
|
|
|
+ len(self.batch.event.io_loop.spawn_callback.mock_calls), 0)
|
|
|
|
+
|
|
|
|
+ def test_batch__event_handler_batch_run_return(self):
|
|
|
|
+ self.batch.event = MagicMock(
|
|
|
|
+ unpack=MagicMock(return_value=('salt/job/1235/ret/foo', {'id': 'foo'})))
|
|
|
|
+ self.batch.start()
|
|
|
|
+ self.batch.active = {'foo'}
|
|
|
|
+ self.batch._BatchAsync__event_handler(MagicMock())
|
|
|
|
+ self.assertEqual(self.batch.active, set())
|
|
|
|
+ self.assertEqual(self.batch.done_minions, {'foo'})
|
|
|
|
+ self.assertEqual(
|
|
|
|
+ self.batch.event.io_loop.call_later.call_args[0],
|
|
|
|
+ (self.batch.batch_delay, self.batch.schedule_next))
|
|
|
|
+
|
|
|
|
+ def test_batch__event_handler_find_job_return(self):
|
|
|
|
+ self.batch.event = MagicMock(
|
|
|
|
+ unpack=MagicMock(return_value=('salt/job/1236/ret/foo', {'id': 'foo'})))
|
|
|
|
+ self.batch.start()
|
|
|
|
+ self.batch._BatchAsync__event_handler(MagicMock())
|
|
|
|
+ self.assertEqual(self.batch.find_job_returned, {'foo'})
|
|
|
|
+
|
|
|
|
+ @tornado.testing.gen_test
|
|
|
|
+ def test_batch__event_handler_end_batch(self):
|
|
|
|
+ self.batch.event = MagicMock(
|
|
|
|
+ unpack=MagicMock(return_value=('salt/job/not-my-jid/ret/foo', {'id': 'foo'})))
|
|
|
|
+ future = tornado.gen.Future()
|
|
|
|
+ future.set_result({'minions': ['foo', 'bar', 'baz']})
|
|
|
|
+ self.batch.local.run_job_async.return_value = future
|
|
|
|
+ self.batch.start()
|
|
|
|
+ self.batch.initialized = True
|
|
|
|
+ self.assertEqual(self.batch.down_minions, {'foo', 'bar', 'baz'})
|
|
|
|
+ self.batch.end_batch = MagicMock()
|
|
|
|
+ self.batch.minions = {'foo', 'bar', 'baz'}
|
|
|
|
+ self.batch.done_minions = {'foo', 'bar'}
|
|
|
|
+ self.batch.timedout_minions = {'baz'}
|
|
|
|
+ self.batch._BatchAsync__event_handler(MagicMock())
|
|
|
|
+ self.assertEqual(len(self.batch.end_batch.mock_calls), 1)
|
|
|
|
+
|
|
|
|
+ @tornado.testing.gen_test
|
|
|
|
+ def test_batch_find_job(self):
|
|
|
|
+ self.batch.event = MagicMock()
|
|
|
|
+ future = tornado.gen.Future()
|
|
|
|
+ future.set_result({})
|
|
|
|
+ self.batch.local.run_job_async.return_value = future
|
|
|
|
+ self.batch.find_job({'foo', 'bar'})
|
|
|
|
+ self.assertEqual(
|
|
|
|
+ self.batch.event.io_loop.call_later.call_args[0],
|
|
|
|
+ (self.batch.opts['gather_job_timeout'], self.batch.check_find_job, {'foo', 'bar'})
|
|
|
|
+ )
|
|
|
|
+
|
|
|
|
+ @tornado.testing.gen_test
|
|
|
|
+ def test_batch_find_job_with_done_minions(self):
|
|
|
|
+ self.batch.done_minions = {'bar'}
|
|
|
|
+ self.batch.event = MagicMock()
|
|
|
|
+ future = tornado.gen.Future()
|
|
|
|
+ future.set_result({})
|
|
|
|
+ self.batch.local.run_job_async.return_value = future
|
|
|
|
+ self.batch.find_job({'foo', 'bar'})
|
|
|
|
+ self.assertEqual(
|
|
|
|
+ self.batch.event.io_loop.call_later.call_args[0],
|
|
|
|
+ (self.batch.opts['gather_job_timeout'], self.batch.check_find_job, {'foo'})
|
|
|
|
+ )
|
|
|
|
+
|
|
|
|
+ def test_batch_check_find_job_did_not_return(self):
|
|
|
|
+ self.batch.event = MagicMock()
|
|
|
|
+ self.batch.active = {'foo'}
|
|
|
|
+ self.batch.find_job_returned = set()
|
|
|
|
+ self.batch.check_find_job({'foo'})
|
|
|
|
+ self.assertEqual(self.batch.find_job_returned, set())
|
|
|
|
+ self.assertEqual(self.batch.active, set())
|
|
|
|
+ self.assertEqual(len(self.batch.event.io_loop.add_callback.mock_calls), 0)
|
|
|
|
+
|
|
|
|
+ def test_batch_check_find_job_did_return(self):
|
|
|
|
+ self.batch.event = MagicMock()
|
|
|
|
+ self.batch.find_job_returned = {'foo'}
|
|
|
|
+ self.batch.check_find_job({'foo'})
|
|
|
|
+ self.assertEqual(
|
|
|
|
+ self.batch.event.io_loop.add_callback.call_args[0],
|
|
|
|
+ (self.batch.find_job, {'foo'})
|
|
|
|
+ )
|
|
|
|
+
|
|
|
|
+ def test_batch_check_find_job_multiple_states(self):
|
|
|
|
+ self.batch.event = MagicMock()
|
|
|
|
+ # currently running minions
|
|
|
|
+ self.batch.active = {'foo', 'bar'}
|
|
|
|
+
|
|
|
|
+ # minion is running and find_job returns
|
|
|
|
+ self.batch.find_job_returned = {'foo'}
|
|
|
|
+
|
|
|
|
+ # minion started running but find_job did not return
|
|
|
|
+ self.batch.timedout_minions = {'faz'}
|
|
|
|
+
|
|
|
|
+ # minion finished
|
|
|
|
+ self.batch.done_minions = {'baz'}
|
|
|
|
+
|
|
|
|
+ # both not yet done but only 'foo' responded to find_job
|
|
|
|
+ not_done = {'foo', 'bar'}
|
|
|
|
+
|
|
|
|
+ self.batch.check_find_job(not_done)
|
|
|
|
+
|
|
|
|
+ # assert 'bar' removed from active
|
|
|
|
+ self.assertEqual(self.batch.active, {'foo'})
|
|
|
|
+
|
|
|
|
+ # assert 'bar' added to timedout_minions
|
|
|
|
+ self.assertEqual(self.batch.timedout_minions, {'bar', 'faz'})
|
|
|
|
+
|
|
|
|
+ # assert 'find_job' schedueled again only for 'foo'
|
|
|
|
+ self.assertEqual(
|
|
|
|
+ self.batch.event.io_loop.add_callback.call_args[0],
|
|
|
|
+ (self.batch.find_job, {'foo'})
|
|
|
|
+ )
|
|
|
|
--
|
2020-04-07 14:14:01 +02:00
|
|
|
2.16.4
|
2019-04-12 11:57:21 +02:00
|
|
|
|
|
|
|
|