Sync from SUSE:ALP:Source:Standard:1.0 venv-salt-minion revision d20f1886d095224efd326fe2da301de4
This commit is contained in:
parent
a72bcaa312
commit
299b51c817
@ -1 +1 @@
|
||||
cc4e56e0465b20664e2f24bfe7034e5fee37232f
|
||||
d7c50289494a6b41234aee2204209083e30fbd8d
|
26
enhance-cleanup-mechanism-after-salt-bundle-upgrade-.patch
Normal file
26
enhance-cleanup-mechanism-after-salt-bundle-upgrade-.patch
Normal file
@ -0,0 +1,26 @@
|
||||
From 994ebca519945c86dc30f1510dff36b3261446c0 Mon Sep 17 00:00:00 2001
|
||||
From: =?UTF-8?q?Pablo=20Su=C3=A1rez=20Hern=C3=A1ndez?=
|
||||
<psuarezhernandez@suse.com>
|
||||
Date: Thu, 10 Oct 2024 11:51:02 +0100
|
||||
Subject: [PATCH] Enhance cleanup mechanism after Salt Bundle upgrade
|
||||
(bsc#1228690) (#685)
|
||||
|
||||
---
|
||||
pkg/common/venv-salt-minion-postinstall.service | 2 +-
|
||||
1 file changed, 1 insertion(+), 1 deletion(-)
|
||||
|
||||
diff --git a/pkg/common/venv-salt-minion-postinstall.service b/pkg/common/venv-salt-minion-postinstall.service
|
||||
index b122d7d6ea..c9db270435 100644
|
||||
--- a/pkg/common/venv-salt-minion-postinstall.service
|
||||
+++ b/pkg/common/venv-salt-minion-postinstall.service
|
||||
@@ -2,6 +2,6 @@
|
||||
Description=Clean old environment for venv-salt-minion
|
||||
|
||||
[Service]
|
||||
-ExecStart=/bin/sh -c '/usr/lib/venv-salt-minion/bin/post_start_cleanup.sh || :'
|
||||
+ExecStart=/bin/sh -c 'WAIT_IF_SALT_JOBS=1 /usr/lib/venv-salt-minion/bin/post_start_cleanup.sh || :'
|
||||
Type=oneshot
|
||||
|
||||
--
|
||||
2.46.1
|
||||
|
@ -6,4 +6,4 @@ else
|
||||
FINDREQ=/usr/lib/rpm/find-requires
|
||||
fi
|
||||
|
||||
$FINDREQ $* | sed -e '/libpython3.*\.so/d; /libzmq.so/d; /libsodium.so/d; /libffi.so/d; /libyaml-0.so/d; /libvirt/d; /libxml/d; /libexslt/d; /libxslt/d; /bin\/python/d; /bin\/env/d; /bin\/sh/d'
|
||||
$FINDREQ $* | sed -e '/libpython3.*\.so/d; /libzmq.so/d; /libsodium.so/d; /libffi.so/d; /libyaml-0.so/d; /libvirt/d; /libxml/d; /libexslt/d; /libxslt/d; /libdbus/d; /libglib/d; /bin\/python/d; /bin\/env/d; /bin\/sh/d'
|
||||
|
87
fix-x509-private-key-tests-and-test_suse-on-sle12-68.patch
Normal file
87
fix-x509-private-key-tests-and-test_suse-on-sle12-68.patch
Normal file
@ -0,0 +1,87 @@
|
||||
From 43e05d3beea1d6e772fe88c051abf006c2a9bf90 Mon Sep 17 00:00:00 2001
|
||||
From: Marek Czernek <marek.czernek@suse.com>
|
||||
Date: Thu, 3 Oct 2024 13:08:16 +0200
|
||||
Subject: [PATCH] Fix x509 private key tests and test_suse on SLE12
|
||||
(#684)
|
||||
|
||||
---
|
||||
.../functional/modules/test_x509_v2.py | 12 +++++++++--
|
||||
.../functional/states/pkgrepo/test_suse.py | 20 +++++++++----------
|
||||
2 files changed, 20 insertions(+), 12 deletions(-)
|
||||
|
||||
diff --git a/tests/pytests/functional/modules/test_x509_v2.py b/tests/pytests/functional/modules/test_x509_v2.py
|
||||
index 7de8f3b01f..3db78c1b63 100644
|
||||
--- a/tests/pytests/functional/modules/test_x509_v2.py
|
||||
+++ b/tests/pytests/functional/modules/test_x509_v2.py
|
||||
@@ -1442,14 +1442,22 @@ def test_create_private_key_with_passphrase(x509, algo):
|
||||
|
||||
@pytest.mark.slow_test
|
||||
def test_create_private_key_der(x509):
|
||||
- res = x509.create_private_key(algo="ec", encoding="der")
|
||||
+ try:
|
||||
+ res = x509.create_private_key(algo="ec", encoding="der")
|
||||
+ except NotImplementedError:
|
||||
+ pytest.skip("Algorithm 'ec' is not supported on this OpenSSL version")
|
||||
assert base64.b64decode(res)
|
||||
|
||||
|
||||
@pytest.mark.slow_test
|
||||
@pytest.mark.parametrize("passphrase", [None, "hunter2"])
|
||||
def test_create_private_key_pkcs12(x509, passphrase):
|
||||
- res = x509.create_private_key(algo="ec", encoding="pkcs12", passphrase=passphrase)
|
||||
+ try:
|
||||
+ res = x509.create_private_key(
|
||||
+ algo="ec", encoding="pkcs12", passphrase=passphrase
|
||||
+ )
|
||||
+ except NotImplementedError:
|
||||
+ pytest.skip("Algorithm 'ec' is not supported on this OpenSSL version")
|
||||
assert base64.b64decode(res)
|
||||
|
||||
|
||||
diff --git a/tests/pytests/functional/states/pkgrepo/test_suse.py b/tests/pytests/functional/states/pkgrepo/test_suse.py
|
||||
index 3bafeedc94..d21a9aeb9d 100644
|
||||
--- a/tests/pytests/functional/states/pkgrepo/test_suse.py
|
||||
+++ b/tests/pytests/functional/states/pkgrepo/test_suse.py
|
||||
@@ -28,14 +28,14 @@ def suse_state_tree(grains, pkgrepo, state_tree):
|
||||
- comments:
|
||||
- '# Salt Test'
|
||||
- refresh: 1
|
||||
- {% if grains['osmajorrelease'] == 15 %}
|
||||
- - baseurl: https://download.opensuse.org/repositories/openSUSE:/Backports:/SLE-15-SP4/standard/
|
||||
- - humanname: openSUSE Backports for SLE 15 SP4
|
||||
- - gpgkey: https://download.opensuse.org/repositories/openSUSE:/Backports:/SLE-15-SP4/standard/repodata/repomd.xml.key
|
||||
- {% elif grains['osfullname'] == 'openSUSE Tumbleweed' %}
|
||||
+ {% if grains['osfullname'] == 'openSUSE Tumbleweed' %}
|
||||
- baseurl: http://download.opensuse.org/tumbleweed/repo/oss/
|
||||
- humanname: openSUSE Tumbleweed OSS
|
||||
- gpgkey: https://download.opensuse.org/tumbleweed/repo/oss/repodata/repomd.xml.key
|
||||
+ {% else %}
|
||||
+ - baseurl: https://download.opensuse.org/repositories/openSUSE:/Backports:/SLE-15-SP4/standard/
|
||||
+ - humanname: openSUSE Backports for SLE 15 SP4
|
||||
+ - gpgkey: https://download.opensuse.org/repositories/openSUSE:/Backports:/SLE-15-SP4/standard/repodata/repomd.xml.key
|
||||
{% endif %}
|
||||
"""
|
||||
|
||||
@@ -53,14 +53,14 @@ def suse_state_tree(grains, pkgrepo, state_tree):
|
||||
- comments:
|
||||
- '# Salt Test (modified)'
|
||||
- refresh: 1
|
||||
- {% if grains['osmajorrelease'] == 15 %}
|
||||
- - baseurl: https://download.opensuse.org/repositories/openSUSE:/Backports:/SLE-15-SP4/standard/
|
||||
- - humanname: Salt modified Backports
|
||||
- - gpgkey: https://download.opensuse.org/repositories/openSUSE:/Backports:/SLE-15-SP4/standard/repodata/repomd.xml.key
|
||||
- {% elif grains['osfullname'] == 'openSUSE Tumbleweed' %}
|
||||
+ {% if grains['osfullname'] == 'openSUSE Tumbleweed' %}
|
||||
- baseurl: http://download.opensuse.org/tumbleweed/repo/oss/
|
||||
- humanname: Salt modified OSS
|
||||
- gpgkey: https://download.opensuse.org/tumbleweed/repo/oss/repodata/repomd.xml.key
|
||||
+ {% else %}
|
||||
+ - baseurl: https://download.opensuse.org/repositories/openSUSE:/Backports:/SLE-15-SP4/standard/
|
||||
+ - humanname: Salt modified Backports
|
||||
+ - gpgkey: https://download.opensuse.org/repositories/openSUSE:/Backports:/SLE-15-SP4/standard/repodata/repomd.xml.key
|
||||
{% endif %}
|
||||
"""
|
||||
|
||||
--
|
||||
2.46.1
|
||||
|
58
handle-logger-flushing-already-closed-file-686.patch
Normal file
58
handle-logger-flushing-already-closed-file-686.patch
Normal file
@ -0,0 +1,58 @@
|
||||
From e23dce108588a9c52d3f7542636892750d6efcbd Mon Sep 17 00:00:00 2001
|
||||
From: Marek Czernek <marek.czernek@suse.com>
|
||||
Date: Thu, 17 Oct 2024 15:52:00 +0200
|
||||
Subject: [PATCH] Handle logger flushing already closed file (#686)
|
||||
|
||||
This is a partial cherry-pick of
|
||||
https://github.com/saltstack/salt/commit/9683260d61668da8559ecde6caf63a52fedd8790
|
||||
---
|
||||
salt/_logging/handlers.py | 4 ++++
|
||||
salt/_logging/impl.py | 10 +++++++++-
|
||||
2 files changed, 13 insertions(+), 1 deletion(-)
|
||||
|
||||
diff --git a/salt/_logging/handlers.py b/salt/_logging/handlers.py
|
||||
index 5a1a1613137..d8bc68a49db 100644
|
||||
--- a/salt/_logging/handlers.py
|
||||
+++ b/salt/_logging/handlers.py
|
||||
@@ -95,6 +95,9 @@ class DeferredStreamHandler(StreamHandler):
|
||||
super().__init__(stream)
|
||||
self.__messages = deque(maxlen=max_queue_size)
|
||||
self.__emitting = False
|
||||
+ import traceback
|
||||
+
|
||||
+ self.stack = "".join(traceback.format_stack())
|
||||
|
||||
def handle(self, record):
|
||||
self.acquire()
|
||||
@@ -116,6 +119,7 @@ class DeferredStreamHandler(StreamHandler):
|
||||
super().handle(record)
|
||||
finally:
|
||||
self.__emitting = False
|
||||
+ # This will raise a ValueError if the file handle has been closed.
|
||||
super().flush()
|
||||
|
||||
def sync_with_handlers(self, handlers=()):
|
||||
diff --git a/salt/_logging/impl.py b/salt/_logging/impl.py
|
||||
index 4d1ebd2495f..9d76c3174e2 100644
|
||||
--- a/salt/_logging/impl.py
|
||||
+++ b/salt/_logging/impl.py
|
||||
@@ -488,7 +488,15 @@ def setup_temp_handler(log_level=None):
|
||||
break
|
||||
else:
|
||||
handler = DeferredStreamHandler(sys.stderr)
|
||||
- atexit.register(handler.flush)
|
||||
+
|
||||
+ def tryflush():
|
||||
+ try:
|
||||
+ handler.flush()
|
||||
+ except ValueError:
|
||||
+ # File handle has already been closed.
|
||||
+ pass
|
||||
+
|
||||
+ atexit.register(tryflush)
|
||||
handler.setLevel(log_level)
|
||||
|
||||
# Set the default temporary console formatter config
|
||||
--
|
||||
2.47.0
|
||||
|
@ -97,3 +97,6 @@ saltbundlepy-devel.rpm$
|
||||
saltbundlepy-salt-test.rpm$
|
||||
saltbundlepy-pathspec.rpm$
|
||||
saltbundlepy-yamllint.rpm$
|
||||
|
||||
# Include DBus module used for snapper
|
||||
saltbundlepy-dbus-python.rpm$
|
||||
|
770
make-minion-reconnecting-on-changing-master-ip-bsc-1.patch
Normal file
770
make-minion-reconnecting-on-changing-master-ip-bsc-1.patch
Normal file
@ -0,0 +1,770 @@
|
||||
From eb6c67e6f535cdfbf685a54c6352018673e37a12 Mon Sep 17 00:00:00 2001
|
||||
From: Victor Zhestkov <vzhestkov@suse.com>
|
||||
Date: Tue, 26 Nov 2024 11:59:08 +0300
|
||||
Subject: [PATCH] Make minion reconnecting on changing master IP
|
||||
(bsc#1228182)
|
||||
|
||||
* Minions check dns when re-connecting to a master
|
||||
|
||||
Check for a chainging dns record anytime a minion gets disconnected from
|
||||
it's master. See github issue #63654 #61482.
|
||||
|
||||
* Regression tests for dns defined masters
|
||||
|
||||
Adding tests to validate we check for changing dns anytime we're
|
||||
disconnected from the currently connected master
|
||||
|
||||
* Update docs for master dns changes
|
||||
|
||||
Update docs to use master_alive_interval to detect master ip changes via
|
||||
DNS.
|
||||
|
||||
* Remove comment which is not true anymore
|
||||
|
||||
* Make minion reconnecting on changing master IP
|
||||
|
||||
with zeromq transport
|
||||
|
||||
* Don't create schedule for alive if no master_alive_interval
|
||||
|
||||
* Skip the tests if running with non-root user
|
||||
|
||||
* Skip if unable to set additional IP address
|
||||
|
||||
* Set master_tries to -1 for minions
|
||||
|
||||
* Fix the tests
|
||||
|
||||
---------
|
||||
|
||||
Co-authored-by: Daniel A. Wozniak <daniel.wozniak@broadcom.com>
|
||||
---
|
||||
conf/minion | 5 +-
|
||||
doc/ref/configuration/minion.rst | 4 +-
|
||||
salt/channel/client.py | 2 -
|
||||
salt/config/__init__.py | 4 +-
|
||||
salt/minion.py | 190 ++++++++----------
|
||||
salt/transport/zeromq.py | 17 +-
|
||||
tests/pytests/scenarios/dns/__init__.py | 0
|
||||
tests/pytests/scenarios/dns/conftest.py | 99 +++++++++
|
||||
.../scenarios/dns/multimaster/conftest.py | 124 ++++++++++++
|
||||
.../scenarios/dns/multimaster/test_dns.py | 54 +++++
|
||||
tests/pytests/scenarios/dns/test_dns.py | 37 ++++
|
||||
.../multimaster/test_failover_master.py | 4 -
|
||||
tests/pytests/unit/test_minion.py | 2 +
|
||||
13 files changed, 422 insertions(+), 120 deletions(-)
|
||||
create mode 100644 tests/pytests/scenarios/dns/__init__.py
|
||||
create mode 100644 tests/pytests/scenarios/dns/conftest.py
|
||||
create mode 100644 tests/pytests/scenarios/dns/multimaster/conftest.py
|
||||
create mode 100644 tests/pytests/scenarios/dns/multimaster/test_dns.py
|
||||
create mode 100644 tests/pytests/scenarios/dns/test_dns.py
|
||||
|
||||
diff --git a/conf/minion b/conf/minion
|
||||
index eeef626fa8..f89e18451f 100644
|
||||
--- a/conf/minion
|
||||
+++ b/conf/minion
|
||||
@@ -271,9 +271,8 @@
|
||||
#ping_interval: 0
|
||||
|
||||
# To auto recover minions if master changes IP address (DDNS)
|
||||
-# auth_tries: 10
|
||||
-# auth_safemode: True
|
||||
-# ping_interval: 2
|
||||
+# master_alive_interval: 10
|
||||
+# master_tries: -1
|
||||
#
|
||||
# Minions won't know master is missing until a ping fails. After the ping fail,
|
||||
# the minion will attempt authentication and likely fails out and cause a restart.
|
||||
diff --git a/doc/ref/configuration/minion.rst b/doc/ref/configuration/minion.rst
|
||||
index 57af5ce4a3..a1b0f2e86e 100644
|
||||
--- a/doc/ref/configuration/minion.rst
|
||||
+++ b/doc/ref/configuration/minion.rst
|
||||
@@ -291,7 +291,9 @@ Default: ``0``
|
||||
|
||||
Configures how often, in seconds, the minion will verify that the current
|
||||
master is alive and responding. The minion will try to establish a connection
|
||||
-to the next master in the list if it finds the existing one is dead.
|
||||
+to the next master in the list if it finds the existing one is dead. This
|
||||
+setting can also be used to detect master DNS record changes when a minion has
|
||||
+been disconnected.
|
||||
|
||||
.. code-block:: yaml
|
||||
|
||||
diff --git a/salt/channel/client.py b/salt/channel/client.py
|
||||
index 76d7a8e5b9..34aafb2c9e 100644
|
||||
--- a/salt/channel/client.py
|
||||
+++ b/salt/channel/client.py
|
||||
@@ -385,8 +385,6 @@ class AsyncPubChannel:
|
||||
# else take the relayed publish_port master reports
|
||||
else:
|
||||
publish_port = self.auth.creds["publish_port"]
|
||||
- # TODO: The zeromq transport does not use connect_callback and
|
||||
- # disconnect_callback.
|
||||
yield self.transport.connect(
|
||||
publish_port, self.connect_callback, self.disconnect_callback
|
||||
)
|
||||
diff --git a/salt/config/__init__.py b/salt/config/__init__.py
|
||||
index b3cd5d85ae..d4865807e6 100644
|
||||
--- a/salt/config/__init__.py
|
||||
+++ b/salt/config/__init__.py
|
||||
@@ -75,7 +75,7 @@ elif salt.utils.platform.is_darwin():
|
||||
else:
|
||||
_DFLT_IPC_MODE = "ipc"
|
||||
_DFLT_FQDNS_GRAINS = False
|
||||
- _MASTER_TRIES = 1
|
||||
+ _MASTER_TRIES = -1
|
||||
_MASTER_USER = salt.utils.user.get_user()
|
||||
|
||||
|
||||
@@ -1272,7 +1272,7 @@ DEFAULT_MINION_OPTS = immutabletypes.freeze(
|
||||
"username": None,
|
||||
"password": None,
|
||||
"zmq_filtering": False,
|
||||
- "zmq_monitor": False,
|
||||
+ "zmq_monitor": True,
|
||||
"cache_sreqs": True,
|
||||
"cmd_safe": True,
|
||||
"sudo_user": "",
|
||||
diff --git a/salt/minion.py b/salt/minion.py
|
||||
index e21a017cfd..834f0848c6 100644
|
||||
--- a/salt/minion.py
|
||||
+++ b/salt/minion.py
|
||||
@@ -2737,10 +2737,64 @@ class Minion(MinionBase):
|
||||
# we are not connected anymore
|
||||
self.connected = False
|
||||
log.info("Connection to master %s lost", self.opts["master"])
|
||||
+ if self.opts["transport"] != "tcp":
|
||||
+ self.schedule.delete_job(name=master_event(type="alive"))
|
||||
+
|
||||
+ log.info("Trying to tune in to next master from master-list")
|
||||
+
|
||||
+ if hasattr(self, "pub_channel"):
|
||||
+ self.pub_channel.on_recv(None)
|
||||
+ if hasattr(self.pub_channel, "auth"):
|
||||
+ self.pub_channel.auth.invalidate()
|
||||
+ if hasattr(self.pub_channel, "close"):
|
||||
+ self.pub_channel.close()
|
||||
+ if hasattr(self, "req_channel") and self.req_channel:
|
||||
+ self.req_channel.close()
|
||||
+ self.req_channel = None
|
||||
+
|
||||
+ # if eval_master finds a new master for us, self.connected
|
||||
+ # will be True again on successful master authentication
|
||||
+ try:
|
||||
+ master, self.pub_channel = yield self.eval_master(
|
||||
+ opts=self.opts,
|
||||
+ failed=True,
|
||||
+ failback=tag.startswith(master_event(type="failback")),
|
||||
+ )
|
||||
+ except SaltClientError:
|
||||
+ pass
|
||||
+
|
||||
+ if self.connected:
|
||||
+ self.opts["master"] = master
|
||||
+
|
||||
+ # re-init the subsystems to work with the new master
|
||||
+ log.info(
|
||||
+ "Re-initialising subsystems for new master %s",
|
||||
+ self.opts["master"],
|
||||
+ )
|
||||
+
|
||||
+ self.req_channel = salt.channel.client.AsyncReqChannel.factory(
|
||||
+ self.opts, io_loop=self.io_loop
|
||||
+ )
|
||||
|
||||
- if self.opts["master_type"] != "failover":
|
||||
- # modify the scheduled job to fire on reconnect
|
||||
- if self.opts["transport"] != "tcp":
|
||||
+ # put the current schedule into the new loaders
|
||||
+ self.opts["schedule"] = self.schedule.option("schedule")
|
||||
+ (
|
||||
+ self.functions,
|
||||
+ self.returners,
|
||||
+ self.function_errors,
|
||||
+ self.executors,
|
||||
+ ) = self._load_modules()
|
||||
+ # make the schedule to use the new 'functions' loader
|
||||
+ self.schedule.functions = self.functions
|
||||
+ self.pub_channel.on_recv(self._handle_payload)
|
||||
+ self._fire_master_minion_start()
|
||||
+ log.info("Minion is ready to receive requests!")
|
||||
+
|
||||
+ # update scheduled job to run with the new master addr
|
||||
+ if (
|
||||
+ self.opts["transport"] != "tcp"
|
||||
+ and self.opts["master_alive_interval"] > 0
|
||||
+ ):
|
||||
schedule = {
|
||||
"function": "status.master",
|
||||
"seconds": self.opts["master_alive_interval"],
|
||||
@@ -2749,116 +2803,35 @@ class Minion(MinionBase):
|
||||
"return_job": False,
|
||||
"kwargs": {
|
||||
"master": self.opts["master"],
|
||||
- "connected": False,
|
||||
+ "connected": True,
|
||||
},
|
||||
}
|
||||
self.schedule.modify_job(
|
||||
name=master_event(type="alive", master=self.opts["master"]),
|
||||
schedule=schedule,
|
||||
)
|
||||
- else:
|
||||
- # delete the scheduled job to don't interfere with the failover process
|
||||
- if self.opts["transport"] != "tcp":
|
||||
- self.schedule.delete_job(name=master_event(type="alive"))
|
||||
-
|
||||
- log.info("Trying to tune in to next master from master-list")
|
||||
-
|
||||
- if hasattr(self, "pub_channel"):
|
||||
- self.pub_channel.on_recv(None)
|
||||
- if hasattr(self.pub_channel, "auth"):
|
||||
- self.pub_channel.auth.invalidate()
|
||||
- if hasattr(self.pub_channel, "close"):
|
||||
- self.pub_channel.close()
|
||||
- del self.pub_channel
|
||||
-
|
||||
- # if eval_master finds a new master for us, self.connected
|
||||
- # will be True again on successful master authentication
|
||||
- try:
|
||||
- master, self.pub_channel = yield self.eval_master(
|
||||
- opts=self.opts,
|
||||
- failed=True,
|
||||
- failback=tag.startswith(master_event(type="failback")),
|
||||
- )
|
||||
- except SaltClientError:
|
||||
- pass
|
||||
-
|
||||
- if self.connected:
|
||||
- self.opts["master"] = master
|
||||
-
|
||||
- # re-init the subsystems to work with the new master
|
||||
- log.info(
|
||||
- "Re-initialising subsystems for new master %s",
|
||||
- self.opts["master"],
|
||||
- )
|
||||
-
|
||||
- self.req_channel = (
|
||||
- salt.transport.client.AsyncReqChannel.factory(
|
||||
- self.opts, io_loop=self.io_loop
|
||||
- )
|
||||
- )
|
||||
-
|
||||
- # put the current schedule into the new loaders
|
||||
- self.opts["schedule"] = self.schedule.option("schedule")
|
||||
- (
|
||||
- self.functions,
|
||||
- self.returners,
|
||||
- self.function_errors,
|
||||
- self.executors,
|
||||
- ) = self._load_modules()
|
||||
- # make the schedule to use the new 'functions' loader
|
||||
- self.schedule.functions = self.functions
|
||||
- self.pub_channel.on_recv(self._handle_payload)
|
||||
- self._fire_master_minion_start()
|
||||
- log.info("Minion is ready to receive requests!")
|
||||
-
|
||||
- # update scheduled job to run with the new master addr
|
||||
- if self.opts["transport"] != "tcp":
|
||||
- schedule = {
|
||||
- "function": "status.master",
|
||||
- "seconds": self.opts["master_alive_interval"],
|
||||
- "jid_include": True,
|
||||
- "maxrunning": 1,
|
||||
- "return_job": False,
|
||||
- "kwargs": {
|
||||
- "master": self.opts["master"],
|
||||
- "connected": True,
|
||||
- },
|
||||
- }
|
||||
- self.schedule.modify_job(
|
||||
- name=master_event(
|
||||
- type="alive", master=self.opts["master"]
|
||||
- ),
|
||||
- schedule=schedule,
|
||||
- )
|
||||
|
||||
- if (
|
||||
- self.opts["master_failback"]
|
||||
- and "master_list" in self.opts
|
||||
- ):
|
||||
- if self.opts["master"] != self.opts["master_list"][0]:
|
||||
- schedule = {
|
||||
- "function": "status.ping_master",
|
||||
- "seconds": self.opts[
|
||||
- "master_failback_interval"
|
||||
- ],
|
||||
- "jid_include": True,
|
||||
- "maxrunning": 1,
|
||||
- "return_job": False,
|
||||
- "kwargs": {
|
||||
- "master": self.opts["master_list"][0]
|
||||
- },
|
||||
- }
|
||||
- self.schedule.modify_job(
|
||||
- name=master_event(type="failback"),
|
||||
- schedule=schedule,
|
||||
- )
|
||||
- else:
|
||||
- self.schedule.delete_job(
|
||||
- name=master_event(type="failback"), persist=True
|
||||
- )
|
||||
- else:
|
||||
- self.restart = True
|
||||
- self.io_loop.stop()
|
||||
+ if self.opts["master_failback"] and "master_list" in self.opts:
|
||||
+ if self.opts["master"] != self.opts["master_list"][0]:
|
||||
+ schedule = {
|
||||
+ "function": "status.ping_master",
|
||||
+ "seconds": self.opts["master_failback_interval"],
|
||||
+ "jid_include": True,
|
||||
+ "maxrunning": 1,
|
||||
+ "return_job": False,
|
||||
+ "kwargs": {"master": self.opts["master_list"][0]},
|
||||
+ }
|
||||
+ self.schedule.modify_job(
|
||||
+ name=master_event(type="failback"),
|
||||
+ schedule=schedule,
|
||||
+ )
|
||||
+ else:
|
||||
+ self.schedule.delete_job(
|
||||
+ name=master_event(type="failback"), persist=True
|
||||
+ )
|
||||
+ else:
|
||||
+ self.restart = True
|
||||
+ self.io_loop.stop()
|
||||
|
||||
elif tag.startswith(master_event(type="connected")):
|
||||
# handle this event only once. otherwise it will pollute the log
|
||||
@@ -2870,7 +2843,10 @@ class Minion(MinionBase):
|
||||
self.connected = True
|
||||
# modify the __master_alive job to only fire,
|
||||
# if the connection is lost again
|
||||
- if self.opts["transport"] != "tcp":
|
||||
+ if (
|
||||
+ self.opts["transport"] != "tcp"
|
||||
+ and self.opts["master_alive_interval"] > 0
|
||||
+ ):
|
||||
schedule = {
|
||||
"function": "status.master",
|
||||
"seconds": self.opts["master_alive_interval"],
|
||||
diff --git a/salt/transport/zeromq.py b/salt/transport/zeromq.py
|
||||
index 7cc6b9987f..89f705190e 100644
|
||||
--- a/salt/transport/zeromq.py
|
||||
+++ b/salt/transport/zeromq.py
|
||||
@@ -1,6 +1,7 @@
|
||||
"""
|
||||
Zeromq transport classes
|
||||
"""
|
||||
+
|
||||
import errno
|
||||
import hashlib
|
||||
import logging
|
||||
@@ -211,6 +212,12 @@ class PublishClient(salt.transport.base.PublishClient):
|
||||
self.master_pub,
|
||||
)
|
||||
log.debug("%r connecting to %s", self, self.master_pub)
|
||||
+ if (
|
||||
+ hasattr(self, "_monitor")
|
||||
+ and self._monitor is not None
|
||||
+ and disconnect_callback is not None
|
||||
+ ):
|
||||
+ self._monitor.disconnect_callback = disconnect_callback
|
||||
self._socket.connect(self.master_pub)
|
||||
connect_callback(True)
|
||||
|
||||
@@ -680,13 +687,21 @@ class ZeroMQSocketMonitor:
|
||||
log.debug("ZeroMQ event: %s", evt)
|
||||
if evt["event"] == zmq.EVENT_MONITOR_STOPPED:
|
||||
self.stop()
|
||||
+ elif evt["event"] == zmq.EVENT_DISCONNECTED:
|
||||
+ if (
|
||||
+ hasattr(self, "disconnect_callback")
|
||||
+ and self.disconnect_callback is not None
|
||||
+ ):
|
||||
+ self.disconnect_callback()
|
||||
|
||||
def stop(self):
|
||||
if self._socket is None:
|
||||
return
|
||||
self._socket.disable_monitor()
|
||||
self._socket = None
|
||||
- self._monitor_socket = None
|
||||
+ if self._monitor_socket is not None:
|
||||
+ self._monitor_socket.close()
|
||||
+ self._monitor_socket = None
|
||||
if self._monitor_stream is not None:
|
||||
self._monitor_stream.close()
|
||||
self._monitor_stream = None
|
||||
diff --git a/tests/pytests/scenarios/dns/__init__.py b/tests/pytests/scenarios/dns/__init__.py
|
||||
new file mode 100644
|
||||
index 0000000000..e69de29bb2
|
||||
diff --git a/tests/pytests/scenarios/dns/conftest.py b/tests/pytests/scenarios/dns/conftest.py
|
||||
new file mode 100644
|
||||
index 0000000000..5a8850719f
|
||||
--- /dev/null
|
||||
+++ b/tests/pytests/scenarios/dns/conftest.py
|
||||
@@ -0,0 +1,99 @@
|
||||
+import logging
|
||||
+import pathlib
|
||||
+import subprocess
|
||||
+
|
||||
+import pytest
|
||||
+
|
||||
+log = logging.getLogger(__name__)
|
||||
+
|
||||
+
|
||||
+@pytest.fixture(scope="package")
|
||||
+def master_alive_interval():
|
||||
+ return 5
|
||||
+
|
||||
+
|
||||
+class HostsFile:
|
||||
+ """
|
||||
+ Simple helper class for tests that need to modify /etc/hosts.
|
||||
+ """
|
||||
+
|
||||
+ def __init__(self, path, orig_text):
|
||||
+ self._path = path
|
||||
+ self._orig_text = orig_text
|
||||
+
|
||||
+ @property
|
||||
+ def orig_text(self):
|
||||
+ return self._orig_text
|
||||
+
|
||||
+ def __getattr__(self, key):
|
||||
+ if key in ["_path", "_orig_text", "orig_text"]:
|
||||
+ return self.__getattribute__(key)
|
||||
+ return getattr(self._path, key)
|
||||
+
|
||||
+
|
||||
+@pytest.fixture
|
||||
+def etc_hosts():
|
||||
+ hosts = pathlib.Path("/etc/hosts")
|
||||
+ orig_text = hosts.read_text(encoding="utf-8")
|
||||
+ hosts = HostsFile(hosts, orig_text)
|
||||
+ try:
|
||||
+ yield hosts
|
||||
+ finally:
|
||||
+ hosts.write_text(orig_text)
|
||||
+
|
||||
+
|
||||
+@pytest.fixture(scope="package")
|
||||
+def master(request, salt_factories):
|
||||
+
|
||||
+ try:
|
||||
+ subprocess.check_output(["ip", "addr", "add", "172.16.0.1/32", "dev", "lo"])
|
||||
+ ip_addr_set = True
|
||||
+ except subprocess.CalledProcessError:
|
||||
+ ip_addr_set = False
|
||||
+
|
||||
+ config_defaults = {
|
||||
+ "open_mode": True,
|
||||
+ "transport": request.config.getoption("--transport"),
|
||||
+ }
|
||||
+ config_overrides = {
|
||||
+ "interface": "0.0.0.0",
|
||||
+ }
|
||||
+ factory = salt_factories.salt_master_daemon(
|
||||
+ "master",
|
||||
+ defaults=config_defaults,
|
||||
+ overrides=config_overrides,
|
||||
+ extra_cli_arguments_after_first_start_failure=["--log-level=info"],
|
||||
+ )
|
||||
+ factory.ip_addr_set = ip_addr_set
|
||||
+ with factory.started(start_timeout=180):
|
||||
+ yield factory
|
||||
+
|
||||
+ try:
|
||||
+ subprocess.check_output(["ip", "addr", "del", "172.16.0.1/32", "dev", "lo"])
|
||||
+ except subprocess.CalledProcessError:
|
||||
+ pass
|
||||
+
|
||||
+
|
||||
+@pytest.fixture(scope="package")
|
||||
+def salt_cli(master):
|
||||
+ return master.salt_cli(timeout=180)
|
||||
+
|
||||
+
|
||||
+@pytest.fixture(scope="package")
|
||||
+def minion(master, master_alive_interval):
|
||||
+ config_defaults = {
|
||||
+ "transport": master.config["transport"],
|
||||
+ }
|
||||
+ port = master.config["ret_port"]
|
||||
+ config_overrides = {
|
||||
+ "master": f"master.local:{port}",
|
||||
+ "publish_port": master.config["publish_port"],
|
||||
+ "master_alive_interval": master_alive_interval,
|
||||
+ }
|
||||
+ factory = master.salt_minion_daemon(
|
||||
+ "minion",
|
||||
+ defaults=config_defaults,
|
||||
+ overrides=config_overrides,
|
||||
+ extra_cli_arguments_after_first_start_failure=["--log-level=info"],
|
||||
+ )
|
||||
+ return factory
|
||||
diff --git a/tests/pytests/scenarios/dns/multimaster/conftest.py b/tests/pytests/scenarios/dns/multimaster/conftest.py
|
||||
new file mode 100644
|
||||
index 0000000000..3333f812ce
|
||||
--- /dev/null
|
||||
+++ b/tests/pytests/scenarios/dns/multimaster/conftest.py
|
||||
@@ -0,0 +1,124 @@
|
||||
+import logging
|
||||
+import os
|
||||
+import shutil
|
||||
+import subprocess
|
||||
+
|
||||
+import pytest
|
||||
+
|
||||
+log = logging.getLogger(__name__)
|
||||
+
|
||||
+
|
||||
+@pytest.fixture(scope="package")
|
||||
+def salt_mm_master_1(request, salt_factories):
|
||||
+
|
||||
+ try:
|
||||
+ subprocess.check_output(["ip", "addr", "add", "172.16.0.1/32", "dev", "lo"])
|
||||
+ ip_addr_set = True
|
||||
+ except subprocess.CalledProcessError:
|
||||
+ ip_addr_set = False
|
||||
+
|
||||
+ config_defaults = {
|
||||
+ "open_mode": True,
|
||||
+ "transport": request.config.getoption("--transport"),
|
||||
+ }
|
||||
+ config_overrides = {
|
||||
+ "interface": "0.0.0.0",
|
||||
+ "master_sign_pubkey": True,
|
||||
+ }
|
||||
+ factory = salt_factories.salt_master_daemon(
|
||||
+ "mm-master-1",
|
||||
+ defaults=config_defaults,
|
||||
+ overrides=config_overrides,
|
||||
+ extra_cli_arguments_after_first_start_failure=["--log-level=info"],
|
||||
+ )
|
||||
+ factory.ip_addr_set = ip_addr_set
|
||||
+ try:
|
||||
+ with factory.started(start_timeout=180):
|
||||
+ yield factory
|
||||
+ finally:
|
||||
+
|
||||
+ try:
|
||||
+ subprocess.check_output(["ip", "addr", "del", "172.16.0.1/32", "dev", "lo"])
|
||||
+ except subprocess.CalledProcessError:
|
||||
+ pass
|
||||
+
|
||||
+
|
||||
+@pytest.fixture(scope="package")
|
||||
+def mm_master_1_salt_cli(salt_mm_master_1):
|
||||
+ return salt_mm_master_1.salt_cli(timeout=180)
|
||||
+
|
||||
+
|
||||
+@pytest.fixture(scope="package")
|
||||
+def salt_mm_master_2(salt_factories, salt_mm_master_1):
|
||||
+ # if salt.utils.platform.is_darwin() or salt.utils.platform.is_freebsd():
|
||||
+ # subprocess.check_output(["ifconfig", "lo0", "alias", "127.0.0.2", "up"])
|
||||
+
|
||||
+ config_defaults = {
|
||||
+ "open_mode": True,
|
||||
+ "transport": salt_mm_master_1.config["transport"],
|
||||
+ }
|
||||
+ config_overrides = {
|
||||
+ "interface": "0.0.0.0",
|
||||
+ "master_sign_pubkey": True,
|
||||
+ }
|
||||
+
|
||||
+ # Use the same ports for both masters, they are binding to different interfaces
|
||||
+ for key in (
|
||||
+ "ret_port",
|
||||
+ "publish_port",
|
||||
+ ):
|
||||
+ config_overrides[key] = salt_mm_master_1.config[key] + 1
|
||||
+ factory = salt_factories.salt_master_daemon(
|
||||
+ "mm-master-2",
|
||||
+ defaults=config_defaults,
|
||||
+ overrides=config_overrides,
|
||||
+ extra_cli_arguments_after_first_start_failure=["--log-level=info"],
|
||||
+ )
|
||||
+
|
||||
+ # Both masters will share the same signing key pair
|
||||
+ for keyfile in ("master_sign.pem", "master_sign.pub"):
|
||||
+ shutil.copyfile(
|
||||
+ os.path.join(salt_mm_master_1.config["pki_dir"], keyfile),
|
||||
+ os.path.join(factory.config["pki_dir"], keyfile),
|
||||
+ )
|
||||
+ with factory.started(start_timeout=180):
|
||||
+ yield factory
|
||||
+
|
||||
+
|
||||
+@pytest.fixture(scope="package")
|
||||
+def mm_master_2_salt_cli(salt_mm_master_2):
|
||||
+ return salt_mm_master_2.salt_cli(timeout=180)
|
||||
+
|
||||
+
|
||||
+@pytest.fixture(scope="package")
|
||||
+def salt_mm_minion_1(salt_mm_master_1, salt_mm_master_2, master_alive_interval):
|
||||
+ config_defaults = {
|
||||
+ "transport": salt_mm_master_1.config["transport"],
|
||||
+ }
|
||||
+
|
||||
+ mm_master_1_port = salt_mm_master_1.config["ret_port"]
|
||||
+ mm_master_2_port = salt_mm_master_2.config["ret_port"]
|
||||
+ config_overrides = {
|
||||
+ "master": [
|
||||
+ f"master1.local:{mm_master_1_port}",
|
||||
+ f"master2.local:{mm_master_2_port}",
|
||||
+ ],
|
||||
+ "publish_port": salt_mm_master_1.config["publish_port"],
|
||||
+ "master_alive_interval": master_alive_interval,
|
||||
+ "master_tries": -1,
|
||||
+ "verify_master_pubkey_sign": True,
|
||||
+ "retry_dns": True,
|
||||
+ }
|
||||
+ factory = salt_mm_master_1.salt_minion_daemon(
|
||||
+ "mm-minion-1",
|
||||
+ defaults=config_defaults,
|
||||
+ overrides=config_overrides,
|
||||
+ extra_cli_arguments_after_first_start_failure=["--log-level=info"],
|
||||
+ )
|
||||
+ # Need to grab the public signing key from the master, either will do
|
||||
+ shutil.copyfile(
|
||||
+ os.path.join(salt_mm_master_1.config["pki_dir"], "master_sign.pub"),
|
||||
+ os.path.join(factory.config["pki_dir"], "master_sign.pub"),
|
||||
+ )
|
||||
+ # with factory.started(start_timeout=180):
|
||||
+ yield factory
|
||||
diff --git a/tests/pytests/scenarios/dns/multimaster/test_dns.py b/tests/pytests/scenarios/dns/multimaster/test_dns.py
|
||||
new file mode 100644
|
||||
index 0000000000..fafb30c12e
|
||||
--- /dev/null
|
||||
+++ b/tests/pytests/scenarios/dns/multimaster/test_dns.py
|
||||
@@ -0,0 +1,54 @@
|
||||
+import logging
|
||||
+import subprocess
|
||||
+import time
|
||||
+
|
||||
+import pytest
|
||||
+
|
||||
+log = logging.getLogger(__name__)
|
||||
+
|
||||
+
|
||||
+@pytest.mark.skip_unless_on_linux
|
||||
+@pytest.mark.skip_if_not_root
|
||||
+def test_multimaster_dns(
|
||||
+ salt_mm_master_1,
|
||||
+ salt_mm_minion_1,
|
||||
+ mm_master_1_salt_cli,
|
||||
+ etc_hosts,
|
||||
+ caplog,
|
||||
+ master_alive_interval,
|
||||
+):
|
||||
+ """
|
||||
+ Verify a minion configured with multimaster hot/hot will pick up a master's
|
||||
+ dns change if it's been disconnected.
|
||||
+ """
|
||||
+
|
||||
+ if not salt_mm_master_1.ip_addr_set:
|
||||
+ pytest.skip("Unable to set additional IP address for master1")
|
||||
+
|
||||
+ etc_hosts.write_text(
|
||||
+ f"{etc_hosts.orig_text}\n172.16.0.1 master1.local master2.local"
|
||||
+ )
|
||||
+
|
||||
+ log.info("Added hosts record for master1.local and master2.local")
|
||||
+
|
||||
+ with salt_mm_minion_1.started(start_timeout=180):
|
||||
+ with caplog.at_level(logging.INFO):
|
||||
+ ret = mm_master_1_salt_cli.run("test.ping", minion_tgt="mm-minion-1")
|
||||
+ assert ret.returncode == 0
|
||||
+ etc_hosts.write_text(
|
||||
+ f"{etc_hosts.orig_text}\n127.0.0.1 master1.local master2.local"
|
||||
+ )
|
||||
+ log.info("Changed hosts record for master1.local and master2.local")
|
||||
+ subprocess.check_output(["ip", "addr", "del", "172.16.0.1/32", "dev", "lo"])
|
||||
+ log.info("Removed secondary master IP address.")
|
||||
+ # Wait for the minion's master_alive_interval, adding a second for
|
||||
+ # reliablity.
|
||||
+ time.sleep(master_alive_interval + 1)
|
||||
+ assert (
|
||||
+ "Master ip address changed from 172.16.0.1 to 127.0.0.1" in caplog.text
|
||||
+ )
|
||||
+ ret = mm_master_1_salt_cli.run("test.ping", minion_tgt="mm-minion-1")
|
||||
+ assert ret.returncode == 0
|
||||
+ assert (
|
||||
+ "Master ip address changed from 172.16.0.1 to 127.0.0.1" in caplog.text
|
||||
+ )
|
||||
diff --git a/tests/pytests/scenarios/dns/test_dns.py b/tests/pytests/scenarios/dns/test_dns.py
|
||||
new file mode 100644
|
||||
index 0000000000..cd33f0e7f0
|
||||
--- /dev/null
|
||||
+++ b/tests/pytests/scenarios/dns/test_dns.py
|
||||
@@ -0,0 +1,37 @@
|
||||
+import logging
|
||||
+import subprocess
|
||||
+import time
|
||||
+
|
||||
+import pytest
|
||||
+
|
||||
+log = logging.getLogger(__name__)
|
||||
+
|
||||
+
|
||||
+@pytest.mark.skip_unless_on_linux
|
||||
+@pytest.mark.skip_if_not_root
|
||||
+def test_dns_change(master, minion, salt_cli, etc_hosts, caplog, master_alive_interval):
|
||||
+ """
|
||||
+ Verify a minion will pick up a master's dns change if it's been disconnected.
|
||||
+ """
|
||||
+
|
||||
+ if not master.ip_addr_set:
|
||||
+ pytest.skip("Unable to set additional IP address for master")
|
||||
+
|
||||
+ etc_hosts.write_text(f"{etc_hosts.orig_text}\n172.16.0.1 master.local")
|
||||
+
|
||||
+ with minion.started(start_timeout=180):
|
||||
+ with caplog.at_level(logging.INFO):
|
||||
+ ret = salt_cli.run("test.ping", minion_tgt="minion")
|
||||
+ assert ret.returncode == 0
|
||||
+ etc_hosts.write_text(f"{etc_hosts.orig_text}\n127.0.0.1 master.local")
|
||||
+ log.info("Changed hosts record for master1.local and master2.local")
|
||||
+ subprocess.check_output(["ip", "addr", "del", "172.16.0.1/32", "dev", "lo"])
|
||||
+ log.info("Removed secondary master IP address.")
|
||||
+ # Wait for the minion's master_alive_interval, adding a second for
|
||||
+ # reliablity.
|
||||
+ time.sleep(master_alive_interval + 1)
|
||||
+ assert (
|
||||
+ "Master ip address changed from 172.16.0.1 to 127.0.0.1" in caplog.text
|
||||
+ )
|
||||
+ ret = salt_cli.run("test.ping", minion_tgt="minion")
|
||||
+ assert ret.returncode == 0
|
||||
diff --git a/tests/pytests/scenarios/failover/multimaster/test_failover_master.py b/tests/pytests/scenarios/failover/multimaster/test_failover_master.py
|
||||
index 9f6251a4d6..ebb2899ff0 100644
|
||||
--- a/tests/pytests/scenarios/failover/multimaster/test_failover_master.py
|
||||
+++ b/tests/pytests/scenarios/failover/multimaster/test_failover_master.py
|
||||
@@ -162,10 +162,6 @@ def test_minions_alive_with_no_master(
|
||||
"""
|
||||
Make sure the minions stay alive after all masters have stopped.
|
||||
"""
|
||||
- if grains["os_family"] == "Debian" and grains["osmajorrelease"] == 9:
|
||||
- pytest.skip(
|
||||
- "Skipping on Debian 9 until flaky issues resolved. See issue #61749"
|
||||
- )
|
||||
start_time = time.time()
|
||||
with salt_mm_failover_master_1.stopped():
|
||||
with salt_mm_failover_master_2.stopped():
|
||||
diff --git a/tests/pytests/unit/test_minion.py b/tests/pytests/unit/test_minion.py
|
||||
index a9e91742a2..017c28d163 100644
|
||||
--- a/tests/pytests/unit/test_minion.py
|
||||
+++ b/tests/pytests/unit/test_minion.py
|
||||
@@ -884,6 +884,8 @@ async def test_master_type_failover(minion_opts):
|
||||
assert opts["master"] == "master2"
|
||||
return MockPubChannel()
|
||||
|
||||
+ minion_opts["master_tries"] = 1
|
||||
+
|
||||
with patch("salt.minion.resolve_dns", mock_resolve_dns), patch(
|
||||
"salt.channel.client.AsyncPubChannel.factory", mock_channel_factory
|
||||
), patch("salt.loader.grains", MagicMock(return_value=[])):
|
||||
--
|
||||
2.47.0
|
||||
|
@ -1,13 +1,57 @@
|
||||
#!/bin/sh
|
||||
#!/bin/bash
|
||||
#
|
||||
# This script performs a cleanup of the prev_env Salt Bundle environment
|
||||
# that is created during a Salt Bundle upgrade.
|
||||
#
|
||||
# In case the script is triggered having WAIT_IF_SALT_JOBS env defined
|
||||
# then it will wait until there is no active Salt jobs running on the system
|
||||
# to perform the cleanup of the old environment.
|
||||
#
|
||||
# The script is triggered automatically via different ways:
|
||||
#
|
||||
# 1. As part of the Salt Highstate (via services.salt-minion).
|
||||
# 2. By venv-salt-minion-postinstall.service, 60 seconds after a Salt Bundle upgrade.
|
||||
# 3. By spec file during package installation/upgrade/removal.
|
||||
#
|
||||
# Expected execution behavior:
|
||||
#
|
||||
# - If there is no old environment detected -> Do nothing.
|
||||
# - If an old environment is found:
|
||||
# a) If WAIT_IF_SALT_JOBS env is NOT present -> Perform the cleanup.
|
||||
# b) If WAIT_IF_SALT_JOBS env is present -> Wait until there are no
|
||||
# other Salt jobs running to perform the cleanup.
|
||||
#
|
||||
# NOTE: In transactional systems, nothing is done when not running in a transaction.
|
||||
#
|
||||
# CAUTION: Do not call this script within a Salt execution and passing
|
||||
# the WAIT_IF_SALT_JOBS env, as it will produce the script to
|
||||
# go into an infinite loop.
|
||||
#
|
||||
|
||||
if builtin type -P transactional-update &> /dev/null && [ "$TRANSACTIONAL_UPDATE" != "true" ]; then
|
||||
# Do not clean previous environment on T-U systems when not in transaction
|
||||
exit 0
|
||||
fi
|
||||
|
||||
if [ -d "VENV_LOCATION_PLACEHOLDER/lib/prev_env" ] || [ -L "VENV_LOCATION_PLACEHOLDER/lib/PLACEHOLDER" ]; then
|
||||
while [ "${WAIT_IF_SALT_JOBS}" ]; do
|
||||
echo "Checking for active Salt jobs..."
|
||||
JOBS_RUNNING=$(venv-salt-call --local --output=json saltutil.running | grep '"jid":' | wc -l)
|
||||
if [ $JOBS_RUNNING -gt 0 ]; then
|
||||
echo "Found active Salt jobs! Waiting 10 seconds..."
|
||||
sleep 10
|
||||
else
|
||||
echo "No active Salt Jobs! Cleaning up the old environment..."
|
||||
break
|
||||
fi
|
||||
done
|
||||
fi
|
||||
|
||||
if [ -d "VENV_LOCATION_PLACEHOLDER/lib/prev_env" ]; then
|
||||
rm -r "VENV_LOCATION_PLACEHOLDER/lib/prev_env" || :
|
||||
echo "Removing VENV_LOCATION_PLACEHOLDER/lib/prev_env"
|
||||
rm -r "VENV_LOCATION_PLACEHOLDER/lib/prev_env" || :
|
||||
fi
|
||||
if [ -L "VENV_LOCATION_PLACEHOLDER/lib/PLACEHOLDER" ]; then
|
||||
unlink "VENV_LOCATION_PLACEHOLDER/lib/PLACEHOLDER" || :
|
||||
echo "Unlinking VENV_LOCATION_PLACEHOLDER/lib/PLACEHOLDER"
|
||||
unlink "VENV_LOCATION_PLACEHOLDER/lib/PLACEHOLDER" || :
|
||||
fi
|
||||
|
65
revert-setting-selinux-context-for-minion-service-bs.patch
Normal file
65
revert-setting-selinux-context-for-minion-service-bs.patch
Normal file
@ -0,0 +1,65 @@
|
||||
From 7217fccdce5df73fda1a8378dc7120642ffe8181 Mon Sep 17 00:00:00 2001
|
||||
From: Victor Zhestkov <vzhestkov@suse.com>
|
||||
Date: Mon, 16 Dec 2024 15:31:26 +0300
|
||||
Subject: [PATCH] Revert setting SELinux context for minion service
|
||||
(bsc#1233667)
|
||||
|
||||
This reverts commit d933c8f0795fdada84a01a2cc754586fa720993d.
|
||||
---
|
||||
pkg/common/salt-minion.service | 1 -
|
||||
pkg/old/deb/salt-minion.service | 1 -
|
||||
pkg/old/suse/salt-minion.service | 1 -
|
||||
pkg/old/suse/salt-minion.service.rhel7 | 1 -
|
||||
4 files changed, 4 deletions(-)
|
||||
|
||||
diff --git a/pkg/common/salt-minion.service b/pkg/common/salt-minion.service
|
||||
index 696d0263c3..69aff18c58 100644
|
||||
--- a/pkg/common/salt-minion.service
|
||||
+++ b/pkg/common/salt-minion.service
|
||||
@@ -9,7 +9,6 @@ Type=notify
|
||||
NotifyAccess=all
|
||||
LimitNOFILE=8192
|
||||
ExecStart=/usr/bin/salt-minion
|
||||
-SELinuxContext=system_u:system_r:unconfined_t:s0
|
||||
|
||||
[Install]
|
||||
WantedBy=multi-user.target
|
||||
diff --git a/pkg/old/deb/salt-minion.service b/pkg/old/deb/salt-minion.service
|
||||
index b0ad82c133..7e6cf14654 100644
|
||||
--- a/pkg/old/deb/salt-minion.service
|
||||
+++ b/pkg/old/deb/salt-minion.service
|
||||
@@ -8,7 +8,6 @@ KillMode=process
|
||||
NotifyAccess=all
|
||||
LimitNOFILE=8192
|
||||
ExecStart=/usr/bin/salt-minion
|
||||
-SELinuxContext=system_u:system_r:unconfined_t:s0
|
||||
|
||||
[Install]
|
||||
WantedBy=multi-user.target
|
||||
diff --git a/pkg/old/suse/salt-minion.service b/pkg/old/suse/salt-minion.service
|
||||
index b99ef06352..12f28314cb 100644
|
||||
--- a/pkg/old/suse/salt-minion.service
|
||||
+++ b/pkg/old/suse/salt-minion.service
|
||||
@@ -10,7 +10,6 @@ ExecStart=/usr/bin/salt-minion
|
||||
KillMode=process
|
||||
Restart=on-failure
|
||||
RestartSec=15
|
||||
-SELinuxContext=system_u:system_r:unconfined_t:s0
|
||||
|
||||
[Install]
|
||||
WantedBy=multi-user.target
|
||||
diff --git a/pkg/old/suse/salt-minion.service.rhel7 b/pkg/old/suse/salt-minion.service.rhel7
|
||||
index 92cc66d32f..6917267714 100644
|
||||
--- a/pkg/old/suse/salt-minion.service.rhel7
|
||||
+++ b/pkg/old/suse/salt-minion.service.rhel7
|
||||
@@ -9,7 +9,6 @@ ExecStart=/usr/bin/salt-minion
|
||||
KillMode=process
|
||||
Restart=on-failure
|
||||
RestartSec=15
|
||||
-SELinuxContext=system_u:system_r:unconfined_t:s0
|
||||
|
||||
[Install]
|
||||
WantedBy=multi-user.target
|
||||
--
|
||||
2.47.1
|
||||
|
@ -1,3 +1,63 @@
|
||||
-------------------------------------------------------------------
|
||||
Wed Jan 8 08:31:57 UTC 2025 - Victor Zhestkov <vzhestkov@suse.com>
|
||||
|
||||
- Include D-Bus python module for SUSE distros (bsc#1231618)
|
||||
|
||||
- Modified:
|
||||
* include-rpm
|
||||
* filter-requires.sh
|
||||
|
||||
-------------------------------------------------------------------
|
||||
Tue Dec 17 08:20:43 UTC 2024 - Victor Zhestkov <vzhestkov@suse.com>
|
||||
|
||||
- Revert setting SELinux context for minion service (bsc#1233667)
|
||||
|
||||
- Added:
|
||||
* revert-setting-selinux-context-for-minion-service-bs.patch
|
||||
|
||||
-------------------------------------------------------------------
|
||||
Tue Nov 26 14:53:34 UTC 2024 - Victor Zhestkov <vzhestkov@suse.com>
|
||||
|
||||
- Make minion reconnecting on changing master IP (bsc#1228182)
|
||||
|
||||
- Added:
|
||||
* make-minion-reconnecting-on-changing-master-ip-bsc-1.patch
|
||||
|
||||
-------------------------------------------------------------------
|
||||
Tue Nov 26 14:53:17 UTC 2024 - Marek Czernek <marek.czernek@suse.com>
|
||||
|
||||
- Fix post_start_cleanup.sh shebang to work on all systems (gh#uyuni-project/uyuni#9477)
|
||||
|
||||
- Modified:
|
||||
* post_start_cleanup.sh
|
||||
|
||||
-------------------------------------------------------------------
|
||||
Wed Oct 23 15:34:36 UTC 2024 - Marek Czernek <marek.czernek@suse.com>
|
||||
|
||||
- Handle logger exception when flushing already closed file
|
||||
|
||||
- Added:
|
||||
* handle-logger-flushing-already-closed-file-686.patch
|
||||
|
||||
-------------------------------------------------------------------
|
||||
Wed Oct 23 15:25:09 UTC 2024 - Pablo Suárez Hernández <psuarezhernandez@suse.com>
|
||||
|
||||
- Make Salt Bundle more tolerant to long running jobs (bsc#1228690)
|
||||
|
||||
- Added:
|
||||
* enhance-cleanup-mechanism-after-salt-bundle-upgrade-.patch
|
||||
|
||||
- Modified:
|
||||
* post_start_cleanup.sh
|
||||
|
||||
-------------------------------------------------------------------
|
||||
Wed Oct 23 15:23:21 UTC 2024 - Marek Czernek <marek.czernek@suse.com>
|
||||
|
||||
- Fix additional x509 tests and test_suse tests for SLE12
|
||||
|
||||
- Added:
|
||||
* fix-x509-private-key-tests-and-test_suse-on-sle12-68.patch
|
||||
|
||||
-------------------------------------------------------------------
|
||||
Wed Oct 23 14:46:45 UTC 2024 - Pablo Suárez Hernández <psuarezhernandez@suse.com>
|
||||
|
||||
|
@ -452,8 +452,20 @@ Patch140: revert-the-change-making-reactor-less-blocking-bsc-1.patch
|
||||
Patch141: allow-namedloadercontexts-to-be-returned-from-loader.patch
|
||||
# PATCH-FIX_UPSTREAM: https://github.com/saltstack/salt/commit/d54407ba6dc664e5e5f3f613e27ae24f828c9648
|
||||
Patch142: avoid-explicit-reading-of-etc-salt-minion-bsc-122035.patch
|
||||
# PATCH-FIX_UPSTREAM: https://github.com/openSUSE/salt/pull/682
|
||||
# PATCH-FIX_OPENSUSE: https://github.com/openSUSE/salt/pull/682
|
||||
Patch143: fix-x509-test-fails-on-old-openssl-systems-682.patch
|
||||
# PATCH-FIX_OPENSUSE: https://github.com/openSUSE/salt/pull/684
|
||||
Patch144: fix-x509-private-key-tests-and-test_suse-on-sle12-68.patch
|
||||
# PATCH-FIX_OPENSUSE: https://github.com/openSUSE/salt/pull/685
|
||||
Patch145: enhance-cleanup-mechanism-after-salt-bundle-upgrade-.patch
|
||||
# PATCH-FIX_UPSTREAM: https://github.com/saltstack/salt/commit/9683260d61668da8559ecde6caf63a52fedd8790
|
||||
Patch146: handle-logger-flushing-already-closed-file-686.patch
|
||||
# PATCH-FIX_UPSTREAM: https://github.com/saltstack/salt/pull/66422
|
||||
# PATCH-FIX_UPSTREAM: https://github.com/saltstack/salt/pull/66757
|
||||
# PATCH-FIX_UPSTREAM: https://github.com/saltstack/salt/pull/66760
|
||||
Patch147: make-minion-reconnecting-on-changing-master-ip-bsc-1.patch
|
||||
# PATCH-FIX_OPENSUSE: https://github.com/openSUSE/salt/pull/690
|
||||
Patch148: revert-setting-selinux-context-for-minion-service-bs.patch
|
||||
|
||||
### IMPORTANT: The line below is used as a snippet marker. Do not touch it.
|
||||
### SALT PATCHES LIST END
|
||||
@ -532,6 +544,7 @@ BuildRequires: saltbundlepy-apt
|
||||
# required for zypper.py
|
||||
%if 0%{?suse_version} != 1110
|
||||
Requires(pre): libzypp(plugin:system) >= 0
|
||||
BuildRequires: saltbundlepy-dbus-python
|
||||
%endif
|
||||
BuildRequires: saltbundlepy-zypp-plugin
|
||||
# required for virtualisation hosts (libvirt)
|
||||
|
Loading…
x
Reference in New Issue
Block a user