salt/fix-salt-warnings-and-testuite-for-python-3.11-635.patch

3861 lines
151 KiB
Diff
Raw Blame History

This file contains ambiguous Unicode characters

This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.

From cdb7211920c9256942518fbcf3bd627a70a99855 Mon Sep 17 00:00:00 2001
From: Victor Zhestkov <vzhestkov@suse.com>
Date: Mon, 18 Mar 2024 09:15:08 +0100
Subject: [PATCH] Fix Salt warnings and testuite for Python 3.11 (#635)
MIME-Version: 1.0
Content-Type: text/plain; charset=UTF-8
Content-Transfer-Encoding: 8bit
* Switch to `FullArgSpec` since Py 3.11 no longer has `ArgSpec`, deprecated since Py 3.0
Signed-off-by: Pedro Algarvio <palgarvio@vmware.com>
* Backport `locale.getdefaultlocale()` into Salt. It's getting removed in Py 3.13
Signed-off-by: Pedro Algarvio <palgarvio@vmware.com>
* Stop using the deprecated `pipes` module
Signed-off-by: Pedro Algarvio <palgarvio@vmware.com>
* Stop using the deprecated `cgi` module.
Signed-off-by: Pedro Algarvio <palgarvio@vmware.com>
* Add `__getstate__` to blacklisted methods, present in Py 3.11
Signed-off-by: Pedro Algarvio <palgarvio@vmware.com>
* Fix test_state test
* Use proper keys since Python's base64 in Py3.11 is more picky
```
artifacts/salt/bin/python3
Python 3.10.11 (main, May 5 2023, 02:31:54) [GCC 11.2.0] on linux
Type "help", "copyright", "credits" or "license" for more information.
>>> import base64
>>> base64.b64decode("AAAAB3NzaC1kcQ9J5bYTEyZ==", validate=True)
b'\x00\x00\x00\x07ssh-dq\x0fI\xe5\xb6\x13\x13&'
```
```
$ artifacts/salt/bin/python3
Python 3.11.3 (main, May 5 2023, 02:31:40) [GCC 11.2.0] on linux
Type "help", "copyright", "credits" or "license" for more information.
>>> import base64
>>> base64.b64decode("AAAAB3NzaC1kcQ9J5bYTEyZ==", validate=True)
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/tmp/testing/artifacts/salt/lib/python3.11/base64.py", line 88, in b64decode
return binascii.a2b_base64(s, strict_mode=validate)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
binascii.Error: Excess data after padding
```
Signed-off-by: Pedro Algarvio <palgarvio@vmware.com>
---------
Signed-off-by: Pedro Algarvio <palgarvio@vmware.com>
Co-authored-by: Pedro Algarvio <palgarvio@vmware.com>
Co-authored-by: Marek Czernek <marek.czernek@suse.com>
---
salt/__init__.py | 43 ++-
salt/grains/core.py | 108 +++----
salt/modules/container_resource.py | 74 ++---
salt/modules/deb_postgres.py | 16 +-
salt/modules/dockermod.py | 178 +++++------
salt/modules/lxc.py | 50 +--
salt/modules/mac_keychain.py | 32 +-
salt/modules/macpackage.py | 45 +--
salt/modules/openstack_config.py | 41 +--
salt/modules/postgres.py | 116 +++----
salt/utils/cloud.py | 300 ++++++++----------
salt/utils/http.py | 48 ++-
salt/utils/jinja.py | 25 +-
salt/utils/locales.py | 39 ++-
tests/integration/states/test_ssh_auth.py | 50 ++-
.../pytests/unit/modules/state/test_state.py | 2 +-
tests/unit/states/test_module.py | 56 ++--
tests/unit/test_master.py | 8 +-
18 files changed, 586 insertions(+), 645 deletions(-)
diff --git a/salt/__init__.py b/salt/__init__.py
index e06b8ad7127..b5fe3677c22 100644
--- a/salt/__init__.py
+++ b/salt/__init__.py
@@ -72,6 +72,44 @@ warnings.filterwarnings(
)
+def __getdefaultlocale(envvars=("LC_ALL", "LC_CTYPE", "LANG", "LANGUAGE")):
+ """
+ This function was backported from Py3.11 which started triggering a
+ deprecation warning about it's removal in 3.13.
+ """
+ import locale
+
+ try:
+ # check if it's supported by the _locale module
+ import _locale
+
+ code, encoding = _locale._getdefaultlocale()
+ except (ImportError, AttributeError):
+ pass
+ else:
+ # make sure the code/encoding values are valid
+ if sys.platform == "win32" and code and code[:2] == "0x":
+ # map windows language identifier to language name
+ code = locale.windows_locale.get(int(code, 0))
+ # ...add other platform-specific processing here, if
+ # necessary...
+ return code, encoding
+
+ # fall back on POSIX behaviour
+ import os
+
+ lookup = os.environ.get
+ for variable in envvars:
+ localename = lookup(variable, None)
+ if localename:
+ if variable == "LANGUAGE":
+ localename = localename.split(":")[0]
+ break
+ else:
+ localename = "C"
+ return locale._parse_localename(localename)
+
+
def __define_global_system_encoding_variable__():
import sys
@@ -90,17 +128,14 @@ def __define_global_system_encoding_variable__():
# If the system is properly configured this should return a valid
# encoding. MS Windows has problems with this and reports the wrong
# encoding
- import locale
try:
- encoding = locale.getdefaultlocale()[-1]
+ encoding = __getdefaultlocale()[-1]
except ValueError:
# A bad locale setting was most likely found:
# https://github.com/saltstack/salt/issues/26063
pass
- # This is now garbage collectable
- del locale
if not encoding:
# This is most likely ascii which is not the best but we were
# unable to find a better encoding. If this fails, we fall all
diff --git a/salt/grains/core.py b/salt/grains/core.py
index 5c125563461..4454c303fed 100644
--- a/salt/grains/core.py
+++ b/salt/grains/core.py
@@ -11,7 +11,6 @@ as those returned here
import datetime
import hashlib
-import locale
import logging
import os
import platform
@@ -34,6 +33,7 @@ import salt.modules.smbios
import salt.utils.args
import salt.utils.dns
import salt.utils.files
+import salt.utils.locales
import salt.utils.network
import salt.utils.path
import salt.utils.pkg.rpm
@@ -290,7 +290,7 @@ def _linux_gpu_data():
devs = []
try:
- lspci_out = __salt__["cmd.run"]("{} -vmm".format(lspci))
+ lspci_out = __salt__["cmd.run"](f"{lspci} -vmm")
cur_dev = {}
error = False
@@ -364,7 +364,7 @@ def _netbsd_gpu_data():
for line in pcictl_out.splitlines():
for vendor in known_vendors:
vendor_match = re.match(
- r"[0-9:]+ ({}) (.+) \(VGA .+\)".format(vendor), line, re.IGNORECASE
+ rf"[0-9:]+ ({vendor}) (.+) \(VGA .+\)", line, re.IGNORECASE
)
if vendor_match:
gpus.append(
@@ -426,18 +426,18 @@ def _bsd_cpudata(osdata):
if sysctl:
cmds.update(
{
- "num_cpus": "{} -n hw.ncpu".format(sysctl),
- "cpuarch": "{} -n hw.machine".format(sysctl),
- "cpu_model": "{} -n hw.model".format(sysctl),
+ "num_cpus": f"{sysctl} -n hw.ncpu",
+ "cpuarch": f"{sysctl} -n hw.machine",
+ "cpu_model": f"{sysctl} -n hw.model",
}
)
if arch and osdata["kernel"] == "OpenBSD":
- cmds["cpuarch"] = "{} -s".format(arch)
+ cmds["cpuarch"] = f"{arch} -s"
if osdata["kernel"] == "Darwin":
- cmds["cpu_model"] = "{} -n machdep.cpu.brand_string".format(sysctl)
- cmds["cpu_flags"] = "{} -n machdep.cpu.features".format(sysctl)
+ cmds["cpu_model"] = f"{sysctl} -n machdep.cpu.brand_string"
+ cmds["cpu_flags"] = f"{sysctl} -n machdep.cpu.features"
grains = {k: __salt__["cmd.run"](v) for k, v in cmds.items()}
@@ -522,7 +522,7 @@ def _aix_cpudata():
grains = {}
cmd = salt.utils.path.which("prtconf")
if cmd:
- data = __salt__["cmd.run"]("{}".format(cmd)) + os.linesep
+ data = __salt__["cmd.run"](f"{cmd}") + os.linesep
for dest, regstring in (
("cpuarch", r"(?im)^\s*Processor\s+Type:\s+(\S+)"),
("cpu_flags", r"(?im)^\s*Processor\s+Version:\s+(\S+)"),
@@ -568,9 +568,9 @@ def _osx_memdata():
sysctl = salt.utils.path.which("sysctl")
if sysctl:
- mem = __salt__["cmd.run"]("{} -n hw.memsize".format(sysctl))
+ mem = __salt__["cmd.run"](f"{sysctl} -n hw.memsize")
swap_total = (
- __salt__["cmd.run"]("{} -n vm.swapusage".format(sysctl))
+ __salt__["cmd.run"](f"{sysctl} -n vm.swapusage")
.split()[2]
.replace(",", ".")
)
@@ -595,20 +595,20 @@ def _bsd_memdata(osdata):
sysctl = salt.utils.path.which("sysctl")
if sysctl:
- mem = __salt__["cmd.run"]("{} -n hw.physmem".format(sysctl))
+ mem = __salt__["cmd.run"](f"{sysctl} -n hw.physmem")
if osdata["kernel"] == "NetBSD" and mem.startswith("-"):
- mem = __salt__["cmd.run"]("{} -n hw.physmem64".format(sysctl))
+ mem = __salt__["cmd.run"](f"{sysctl} -n hw.physmem64")
grains["mem_total"] = int(mem) // 1024 // 1024
if osdata["kernel"] in ["OpenBSD", "NetBSD"]:
swapctl = salt.utils.path.which("swapctl")
- swap_data = __salt__["cmd.run"]("{} -sk".format(swapctl))
+ swap_data = __salt__["cmd.run"](f"{swapctl} -sk")
if swap_data == "no swap devices configured":
swap_total = 0
else:
swap_total = swap_data.split(" ")[1]
else:
- swap_total = __salt__["cmd.run"]("{} -n vm.swap_total".format(sysctl))
+ swap_total = __salt__["cmd.run"](f"{sysctl} -n vm.swap_total")
grains["swap_total"] = int(swap_total) // 1024 // 1024
return grains
@@ -626,7 +626,7 @@ def _sunos_memdata():
grains["mem_total"] = int(comps[2].strip())
swap_cmd = salt.utils.path.which("swap")
- swap_data = __salt__["cmd.run"]("{} -s".format(swap_cmd)).split()
+ swap_data = __salt__["cmd.run"](f"{swap_cmd} -s").split()
try:
swap_avail = int(swap_data[-2][:-1])
swap_used = int(swap_data[-4][:-1])
@@ -654,7 +654,7 @@ def _aix_memdata():
swap_cmd = salt.utils.path.which("swap")
if swap_cmd:
- swap_data = __salt__["cmd.run"]("{} -s".format(swap_cmd)).split()
+ swap_data = __salt__["cmd.run"](f"{swap_cmd} -s").split()
try:
swap_total = (int(swap_data[-2]) + int(swap_data[-6])) * 4
except ValueError:
@@ -707,7 +707,7 @@ def _aix_get_machine_id():
grains = {}
cmd = salt.utils.path.which("lsattr")
if cmd:
- data = __salt__["cmd.run"]("{} -El sys0".format(cmd)) + os.linesep
+ data = __salt__["cmd.run"](f"{cmd} -El sys0") + os.linesep
uuid_regexes = [re.compile(r"(?im)^\s*os_uuid\s+(\S+)\s+(.*)")]
for regex in uuid_regexes:
res = regex.search(data)
@@ -1018,7 +1018,7 @@ def _virtual(osdata):
subtype_cmd = "{} -c current get -H -o value {}-role".format(
command, role
)
- ret = __salt__["cmd.run"]("{}".format(subtype_cmd))
+ ret = __salt__["cmd.run"](f"{subtype_cmd}")
if ret == "true":
roles.append(role)
if roles:
@@ -1164,14 +1164,14 @@ def _virtual(osdata):
elif osdata["kernel"] == "FreeBSD":
kenv = salt.utils.path.which("kenv")
if kenv:
- product = __salt__["cmd.run"]("{} smbios.system.product".format(kenv))
- maker = __salt__["cmd.run"]("{} smbios.system.maker".format(kenv))
+ product = __salt__["cmd.run"](f"{kenv} smbios.system.product")
+ maker = __salt__["cmd.run"](f"{kenv} smbios.system.maker")
if product.startswith("VMware"):
grains["virtual"] = "VMware"
if product.startswith("VirtualBox"):
grains["virtual"] = "VirtualBox"
if maker.startswith("Xen"):
- grains["virtual_subtype"] = "{} {}".format(maker, product)
+ grains["virtual_subtype"] = f"{maker} {product}"
grains["virtual"] = "xen"
if maker.startswith("Microsoft") and product.startswith("Virtual"):
grains["virtual"] = "VirtualPC"
@@ -1182,9 +1182,9 @@ def _virtual(osdata):
if maker.startswith("Amazon EC2"):
grains["virtual"] = "Nitro"
if sysctl:
- hv_vendor = __salt__["cmd.run"]("{} -n hw.hv_vendor".format(sysctl))
- model = __salt__["cmd.run"]("{} -n hw.model".format(sysctl))
- jail = __salt__["cmd.run"]("{} -n security.jail.jailed".format(sysctl))
+ hv_vendor = __salt__["cmd.run"](f"{sysctl} -n hw.hv_vendor")
+ model = __salt__["cmd.run"](f"{sysctl} -n hw.model")
+ jail = __salt__["cmd.run"](f"{sysctl} -n security.jail.jailed")
if "bhyve" in hv_vendor:
grains["virtual"] = "bhyve"
elif "QEMU Virtual CPU" in model:
@@ -1200,22 +1200,19 @@ def _virtual(osdata):
elif osdata["kernel"] == "NetBSD":
if sysctl:
if "QEMU Virtual CPU" in __salt__["cmd.run"](
- "{} -n machdep.cpu_brand".format(sysctl)
+ f"{sysctl} -n machdep.cpu_brand"
):
grains["virtual"] = "kvm"
elif "invalid" not in __salt__["cmd.run"](
- "{} -n machdep.xen.suspend".format(sysctl)
+ f"{sysctl} -n machdep.xen.suspend"
):
grains["virtual"] = "Xen PV DomU"
elif "VMware" in __salt__["cmd.run"](
- "{} -n machdep.dmi.system-vendor".format(sysctl)
+ f"{sysctl} -n machdep.dmi.system-vendor"
):
grains["virtual"] = "VMware"
# NetBSD has Xen dom0 support
- elif (
- __salt__["cmd.run"]("{} -n machdep.idle-mechanism".format(sysctl))
- == "xen"
- ):
+ elif __salt__["cmd.run"](f"{sysctl} -n machdep.idle-mechanism") == "xen":
if os.path.isfile("/var/run/xenconsoled.pid"):
grains["virtual_subtype"] = "Xen Dom0"
elif osdata["kernel"] == "SunOS":
@@ -1223,7 +1220,7 @@ def _virtual(osdata):
# check the zonename here as fallback
zonename = salt.utils.path.which("zonename")
if zonename:
- zone = __salt__["cmd.run"]("{}".format(zonename))
+ zone = __salt__["cmd.run"](f"{zonename}")
if zone != "global":
grains["virtual"] = "zone"
@@ -1252,7 +1249,7 @@ def _virtual(osdata):
r".*Product Name: ([^\r\n]*).*", output, flags=re.DOTALL
)
if product:
- grains["virtual_subtype"] = "Amazon EC2 ({})".format(product[1])
+ grains["virtual_subtype"] = f"Amazon EC2 ({product[1]})"
elif re.match(r".*Version: [^\r\n]+\.amazon.*", output, flags=re.DOTALL):
grains["virtual_subtype"] = "Amazon EC2"
@@ -1284,9 +1281,7 @@ def _virtual_hv(osdata):
try:
version = {}
for fn in ("major", "minor", "extra"):
- with salt.utils.files.fopen(
- "/sys/hypervisor/version/{}".format(fn), "r"
- ) as fhr:
+ with salt.utils.files.fopen(f"/sys/hypervisor/version/{fn}", "r") as fhr:
version[fn] = salt.utils.stringutils.to_unicode(fhr.read().strip())
grains["virtual_hv_version"] = "{}.{}{}".format(
version["major"], version["minor"], version["extra"]
@@ -1442,7 +1437,7 @@ def _windows_os_release_grain(caption, product_type):
# ie: R2
if re.match(r"^R\d+$", item):
release = item
- os_release = "{}Server{}".format(version, release)
+ os_release = f"{version}Server{release}"
else:
for item in caption.split(" "):
# If it's a number, decimal number, Thin or Vista, then it's the
@@ -1633,7 +1628,7 @@ def _linux_devicetree_platform_data():
try:
# /proc/device-tree should be used instead of /sys/firmware/devicetree/base
# see https://github.com/torvalds/linux/blob/v5.13/Documentation/ABI/testing/sysfs-firmware-ofw#L14
- loc = "/proc/device-tree/{}".format(path)
+ loc = f"/proc/device-tree/{path}"
if os.path.isfile(loc):
with salt.utils.files.fopen(loc, mode="r") as f:
return f.read().rstrip("\x00") # all strings are null-terminated
@@ -1872,18 +1867,13 @@ def _linux_bin_exists(binary):
"""
for search_cmd in ("which", "type -ap"):
try:
- return __salt__["cmd.retcode"]("{} {}".format(search_cmd, binary)) == 0
+ return __salt__["cmd.retcode"](f"{search_cmd} {binary}") == 0
except salt.exceptions.CommandExecutionError:
pass
try:
return (
- len(
- __salt__["cmd.run_all"]("whereis -b {}".format(binary))[
- "stdout"
- ].split()
- )
- > 1
+ len(__salt__["cmd.run_all"](f"whereis -b {binary}")["stdout"].split()) > 1
)
except salt.exceptions.CommandExecutionError:
return False
@@ -1901,7 +1891,7 @@ def _parse_lsb_release():
pass
else:
# Adds lsb_distrib_{id,release,codename,description}
- ret["lsb_{}".format(key.lower())] = value.rstrip()
+ ret[f"lsb_{key.lower()}"] = value.rstrip()
except OSError as exc:
log.trace("Failed to parse /etc/lsb-release: %s", exc)
return ret
@@ -2634,7 +2624,7 @@ def os_data():
osbuild = __salt__["cmd.run"]("sw_vers -buildVersion")
grains["os"] = "MacOS"
grains["os_family"] = "MacOS"
- grains["osfullname"] = "{} {}".format(osname, osrelease)
+ grains["osfullname"] = f"{osname} {osrelease}"
grains["osrelease"] = osrelease
grains["osbuild"] = osbuild
grains["init"] = "launchd"
@@ -2708,7 +2698,7 @@ def locale_info():
(
grains["locale_info"]["defaultlanguage"],
grains["locale_info"]["defaultencoding"],
- ) = locale.getdefaultlocale()
+ ) = salt.utils.locales.getdefaultlocale()
except Exception: # pylint: disable=broad-except
# locale.getdefaultlocale can ValueError!! Catch anything else it
# might do, per #2205
@@ -3175,7 +3165,7 @@ def _hw_data(osdata):
"productname": "DeviceDesc",
}
for grain_name, cmd_key in hwdata.items():
- result = __salt__["cmd.run_all"]("fw_printenv {}".format(cmd_key))
+ result = __salt__["cmd.run_all"](f"fw_printenv {cmd_key}")
if result["retcode"] == 0:
uboot_keyval = result["stdout"].split("=")
grains[grain_name] = _clean_value(grain_name, uboot_keyval[1])
@@ -3195,7 +3185,7 @@ def _hw_data(osdata):
"uuid": "smbios.system.uuid",
}
for key, val in fbsd_hwdata.items():
- value = __salt__["cmd.run"]("{} {}".format(kenv, val))
+ value = __salt__["cmd.run"](f"{kenv} {val}")
grains[key] = _clean_value(key, value)
elif osdata["kernel"] == "OpenBSD":
sysctl = salt.utils.path.which("sysctl")
@@ -3207,7 +3197,7 @@ def _hw_data(osdata):
"uuid": "hw.uuid",
}
for key, oid in hwdata.items():
- value = __salt__["cmd.run"]("{} -n {}".format(sysctl, oid))
+ value = __salt__["cmd.run"](f"{sysctl} -n {oid}")
if not value.endswith(" value is not available"):
grains[key] = _clean_value(key, value)
elif osdata["kernel"] == "NetBSD":
@@ -3222,7 +3212,7 @@ def _hw_data(osdata):
"uuid": "machdep.dmi.system-uuid",
}
for key, oid in nbsd_hwdata.items():
- result = __salt__["cmd.run_all"]("{} -n {}".format(sysctl, oid))
+ result = __salt__["cmd.run_all"](f"{sysctl} -n {oid}")
if result["retcode"] == 0:
grains[key] = _clean_value(key, result["stdout"])
elif osdata["kernel"] == "Darwin":
@@ -3230,7 +3220,7 @@ def _hw_data(osdata):
sysctl = salt.utils.path.which("sysctl")
hwdata = {"productname": "hw.model"}
for key, oid in hwdata.items():
- value = __salt__["cmd.run"]("{} -b {}".format(sysctl, oid))
+ value = __salt__["cmd.run"](f"{sysctl} -b {oid}")
if not value.endswith(" is invalid"):
grains[key] = _clean_value(key, value)
elif osdata["kernel"] == "SunOS" and osdata["cpuarch"].startswith("sparc"):
@@ -3244,7 +3234,7 @@ def _hw_data(osdata):
("/usr/sbin/virtinfo", "-a"),
):
if salt.utils.path.which(cmd): # Also verifies that cmd is executable
- data += __salt__["cmd.run"]("{} {}".format(cmd, args))
+ data += __salt__["cmd.run"](f"{cmd} {args}")
data += "\n"
sn_regexes = [
@@ -3359,7 +3349,7 @@ def _hw_data(osdata):
elif osdata["kernel"] == "AIX":
cmd = salt.utils.path.which("prtconf")
if cmd:
- data = __salt__["cmd.run"]("{}".format(cmd)) + os.linesep
+ data = __salt__["cmd.run"](f"{cmd}") + os.linesep
for dest, regstring in (
("serialnumber", r"(?im)^\s*Machine\s+Serial\s+Number:\s+(\S+)"),
("systemfirmware", r"(?im)^\s*Firmware\s+Version:\s+(.*)"),
@@ -3480,14 +3470,14 @@ def default_gateway():
for line in out.splitlines():
if line.startswith("default"):
grains["ip_gw"] = True
- grains["ip{}_gw".format(ip_version)] = True
+ grains[f"ip{ip_version}_gw"] = True
try:
via, gw_ip = line.split()[1:3]
except ValueError:
pass
else:
if via == "via":
- grains["ip{}_gw".format(ip_version)] = gw_ip
+ grains[f"ip{ip_version}_gw"] = gw_ip
break
except Exception: # pylint: disable=broad-except
continue
diff --git a/salt/modules/container_resource.py b/salt/modules/container_resource.py
index a29cba2e468..ceec72a7b20 100644
--- a/salt/modules/container_resource.py
+++ b/salt/modules/container_resource.py
@@ -8,13 +8,11 @@ These functions are not designed to be called directly, but instead from the
:mod:`docker <salt.modules.docker>` execution modules. They provide for
common logic to be re-used for common actions.
"""
-
-
import copy
import functools
import logging
import os
-import pipes
+import shlex
import time
import traceback
@@ -68,14 +66,14 @@ def _nsenter(pid):
"""
Return the nsenter command to attach to the named container
"""
- return "nsenter --target {} --mount --uts --ipc --net --pid".format(pid)
+ return f"nsenter --target {pid} --mount --uts --ipc --net --pid"
def _get_md5(name, path, run_func):
"""
Get the MD5 checksum of a file from a container
"""
- output = run_func(name, "md5sum {}".format(pipes.quote(path)), ignore_retcode=True)[
+ output = run_func(name, f"md5sum {shlex.quote(path)}", ignore_retcode=True)[
"stdout"
]
try:
@@ -102,10 +100,10 @@ def cache_file(source):
if source.startswith("salt://"):
cached_source = __salt__["cp.cache_file"](source)
if not cached_source:
- raise CommandExecutionError("Unable to cache {}".format(source))
+ raise CommandExecutionError(f"Unable to cache {source}")
return cached_source
except AttributeError:
- raise SaltInvocationError("Invalid source file {}".format(source))
+ raise SaltInvocationError(f"Invalid source file {source}")
return source
@@ -164,55 +162,47 @@ def run(
if exec_driver == "lxc-attach":
full_cmd = "lxc-attach "
if path:
- full_cmd += "-P {} ".format(pipes.quote(path))
+ full_cmd += f"-P {shlex.quote(path)} "
if keep_env is not True:
full_cmd += "--clear-env "
if "PATH" not in to_keep:
- full_cmd += "--set-var {} ".format(PATH)
+ full_cmd += f"--set-var {PATH} "
# --clear-env results in a very restrictive PATH
# (/bin:/usr/bin), use a good fallback.
full_cmd += " ".join(
[
- "--set-var {}={}".format(x, pipes.quote(os.environ[x]))
+ f"--set-var {x}={shlex.quote(os.environ[x])}"
for x in to_keep
if x in os.environ
]
)
- full_cmd += " -n {} -- {}".format(pipes.quote(name), cmd)
+ full_cmd += f" -n {shlex.quote(name)} -- {cmd}"
elif exec_driver == "nsenter":
- pid = __salt__["{}.pid".format(container_type)](name)
- full_cmd = "nsenter --target {} --mount --uts --ipc --net --pid -- ".format(pid)
+ pid = __salt__[f"{container_type}.pid"](name)
+ full_cmd = f"nsenter --target {pid} --mount --uts --ipc --net --pid -- "
if keep_env is not True:
full_cmd += "env -i "
if "PATH" not in to_keep:
- full_cmd += "{} ".format(PATH)
+ full_cmd += f"{PATH} "
full_cmd += " ".join(
- [
- "{}={}".format(x, pipes.quote(os.environ[x]))
- for x in to_keep
- if x in os.environ
- ]
+ [f"{x}={shlex.quote(os.environ[x])}" for x in to_keep if x in os.environ]
)
- full_cmd += " {}".format(cmd)
+ full_cmd += f" {cmd}"
elif exec_driver == "docker-exec":
# We're using docker exec on the CLI as opposed to via docker-py, since
# the Docker API doesn't return stdout and stderr separately.
full_cmd = "docker exec "
if stdin:
full_cmd += "-i "
- full_cmd += "{} ".format(name)
+ full_cmd += f"{name} "
if keep_env is not True:
full_cmd += "env -i "
if "PATH" not in to_keep:
- full_cmd += "{} ".format(PATH)
+ full_cmd += f"{PATH} "
full_cmd += " ".join(
- [
- "{}={}".format(x, pipes.quote(os.environ[x]))
- for x in to_keep
- if x in os.environ
- ]
+ [f"{x}={shlex.quote(os.environ[x])}" for x in to_keep if x in os.environ]
)
- full_cmd += " {}".format(cmd)
+ full_cmd += f" {cmd}"
if not use_vt:
ret = __salt__[cmd_func](
@@ -299,13 +289,13 @@ def copy_to(
salt myminion container_resource.copy_to mycontainer /local/file/path /container/file/path container_type=docker exec_driver=nsenter
"""
# Get the appropriate functions
- state = __salt__["{}.state".format(container_type)]
+ state = __salt__[f"{container_type}.state"]
def run_all(*args, **akwargs):
akwargs = copy.deepcopy(akwargs)
if container_type in ["lxc"] and "path" not in akwargs:
akwargs["path"] = path
- return __salt__["{}.run_all".format(container_type)](*args, **akwargs)
+ return __salt__[f"{container_type}.run_all"](*args, **akwargs)
state_kwargs = {}
cmd_kwargs = {"ignore_retcode": True}
@@ -321,7 +311,7 @@ def copy_to(
c_state = _state(name)
if c_state != "running":
- raise CommandExecutionError("Container '{}' is not running".format(name))
+ raise CommandExecutionError(f"Container '{name}' is not running")
local_file = cache_file(source)
source_dir, source_name = os.path.split(local_file)
@@ -330,17 +320,14 @@ def copy_to(
if not os.path.isabs(local_file):
raise SaltInvocationError("Source path must be absolute")
elif not os.path.exists(local_file):
- raise SaltInvocationError("Source file {} does not exist".format(local_file))
+ raise SaltInvocationError(f"Source file {local_file} does not exist")
elif not os.path.isfile(local_file):
raise SaltInvocationError("Source must be a regular file")
# Destination file sanity checks
if not os.path.isabs(dest):
raise SaltInvocationError("Destination path must be absolute")
- if (
- run_all(name, "test -d {}".format(pipes.quote(dest)), **cmd_kwargs)["retcode"]
- == 0
- ):
+ if run_all(name, f"test -d {shlex.quote(dest)}", **cmd_kwargs)["retcode"] == 0:
# Destination is a directory, full path to dest file will include the
# basename of the source file.
dest = os.path.join(dest, source_name)
@@ -350,14 +337,12 @@ def copy_to(
# parent directory.
dest_dir, dest_name = os.path.split(dest)
if (
- run_all(name, "test -d {}".format(pipes.quote(dest_dir)), **cmd_kwargs)[
- "retcode"
- ]
+ run_all(name, f"test -d {shlex.quote(dest_dir)}", **cmd_kwargs)["retcode"]
!= 0
):
if makedirs:
result = run_all(
- name, "mkdir -p {}".format(pipes.quote(dest_dir)), **cmd_kwargs
+ name, f"mkdir -p {shlex.quote(dest_dir)}", **cmd_kwargs
)
if result["retcode"] != 0:
error = (
@@ -375,10 +360,7 @@ def copy_to(
)
if (
not overwrite
- and run_all(name, "test -e {}".format(pipes.quote(dest)), **cmd_kwargs)[
- "retcode"
- ]
- == 0
+ and run_all(name, f"test -e {shlex.quote(dest)}", **cmd_kwargs)["retcode"] == 0
):
raise CommandExecutionError(
"Destination path {} already exists. Use overwrite=True to "
@@ -401,14 +383,14 @@ def copy_to(
if exec_driver == "lxc-attach":
lxcattach = "lxc-attach"
if path:
- lxcattach += " -P {}".format(pipes.quote(path))
+ lxcattach += f" -P {shlex.quote(path)}"
copy_cmd = (
'cat "{0}" | {4} --clear-env --set-var {1} -n {2} -- tee "{3}"'.format(
local_file, PATH, name, dest, lxcattach
)
)
elif exec_driver == "nsenter":
- pid = __salt__["{}.pid".format(container_type)](name)
+ pid = __salt__[f"{container_type}.pid"](name)
copy_cmd = 'cat "{}" | {} env -i {} tee "{}"'.format(
local_file, _nsenter(pid), PATH, dest
)
diff --git a/salt/modules/deb_postgres.py b/salt/modules/deb_postgres.py
index 3ecd4a8ba49..d92859562d4 100644
--- a/salt/modules/deb_postgres.py
+++ b/salt/modules/deb_postgres.py
@@ -2,10 +2,8 @@
Module to provide Postgres compatibility to salt for debian family specific tools.
"""
-
-
import logging
-import pipes
+import shlex
import salt.utils.path
@@ -76,7 +74,7 @@ def cluster_create(
cmd += ["--data-checksums"]
if wal_segsize:
cmd += ["--wal-segsize", wal_segsize]
- cmdstr = " ".join([pipes.quote(c) for c in cmd])
+ cmdstr = " ".join([shlex.quote(c) for c in cmd])
ret = __salt__["cmd.run_all"](cmdstr, python_shell=False)
if ret.get("retcode", 0) != 0:
log.error("Error creating a Postgresql cluster %s/%s", version, name)
@@ -97,7 +95,7 @@ def cluster_list(verbose=False):
salt '*' postgres.cluster_list verbose=True
"""
cmd = [salt.utils.path.which("pg_lsclusters"), "--no-header"]
- ret = __salt__["cmd.run_all"](" ".join([pipes.quote(c) for c in cmd]))
+ ret = __salt__["cmd.run_all"](" ".join([shlex.quote(c) for c in cmd]))
if ret.get("retcode", 0) != 0:
log.error("Error listing clusters")
cluster_dict = _parse_pg_lscluster(ret["stdout"])
@@ -118,7 +116,7 @@ def cluster_exists(version, name="main"):
salt '*' postgres.cluster_exists '9.3' 'main'
"""
- return "{}/{}".format(version, name) in cluster_list()
+ return f"{version}/{name}" in cluster_list()
def cluster_remove(version, name="main", stop=False):
@@ -141,13 +139,13 @@ def cluster_remove(version, name="main", stop=False):
if stop:
cmd += ["--stop"]
cmd += [str(version), name]
- cmdstr = " ".join([pipes.quote(c) for c in cmd])
+ cmdstr = " ".join([shlex.quote(c) for c in cmd])
ret = __salt__["cmd.run_all"](cmdstr, python_shell=False)
# FIXME - return Boolean ?
if ret.get("retcode", 0) != 0:
log.error("Error removing a Postgresql cluster %s/%s", version, name)
else:
- ret["changes"] = "Successfully removed cluster {}/{}".format(version, name)
+ ret["changes"] = f"Successfully removed cluster {version}/{name}"
return ret
@@ -158,7 +156,7 @@ def _parse_pg_lscluster(output):
cluster_dict = {}
for line in output.splitlines():
version, name, port, status, user, datadir, log = line.split()
- cluster_dict["{}/{}".format(version, name)] = {
+ cluster_dict[f"{version}/{name}"] = {
"port": int(port),
"status": status,
"user": user,
diff --git a/salt/modules/dockermod.py b/salt/modules/dockermod.py
index 69b722f0c95..331b6bb7482 100644
--- a/salt/modules/dockermod.py
+++ b/salt/modules/dockermod.py
@@ -206,8 +206,8 @@ import json
import logging
import os
import pathlib
-import pipes
import re
+import shlex
import shutil
import string
import subprocess
@@ -257,7 +257,6 @@ except ImportError:
HAS_NSENTER = bool(salt.utils.path.which("nsenter"))
-# Set up logging
log = logging.getLogger(__name__)
# Don't shadow built-in's.
@@ -397,7 +396,7 @@ def _get_client(timeout=NOTSET, **kwargs):
)
except Exception as exc: # pylint: disable=broad-except
raise CommandExecutionError(
- "Docker machine {} failed: {}".format(docker_machine, exc)
+ f"Docker machine {docker_machine} failed: {exc}"
)
try:
# docker-py 2.0 renamed this client attribute
@@ -497,7 +496,7 @@ def _change_state(name, action, expected, *args, **kwargs):
return {
"result": False,
"state": {"old": expected, "new": expected},
- "comment": "Container '{}' already {}".format(name, expected),
+ "comment": f"Container '{name}' already {expected}",
}
_client_wrapper(action, name, *args, **kwargs)
_clear_context()
@@ -535,9 +534,7 @@ def _get_md5(name, path):
"""
Get the MD5 checksum of a file from a container
"""
- output = run_stdout(
- name, "md5sum {}".format(pipes.quote(path)), ignore_retcode=True
- )
+ output = run_stdout(name, f"md5sum {shlex.quote(path)}", ignore_retcode=True)
try:
return output.split()[0]
except IndexError:
@@ -616,7 +613,7 @@ def _scrub_links(links, name):
if isinstance(links, list):
ret = []
for l in links:
- ret.append(l.replace("/{}/".format(name), "/", 1))
+ ret.append(l.replace(f"/{name}/", "/", 1))
else:
ret = links
@@ -639,11 +636,11 @@ def _size_fmt(num):
try:
num = int(num)
if num < 1024:
- return "{} bytes".format(num)
+ return f"{num} bytes"
num /= 1024.0
for unit in ("KiB", "MiB", "GiB", "TiB", "PiB"):
if num < 1024.0:
- return "{:3.1f} {}".format(num, unit)
+ return f"{num:3.1f} {unit}"
num /= 1024.0
except Exception: # pylint: disable=broad-except
log.error("Unable to format file size for '%s'", num)
@@ -658,7 +655,7 @@ def _client_wrapper(attr, *args, **kwargs):
catch_api_errors = kwargs.pop("catch_api_errors", True)
func = getattr(__context__["docker.client"], attr, None)
if func is None or not hasattr(func, "__call__"):
- raise SaltInvocationError("Invalid client action '{}'".format(attr))
+ raise SaltInvocationError(f"Invalid client action '{attr}'")
if attr in ("push", "pull"):
try:
# Refresh auth config from config.json
@@ -678,7 +675,7 @@ def _client_wrapper(attr, *args, **kwargs):
if catch_api_errors:
# Generic handling of Docker API errors
raise CommandExecutionError(
- "Error {}: {}".format(exc.response.status_code, exc.explanation)
+ f"Error {exc.response.status_code}: {exc.explanation}"
)
else:
# Allow API errors to be caught further up the stack
@@ -693,9 +690,9 @@ def _client_wrapper(attr, *args, **kwargs):
# If we're here, it's because an exception was caught earlier, and the
# API command failed.
- msg = "Unable to perform {}".format(attr)
+ msg = f"Unable to perform {attr}"
if err:
- msg += ": {}".format(err)
+ msg += f": {err}"
raise CommandExecutionError(msg)
@@ -722,7 +719,7 @@ def _import_status(data, item, repo_name, repo_tag):
return
elif all(x in string.hexdigits for x in status):
# Status is an image ID
- data["Image"] = "{}:{}".format(repo_name, repo_tag)
+ data["Image"] = f"{repo_name}:{repo_tag}"
data["Id"] = status
except (AttributeError, TypeError):
pass
@@ -881,7 +878,7 @@ def _get_create_kwargs(
ignore_collisions=False,
validate_ip_addrs=True,
client_args=None,
- **kwargs
+ **kwargs,
):
"""
Take input kwargs and return a kwargs dict to pass to docker-py's
@@ -899,7 +896,7 @@ def _get_create_kwargs(
skip_translate=skip_translate,
ignore_collisions=ignore_collisions,
validate_ip_addrs=validate_ip_addrs,
- **__utils__["args.clean_kwargs"](**kwargs)
+ **__utils__["args.clean_kwargs"](**kwargs),
)
if networks:
@@ -912,7 +909,7 @@ def _get_create_kwargs(
log.error(
"docker.create: Error getting client args: '%s'", exc, exc_info=True
)
- raise CommandExecutionError("Failed to get client args: {}".format(exc))
+ raise CommandExecutionError(f"Failed to get client args: {exc}")
full_host_config = {}
host_kwargs = {}
@@ -1473,15 +1470,15 @@ def login(*registries):
results = ret.setdefault("Results", {})
for registry in registries:
if registry not in registry_auth:
- errors.append("No match found for registry '{}'".format(registry))
+ errors.append(f"No match found for registry '{registry}'")
continue
try:
username = registry_auth[registry]["username"]
password = registry_auth[registry]["password"]
except TypeError:
- errors.append("Invalid configuration for registry '{}'".format(registry))
+ errors.append(f"Invalid configuration for registry '{registry}'")
except KeyError as exc:
- errors.append("Missing {} for registry '{}'".format(exc, registry))
+ errors.append(f"Missing {exc} for registry '{registry}'")
else:
cmd = ["docker", "login", "-u", username, "-p", password]
if registry.lower() != "hub":
@@ -1567,7 +1564,7 @@ def logout(*registries):
results = ret.setdefault("Results", {})
for registry in registries:
if registry not in registry_auth:
- errors.append("No match found for registry '{}'".format(registry))
+ errors.append(f"No match found for registry '{registry}'")
continue
else:
cmd = ["docker", "logout"]
@@ -1689,7 +1686,7 @@ def exists(name):
salt myminion docker.exists mycontainer
"""
- contextkey = "docker.exists.{}".format(name)
+ contextkey = f"docker.exists.{name}"
if contextkey in __context__:
return __context__[contextkey]
try:
@@ -1780,7 +1777,7 @@ def history(name, quiet=False):
)
for param in ("Size",):
if param in step:
- step["{}_Human".format(param)] = _size_fmt(step[param])
+ step[f"{param}_Human"] = _size_fmt(step[param])
ret.append(copy.deepcopy(step))
if quiet:
return [x.get("Command") for x in ret]
@@ -1842,9 +1839,7 @@ def images(verbose=False, **kwargs):
)
for param in ("Size", "VirtualSize"):
if param in bucket.get(img_id, {}):
- bucket[img_id]["{}_Human".format(param)] = _size_fmt(
- bucket[img_id][param]
- )
+ bucket[img_id][f"{param}_Human"] = _size_fmt(bucket[img_id][param])
context_data = __context__.get("docker.images", {})
ret = copy.deepcopy(context_data.get("tagged", {}))
@@ -1927,7 +1922,7 @@ def inspect(name):
raise
raise CommandExecutionError(
- "Error 404: No such image/container/volume/network: {}".format(name)
+ f"Error 404: No such image/container/volume/network: {name}"
)
@@ -1983,7 +1978,7 @@ def inspect_image(name):
ret = _client_wrapper("inspect_image", name)
for param in ("Size", "VirtualSize"):
if param in ret:
- ret["{}_Human".format(param)] = _size_fmt(ret[param])
+ ret[f"{param}_Human"] = _size_fmt(ret[param])
return ret
@@ -2277,7 +2272,7 @@ def port(name, private_port=None):
else:
# Sanity checks
if isinstance(private_port, int):
- pattern = "{}/*".format(private_port)
+ pattern = f"{private_port}/*"
else:
err = (
"Invalid private_port '{}'. Must either be a port number, "
@@ -2398,7 +2393,7 @@ def state(name):
salt myminion docker.state mycontainer
"""
- contextkey = "docker.state.{}".format(name)
+ contextkey = f"docker.state.{name}"
if contextkey in __context__:
return __context__[contextkey]
__context__[contextkey] = _get_state(inspect_container(name))
@@ -2438,9 +2433,7 @@ def search(name, official=False, trusted=False):
"""
response = _client_wrapper("search", name)
if not response:
- raise CommandExecutionError(
- "No images matched the search string '{}'".format(name)
- )
+ raise CommandExecutionError(f"No images matched the search string '{name}'")
key_map = {
"description": "Description",
@@ -2555,7 +2548,7 @@ def create(
ignore_collisions=False,
validate_ip_addrs=True,
client_timeout=salt.utils.dockermod.CLIENT_TIMEOUT,
- **kwargs
+ **kwargs,
):
"""
Create a new container
@@ -3281,7 +3274,7 @@ def create(
skip_translate=skip_translate,
ignore_collisions=ignore_collisions,
validate_ip_addrs=validate_ip_addrs,
- **kwargs
+ **kwargs,
)
if unused_kwargs:
@@ -3293,7 +3286,7 @@ def create(
log.debug(
"docker.create: creating container %susing the following arguments: %s",
- "with name '{}' ".format(name) if name is not None else "",
+ f"with name '{name}' " if name is not None else "",
kwargs,
)
time_started = time.time()
@@ -3331,7 +3324,7 @@ def run_container(
replace=False,
force=False,
networks=None,
- **kwargs
+ **kwargs,
):
"""
.. versionadded:: 2018.3.0
@@ -3433,7 +3426,7 @@ def run_container(
skip_translate=skip_translate,
ignore_collisions=ignore_collisions,
validate_ip_addrs=validate_ip_addrs,
- **kwargs
+ **kwargs,
)
# _get_create_kwargs() will have processed auto_remove and put it into the
@@ -3458,7 +3451,7 @@ def run_container(
log.debug(
"docker.create: creating container %susing the following arguments: %s",
- "with name '{}' ".format(name) if name is not None else "",
+ f"with name '{name}' " if name is not None else "",
kwargs,
)
@@ -3498,7 +3491,7 @@ def run_container(
rm_(name)
except CommandExecutionError as rm_exc:
exc_info.setdefault("other_errors", []).append(
- "Failed to auto_remove container: {}".format(rm_exc)
+ f"Failed to auto_remove container: {rm_exc}"
)
# Raise original exception with additional info
raise CommandExecutionError(exc.__str__(), info=exc_info)
@@ -3593,7 +3586,7 @@ def copy_from(name, source, dest, overwrite=False, makedirs=False):
"""
c_state = state(name)
if c_state != "running":
- raise CommandExecutionError("Container '{}' is not running".format(name))
+ raise CommandExecutionError(f"Container '{name}' is not running")
# Destination file sanity checks
if not os.path.isabs(dest):
@@ -3619,9 +3612,7 @@ def copy_from(name, source, dest, overwrite=False, makedirs=False):
)
)
else:
- raise SaltInvocationError(
- "Directory {} does not exist".format(dest_dir)
- )
+ raise SaltInvocationError(f"Directory {dest_dir} does not exist")
if not overwrite and os.path.exists(dest):
raise CommandExecutionError(
"Destination path {} already exists. Use overwrite=True to "
@@ -3632,19 +3623,14 @@ def copy_from(name, source, dest, overwrite=False, makedirs=False):
if not os.path.isabs(source):
raise SaltInvocationError("Source path must be absolute")
else:
- if (
- retcode(name, "test -e {}".format(pipes.quote(source)), ignore_retcode=True)
- == 0
- ):
+ if retcode(name, f"test -e {shlex.quote(source)}", ignore_retcode=True) == 0:
if (
- retcode(
- name, "test -f {}".format(pipes.quote(source)), ignore_retcode=True
- )
+ retcode(name, f"test -f {shlex.quote(source)}", ignore_retcode=True)
!= 0
):
raise SaltInvocationError("Source must be a regular file")
else:
- raise SaltInvocationError("Source file {} does not exist".format(source))
+ raise SaltInvocationError(f"Source file {source} does not exist")
# Before we try to replace the file, compare checksums.
source_md5 = _get_md5(name, source)
@@ -3657,7 +3643,7 @@ def copy_from(name, source, dest, overwrite=False, makedirs=False):
try:
src_path = ":".join((name, source))
except TypeError:
- src_path = "{}:{}".format(name, source)
+ src_path = f"{name}:{source}"
cmd = ["docker", "cp", src_path, dest_dir]
__salt__["cmd.run"](cmd, python_shell=False)
return source_md5 == __salt__["file.get_sum"](dest, "md5")
@@ -3784,7 +3770,7 @@ def export(name, path, overwrite=False, makedirs=False, compression=None, **kwar
salt myminion docker.export mycontainer /tmp/mycontainer.tar
salt myminion docker.export mycontainer /tmp/mycontainer.tar.xz push=True
"""
- err = "Path '{}' is not absolute".format(path)
+ err = f"Path '{path}' is not absolute"
try:
if not os.path.isabs(path):
raise SaltInvocationError(err)
@@ -3792,7 +3778,7 @@ def export(name, path, overwrite=False, makedirs=False, compression=None, **kwar
raise SaltInvocationError(err)
if os.path.exists(path) and not overwrite:
- raise CommandExecutionError("{} already exists".format(path))
+ raise CommandExecutionError(f"{path} already exists")
if compression is None:
if path.endswith(".tar.gz") or path.endswith(".tgz"):
@@ -3815,7 +3801,7 @@ def export(name, path, overwrite=False, makedirs=False, compression=None, **kwar
compression = "xz"
if compression and compression not in ("gzip", "bzip2", "xz"):
- raise SaltInvocationError("Invalid compression type '{}'".format(compression))
+ raise SaltInvocationError(f"Invalid compression type '{compression}'")
parent_dir = os.path.dirname(path)
if not os.path.isdir(parent_dir):
@@ -3828,16 +3814,14 @@ def export(name, path, overwrite=False, makedirs=False, compression=None, **kwar
os.makedirs(parent_dir)
except OSError as exc:
raise CommandExecutionError(
- "Unable to make parent dir {}: {}".format(parent_dir, exc)
+ f"Unable to make parent dir {parent_dir}: {exc}"
)
if compression == "gzip":
try:
out = gzip.open(path, "wb")
except OSError as exc:
- raise CommandExecutionError(
- "Unable to open {} for writing: {}".format(path, exc)
- )
+ raise CommandExecutionError(f"Unable to open {path} for writing: {exc}")
elif compression == "bzip2":
compressor = bz2.BZ2Compressor()
elif compression == "xz":
@@ -3875,9 +3859,7 @@ def export(name, path, overwrite=False, makedirs=False, compression=None, **kwar
os.remove(path)
except OSError:
pass
- raise CommandExecutionError(
- "Error occurred during container export: {}".format(exc)
- )
+ raise CommandExecutionError(f"Error occurred during container export: {exc}")
finally:
out.close()
ret = {"Time_Elapsed": time.time() - time_started}
@@ -4112,7 +4094,7 @@ def build(
# For the build function in the low-level API, the "tag" refers to the full
# tag (e.g. myuser/myimage:mytag). This is different than in other
# functions, where the repo and tag are passed separately.
- image_tag = "{}:{}".format(repository, tag) if repository and tag else None
+ image_tag = f"{repository}:{tag}" if repository and tag else None
time_started = time.time()
response = _client_wrapper(
@@ -4131,7 +4113,7 @@ def build(
if not response:
raise CommandExecutionError(
- "Build failed for {}, no response returned from Docker API".format(path)
+ f"Build failed for {path}, no response returned from Docker API"
)
stream_data = []
@@ -4168,7 +4150,7 @@ def build(
if "Id" not in ret:
# API returned information, but there was no confirmation of a
# successful build.
- msg = "Build failed for {}".format(path)
+ msg = f"Build failed for {path}"
log.error(msg)
log.error(stream_data)
if errors:
@@ -4179,7 +4161,7 @@ def build(
if resolved_tag:
ret["Image"] = resolved_tag
else:
- ret["Warning"] = "Failed to tag image as {}".format(image_tag)
+ ret["Warning"] = f"Failed to tag image as {image_tag}"
if api_response:
ret["API_Response"] = stream_data
@@ -4386,7 +4368,7 @@ def import_(source, repository, tag="latest", api_response=False):
if not response:
raise CommandExecutionError(
- "Import failed for {}, no response returned from Docker API".format(source)
+ f"Import failed for {source}, no response returned from Docker API"
)
elif api_response:
ret["API_Response"] = response
@@ -4406,7 +4388,7 @@ def import_(source, repository, tag="latest", api_response=False):
if "Id" not in ret:
# API returned information, but there was no confirmation of a
# successful push.
- msg = "Import failed for {}".format(source)
+ msg = f"Import failed for {source}"
if errors:
msg += ". Error(s) follow:\n\n{}".format("\n\n".join(errors))
raise CommandExecutionError(msg)
@@ -4481,7 +4463,7 @@ def load(path, repository=None, tag=None):
local_path = __salt__["container_resource.cache_file"](path)
if not os.path.isfile(local_path):
- raise CommandExecutionError("Source file {} does not exist".format(path))
+ raise CommandExecutionError(f"Source file {path} does not exist")
pre = images(all=True)
cmd = ["docker", "load", "-i", local_path]
@@ -4491,7 +4473,7 @@ def load(path, repository=None, tag=None):
_clear_context()
post = images(all=True)
if result["retcode"] != 0:
- msg = "Failed to load image(s) from {}".format(path)
+ msg = f"Failed to load image(s) from {path}"
if result["stderr"]:
msg += ": {}".format(result["stderr"])
raise CommandExecutionError(msg)
@@ -4512,7 +4494,7 @@ def load(path, repository=None, tag=None):
# strings when passed (e.g. a numeric tag would be loaded as an int
# or float), and because the tag_ function will stringify them if
# need be, a str.format is the correct thing to do here.
- tagged_image = "{}:{}".format(repository, tag)
+ tagged_image = f"{repository}:{tag}"
try:
result = tag_(top_level_images[0], repository=repository, tag=tag)
ret["Image"] = tagged_image
@@ -4549,7 +4531,7 @@ def layers(name):
):
ret.append(line)
if not ret:
- raise CommandExecutionError("Image '{}' not found".format(name))
+ raise CommandExecutionError(f"Image '{name}' not found")
return ret
@@ -4620,7 +4602,7 @@ def pull(
if not response:
raise CommandExecutionError(
- "Pull failed for {}, no response returned from Docker API".format(image)
+ f"Pull failed for {image}, no response returned from Docker API"
)
elif api_response:
ret["API_Response"] = response
@@ -4633,7 +4615,7 @@ def pull(
event = salt.utils.json.loads(event)
except Exception as exc: # pylint: disable=broad-except
raise CommandExecutionError(
- "Unable to interpret API event: '{}'".format(event),
+ f"Unable to interpret API event: '{event}'",
info={"Error": exc.__str__()},
)
try:
@@ -4715,7 +4697,7 @@ def push(
if not response:
raise CommandExecutionError(
- "Push failed for {}, no response returned from Docker API".format(image)
+ f"Push failed for {image}, no response returned from Docker API"
)
elif api_response:
ret["API_Response"] = response
@@ -4727,7 +4709,7 @@ def push(
event = salt.utils.json.loads(event)
except Exception as exc: # pylint: disable=broad-except
raise CommandExecutionError(
- "Unable to interpret API event: '{}'".format(event),
+ f"Unable to interpret API event: '{event}'",
info={"Error": exc.__str__()},
)
try:
@@ -4807,9 +4789,7 @@ def rmi(*names, **kwargs):
err += "image(s): {}".format(", ".join(deps["Images"]))
errors.append(err)
else:
- errors.append(
- "Error {}: {}".format(exc.response.status_code, exc.explanation)
- )
+ errors.append(f"Error {exc.response.status_code}: {exc.explanation}")
_clear_context()
ret = {
@@ -4897,7 +4877,7 @@ def save(name, path, overwrite=False, makedirs=False, compression=None, **kwargs
salt myminion docker.save centos:7 /tmp/cent7.tar
salt myminion docker.save 0123456789ab cdef01234567 /tmp/saved.tar
"""
- err = "Path '{}' is not absolute".format(path)
+ err = f"Path '{path}' is not absolute"
try:
if not os.path.isabs(path):
raise SaltInvocationError(err)
@@ -4905,7 +4885,7 @@ def save(name, path, overwrite=False, makedirs=False, compression=None, **kwargs
raise SaltInvocationError(err)
if os.path.exists(path) and not overwrite:
- raise CommandExecutionError("{} already exists".format(path))
+ raise CommandExecutionError(f"{path} already exists")
if compression is None:
if path.endswith(".tar.gz") or path.endswith(".tgz"):
@@ -4928,7 +4908,7 @@ def save(name, path, overwrite=False, makedirs=False, compression=None, **kwargs
compression = "xz"
if compression and compression not in ("gzip", "bzip2", "xz"):
- raise SaltInvocationError("Invalid compression type '{}'".format(compression))
+ raise SaltInvocationError(f"Invalid compression type '{compression}'")
parent_dir = os.path.dirname(path)
if not os.path.isdir(parent_dir):
@@ -4950,7 +4930,7 @@ def save(name, path, overwrite=False, makedirs=False, compression=None, **kwargs
time_started = time.time()
result = __salt__["cmd.run_all"](cmd, python_shell=False)
if result["retcode"] != 0:
- err = "Failed to save image(s) to {}".format(path)
+ err = f"Failed to save image(s) to {path}"
if result["stderr"]:
err += ": {}".format(result["stderr"])
raise CommandExecutionError(err)
@@ -4960,9 +4940,7 @@ def save(name, path, overwrite=False, makedirs=False, compression=None, **kwargs
try:
out = gzip.open(path, "wb")
except OSError as exc:
- raise CommandExecutionError(
- "Unable to open {} for writing: {}".format(path, exc)
- )
+ raise CommandExecutionError(f"Unable to open {path} for writing: {exc}")
elif compression == "bzip2":
compressor = bz2.BZ2Compressor()
elif compression == "xz":
@@ -4998,9 +4976,7 @@ def save(name, path, overwrite=False, makedirs=False, compression=None, **kwargs
os.remove(path)
except OSError:
pass
- raise CommandExecutionError(
- "Error occurred during image save: {}".format(exc)
- )
+ raise CommandExecutionError(f"Error occurred during image save: {exc}")
finally:
try:
# Clean up temp file
@@ -5120,7 +5096,7 @@ def create_network(
ignore_collisions=False,
validate_ip_addrs=True,
client_timeout=salt.utils.dockermod.CLIENT_TIMEOUT,
- **kwargs
+ **kwargs,
):
"""
.. versionchanged:: 2018.3.0
@@ -5360,7 +5336,7 @@ def create_network(
skip_translate=skip_translate,
ignore_collisions=ignore_collisions,
validate_ip_addrs=validate_ip_addrs,
- **__utils__["args.clean_kwargs"](**kwargs)
+ **__utils__["args.clean_kwargs"](**kwargs),
)
if "ipam" not in kwargs:
@@ -5692,7 +5668,7 @@ def pause(name):
return {
"result": False,
"state": {"old": orig_state, "new": orig_state},
- "comment": "Container '{}' is stopped, cannot pause".format(name),
+ "comment": f"Container '{name}' is stopped, cannot pause",
}
return _change_state(name, "pause", "paused")
@@ -5791,7 +5767,7 @@ def start_(name):
return {
"result": False,
"state": {"old": orig_state, "new": orig_state},
- "comment": "Container '{}' is paused, cannot start".format(name),
+ "comment": f"Container '{name}' is paused, cannot start",
}
return _change_state(name, "start", "running")
@@ -5896,7 +5872,7 @@ def unpause(name):
return {
"result": False,
"state": {"old": orig_state, "new": orig_state},
- "comment": "Container '{}' is stopped, cannot unpause".format(name),
+ "comment": f"Container '{name}' is stopped, cannot unpause",
}
return _change_state(name, "unpause", "running")
@@ -5945,7 +5921,7 @@ def wait(name, ignore_already_stopped=False, fail_on_exit_status=False):
# Container doesn't exist anymore
return {
"result": ignore_already_stopped,
- "comment": "Container '{}' absent".format(name),
+ "comment": f"Container '{name}' absent",
}
already_stopped = pre == "stopped"
response = _client_wrapper("wait", name)
@@ -5969,7 +5945,7 @@ def wait(name, ignore_already_stopped=False, fail_on_exit_status=False):
"exit_status": response,
}
if already_stopped:
- result["comment"] = "Container '{}' already stopped".format(name)
+ result["comment"] = f"Container '{name}' already stopped"
if fail_on_exit_status and result["result"]:
result["result"] = result["exit_status"] == 0
return result
@@ -5982,7 +5958,7 @@ def prune(
build=False,
volumes=False,
system=None,
- **filters
+ **filters,
):
"""
.. versionadded:: 2019.2.0
@@ -6668,7 +6644,7 @@ def script_retcode(
def _generate_tmp_path():
- return os.path.join("/tmp", "salt.docker.{}".format(uuid.uuid4().hex[:6]))
+ return os.path.join("/tmp", f"salt.docker.{uuid.uuid4().hex[:6]}")
def _prepare_trans_tar(name, sls_opts, mods=None, pillar=None, extra_filerefs=""):
@@ -6929,7 +6905,7 @@ def call(name, function, *args, **kwargs):
]
+ list(args)
+ [
- "{}={}".format(key, value)
+ f"{key}={value}"
for (key, value) in kwargs.items()
if not key.startswith("__")
]
diff --git a/salt/modules/lxc.py b/salt/modules/lxc.py
index bea6445db98..d2c1e66491e 100644
--- a/salt/modules/lxc.py
+++ b/salt/modules/lxc.py
@@ -12,9 +12,9 @@ import datetime
import difflib
import logging
import os
-import pipes
import random
import re
+import shlex
import shutil
import string
import tempfile
@@ -1834,8 +1834,8 @@ def _after_ignition_network_profile(cmd, ret, name, network_profile, path, nic_o
# destroy the container if it was partially created
cmd = "lxc-destroy"
if path:
- cmd += " -P {}".format(pipes.quote(path))
- cmd += " -n {}".format(name)
+ cmd += f" -P {shlex.quote(path)}"
+ cmd += f" -n {name}"
__salt__["cmd.retcode"](cmd, python_shell=False)
raise CommandExecutionError(
"Container could not be created with cmd '{}': {}".format(
@@ -1997,7 +1997,7 @@ def create(
)
options["imgtar"] = img_tar
if path:
- cmd += " -P {}".format(pipes.quote(path))
+ cmd += f" -P {shlex.quote(path)}"
if not os.path.exists(path):
os.makedirs(path)
if config:
@@ -2138,7 +2138,7 @@ def clone(name, orig, profile=None, network_profile=None, nic_opts=None, **kwarg
cmd = "lxc-clone"
cmd += " {} -o {} -n {}".format(snapshot, orig, name)
if path:
- cmd += " -P {}".format(pipes.quote(path))
+ cmd += f" -P {shlex.quote(path)}"
if not os.path.exists(path):
os.makedirs(path)
if backing:
@@ -2186,7 +2186,7 @@ def ls_(active=None, cache=True, path=None):
ret = []
cmd = "lxc-ls"
if path:
- cmd += " -P {}".format(pipes.quote(path))
+ cmd += f" -P {shlex.quote(path)}"
if active:
cmd += " --active"
output = __salt__["cmd.run_stdout"](cmd, python_shell=False)
@@ -2242,8 +2242,8 @@ def list_(extra=False, limit=None, path=None):
for container in ctnrs:
cmd = "lxc-info"
if path:
- cmd += " -P {}".format(pipes.quote(path))
- cmd += " -n {}".format(container)
+ cmd += f" -P {shlex.quote(path)}"
+ cmd += f" -n {container}"
c_info = __salt__["cmd.run"](cmd, python_shell=False, output_loglevel="debug")
c_state = None
for line in c_info.splitlines():
@@ -2301,13 +2301,13 @@ def _change_state(
# Kill the container first
scmd = "lxc-stop"
if path:
- scmd += " -P {}".format(pipes.quote(path))
- scmd += " -k -n {}".format(name)
+ scmd += f" -P {shlex.quote(path)}"
+ scmd += f" -k -n {name}"
__salt__["cmd.run"](scmd, python_shell=False)
if path and " -P " not in cmd:
- cmd += " -P {}".format(pipes.quote(path))
- cmd += " -n {}".format(name)
+ cmd += f" -P {shlex.quote(path)}"
+ cmd += f" -n {name}"
# certain lxc commands need to be taken with care (lxc-start)
# as te command itself mess with double forks; we must not
@@ -2337,8 +2337,8 @@ def _change_state(
# some commands do not wait, so we will
rcmd = "lxc-wait"
if path:
- rcmd += " -P {}".format(pipes.quote(path))
- rcmd += " -n {} -s {}".format(name, expected.upper())
+ rcmd += f" -P {shlex.quote(path)}"
+ rcmd += f" -n {name} -s {expected.upper()}"
__salt__["cmd.run"](rcmd, python_shell=False, timeout=30)
_clear_context()
post = state(name, path=path)
@@ -2459,7 +2459,7 @@ def start(name, **kwargs):
lxc_config = os.path.join(cpath, name, "config")
# we try to start, even without config, if global opts are there
if os.path.exists(lxc_config):
- cmd += " -f {}".format(pipes.quote(lxc_config))
+ cmd += f" -f {shlex.quote(lxc_config)}"
cmd += " -d"
_ensure_exists(name, path=path)
if state(name, path=path) == "frozen":
@@ -2564,7 +2564,7 @@ def freeze(name, **kwargs):
start(name, path=path)
cmd = "lxc-freeze"
if path:
- cmd += " -P {}".format(pipes.quote(path))
+ cmd += f" -P {shlex.quote(path)}"
ret = _change_state(cmd, name, "frozen", use_vt=use_vt, path=path)
if orig_state == "stopped" and start_:
ret["state"]["old"] = orig_state
@@ -2599,7 +2599,7 @@ def unfreeze(name, path=None, use_vt=None):
raise CommandExecutionError("Container '{}' is stopped".format(name))
cmd = "lxc-unfreeze"
if path:
- cmd += " -P {}".format(pipes.quote(path))
+ cmd += f" -P {shlex.quote(path)}"
return _change_state(cmd, name, "running", path=path, use_vt=use_vt)
@@ -2693,8 +2693,8 @@ def state(name, path=None):
else:
cmd = "lxc-info"
if path:
- cmd += " -P {}".format(pipes.quote(path))
- cmd += " -n {}".format(name)
+ cmd += f" -P {shlex.quote(path)}"
+ cmd += f" -n {name}"
ret = __salt__["cmd.run_all"](cmd, python_shell=False)
if ret["retcode"] != 0:
_clear_context()
@@ -2731,8 +2731,8 @@ def get_parameter(name, parameter, path=None):
_ensure_exists(name, path=path)
cmd = "lxc-cgroup"
if path:
- cmd += " -P {}".format(pipes.quote(path))
- cmd += " -n {} {}".format(name, parameter)
+ cmd += f" -P {shlex.quote(path)}"
+ cmd += f" -n {name} {parameter}"
ret = __salt__["cmd.run_all"](cmd, python_shell=False)
if ret["retcode"] != 0:
raise CommandExecutionError(
@@ -2762,8 +2762,8 @@ def set_parameter(name, parameter, value, path=None):
cmd = "lxc-cgroup"
if path:
- cmd += " -P {}".format(pipes.quote(path))
- cmd += " -n {} {} {}".format(name, parameter, value)
+ cmd += f" -P {shlex.quote(path)}"
+ cmd += f" -n {name} {parameter} {value}"
ret = __salt__["cmd.run_all"](cmd, python_shell=False)
if ret["retcode"] != 0:
return False
@@ -3662,8 +3662,8 @@ def attachable(name, path=None):
log.debug("Checking if LXC container %s is attachable", name)
cmd = "lxc-attach"
if path:
- cmd += " -P {}".format(pipes.quote(path))
- cmd += " --clear-env -n {} -- /usr/bin/env".format(name)
+ cmd += f" -P {shlex.quote(path)}"
+ cmd += f" --clear-env -n {name} -- /usr/bin/env"
result = (
__salt__["cmd.retcode"](
cmd, python_shell=False, output_loglevel="quiet", ignore_retcode=True
diff --git a/salt/modules/mac_keychain.py b/salt/modules/mac_keychain.py
index a823c428b76..7fdc162b9aa 100644
--- a/salt/modules/mac_keychain.py
+++ b/salt/modules/mac_keychain.py
@@ -11,20 +11,6 @@ import shlex
import salt.utils.platform
-try:
- import pipes
-
- HAS_DEPS = True
-except ImportError:
- HAS_DEPS = False
-
-if hasattr(shlex, "quote"):
- _quote = shlex.quote
-elif HAS_DEPS and hasattr(pipes, "quote"):
- _quote = pipes.quote
-else:
- _quote = None
-
log = logging.getLogger(__name__)
__virtualname__ = "keychain"
@@ -34,7 +20,7 @@ def __virtual__():
"""
Only work on Mac OS
"""
- if salt.utils.platform.is_darwin() and _quote is not None:
+ if salt.utils.platform.is_darwin():
return __virtualname__
return (False, "Only available on Mac OS systems with pipes")
@@ -82,7 +68,7 @@ def install(
if keychain_password is not None:
unlock_keychain(keychain, keychain_password)
- cmd = "security import {} -P {} -k {}".format(cert, password, keychain)
+ cmd = f"security import {cert} -P {password} -k {keychain}"
if allow_any:
cmd += " -A"
return __salt__["cmd.run"](cmd)
@@ -117,7 +103,7 @@ def uninstall(
if keychain_password is not None:
unlock_keychain(keychain, keychain_password)
- cmd = 'security delete-certificate -c "{}" {}'.format(cert_name, keychain)
+ cmd = f'security delete-certificate -c "{cert_name}" {keychain}'
return __salt__["cmd.run"](cmd)
@@ -137,7 +123,7 @@ def list_certs(keychain="/Library/Keychains/System.keychain"):
"""
cmd = (
'security find-certificate -a {} | grep -o "alis".*\\" | '
- "grep -o '\\\"[-A-Za-z0-9.:() ]*\\\"'".format(_quote(keychain))
+ "grep -o '\\\"[-A-Za-z0-9.:() ]*\\\"'".format(shlex.quote(keychain))
)
out = __salt__["cmd.run"](cmd, python_shell=True)
return out.replace('"', "").split("\n")
@@ -165,7 +151,7 @@ def get_friendly_name(cert, password):
"""
cmd = (
"openssl pkcs12 -in {} -passin pass:{} -info -nodes -nokeys 2> /dev/null | "
- "grep friendlyName:".format(_quote(cert), _quote(password))
+ "grep friendlyName:".format(shlex.quote(cert), shlex.quote(password))
)
out = __salt__["cmd.run"](cmd, python_shell=True)
return out.replace("friendlyName: ", "").strip()
@@ -187,7 +173,7 @@ def get_default_keychain(user=None, domain="user"):
salt '*' keychain.get_default_keychain
"""
- cmd = "security default-keychain -d {}".format(domain)
+ cmd = f"security default-keychain -d {domain}"
return __salt__["cmd.run"](cmd, runas=user)
@@ -210,7 +196,7 @@ def set_default_keychain(keychain, domain="user", user=None):
salt '*' keychain.set_keychain /Users/fred/Library/Keychains/login.keychain
"""
- cmd = "security default-keychain -d {} -s {}".format(domain, keychain)
+ cmd = f"security default-keychain -d {domain} -s {keychain}"
return __salt__["cmd.run"](cmd, runas=user)
@@ -233,7 +219,7 @@ def unlock_keychain(keychain, password):
salt '*' keychain.unlock_keychain /tmp/test.p12 test123
"""
- cmd = "security unlock-keychain -p {} {}".format(password, keychain)
+ cmd = f"security unlock-keychain -p {password} {keychain}"
__salt__["cmd.run"](cmd)
@@ -261,7 +247,7 @@ def get_hash(name, password=None):
name, password
)
else:
- cmd = 'security find-certificate -c "{}" -m -p'.format(name)
+ cmd = f'security find-certificate -c "{name}" -m -p'
out = __salt__["cmd.run"](cmd)
matches = re.search(
diff --git a/salt/modules/macpackage.py b/salt/modules/macpackage.py
index faf5810d4fc..f9a6b7bb95c 100644
--- a/salt/modules/macpackage.py
+++ b/salt/modules/macpackage.py
@@ -9,31 +9,16 @@ import shlex
import salt.utils.platform
-try:
- import pipes
-
- HAS_DEPS = True
-except ImportError:
- HAS_DEPS = False
-
-
log = logging.getLogger(__name__)
-__virtualname__ = "macpackage"
-
-if hasattr(shlex, "quote"):
- _quote = shlex.quote
-elif HAS_DEPS and hasattr(pipes, "quote"):
- _quote = pipes.quote
-else:
- _quote = None
+__virtualname__ = "macpackage"
def __virtual__():
"""
Only work on Mac OS
"""
- if salt.utils.platform.is_darwin() and _quote is not None:
+ if salt.utils.platform.is_darwin():
return __virtualname__
return (False, "Only available on Mac OS systems with pipes")
@@ -60,11 +45,11 @@ def install(pkg, target="LocalSystem", store=False, allow_untrusted=False):
"""
if "*." not in pkg:
# If we use wildcards, we cannot use quotes
- pkg = _quote(pkg)
+ pkg = shlex.quote(pkg)
- target = _quote(target)
+ target = shlex.quote(target)
- cmd = "installer -pkg {} -target {}".format(pkg, target)
+ cmd = f"installer -pkg {pkg} -target {target}"
if store:
cmd += " -store"
if allow_untrusted:
@@ -109,7 +94,7 @@ def install_app(app, target="/Applications/"):
if not app[-1] == "/":
app += "/"
- cmd = 'rsync -a --delete "{}" "{}"'.format(app, target)
+ cmd = f'rsync -a --delete "{app}" "{target}"'
return __salt__["cmd.run"](cmd)
@@ -154,7 +139,7 @@ def mount(dmg):
temp_dir = __salt__["temp.dir"](prefix="dmg-")
- cmd = 'hdiutil attach -readonly -nobrowse -mountpoint {} "{}"'.format(temp_dir, dmg)
+ cmd = f'hdiutil attach -readonly -nobrowse -mountpoint {temp_dir} "{dmg}"'
return __salt__["cmd.run"](cmd), temp_dir
@@ -176,7 +161,7 @@ def unmount(mountpoint):
salt '*' macpackage.unmount /dev/disk2
"""
- cmd = 'hdiutil detach "{}"'.format(mountpoint)
+ cmd = f'hdiutil detach "{mountpoint}"'
return __salt__["cmd.run"](cmd)
@@ -216,7 +201,7 @@ def get_pkg_id(pkg):
salt '*' macpackage.get_pkg_id /tmp/test.pkg
"""
- pkg = _quote(pkg)
+ pkg = shlex.quote(pkg)
package_ids = []
# Create temp directory
@@ -224,7 +209,7 @@ def get_pkg_id(pkg):
try:
# List all of the PackageInfo files
- cmd = "xar -t -f {} | grep PackageInfo".format(pkg)
+ cmd = f"xar -t -f {pkg} | grep PackageInfo"
out = __salt__["cmd.run"](cmd, python_shell=True, output_loglevel="quiet")
files = out.split("\n")
@@ -264,12 +249,12 @@ def get_mpkg_ids(mpkg):
salt '*' macpackage.get_mpkg_ids /dev/disk2
"""
- mpkg = _quote(mpkg)
+ mpkg = shlex.quote(mpkg)
package_infos = []
base_path = os.path.dirname(mpkg)
# List all of the .pkg files
- cmd = "find {} -name *.pkg".format(base_path)
+ cmd = f"find {base_path} -name *.pkg"
out = __salt__["cmd.run"](cmd, python_shell=True)
pkg_files = out.split("\n")
@@ -281,7 +266,7 @@ def get_mpkg_ids(mpkg):
def _get_pkg_id_from_pkginfo(pkginfo):
# Find our identifiers
- pkginfo = _quote(pkginfo)
+ pkginfo = shlex.quote(pkginfo)
cmd = "cat {} | grep -Eo 'identifier=\"[a-zA-Z.0-9\\-]*\"' | cut -c 13- | tr -d '\"'".format(
pkginfo
)
@@ -294,8 +279,8 @@ def _get_pkg_id_from_pkginfo(pkginfo):
def _get_pkg_id_dir(path):
- path = _quote(os.path.join(path, "Contents/Info.plist"))
- cmd = '/usr/libexec/PlistBuddy -c "print :CFBundleIdentifier" {}'.format(path)
+ path = shlex.quote(os.path.join(path, "Contents/Info.plist"))
+ cmd = f'/usr/libexec/PlistBuddy -c "print :CFBundleIdentifier" {path}'
# We can only use wildcards in python_shell which is
# sent by the macpackage state
diff --git a/salt/modules/openstack_config.py b/salt/modules/openstack_config.py
index 823afbf1c60..937c10da61a 100644
--- a/salt/modules/openstack_config.py
+++ b/salt/modules/openstack_config.py
@@ -13,28 +13,11 @@ import shlex
import salt.exceptions
import salt.utils.decorators.path
-try:
- import pipes
-
- HAS_DEPS = True
-except ImportError:
- HAS_DEPS = False
-
-if hasattr(shlex, "quote"):
- _quote = shlex.quote
-elif HAS_DEPS and hasattr(pipes, "quote"):
- _quote = pipes.quote
-else:
- _quote = None
-
-
# Don't shadow built-in's.
__func_alias__ = {"set_": "set"}
def __virtual__():
- if _quote is None and not HAS_DEPS:
- return (False, "Missing dependencies")
return True
@@ -69,10 +52,10 @@ def set_(filename, section, parameter, value):
salt-call openstack_config.set /etc/keystone/keystone.conf sql connection foo
"""
- filename = _quote(filename)
- section = _quote(section)
- parameter = _quote(parameter)
- value = _quote(str(value))
+ filename = shlex.quote(filename)
+ section = shlex.quote(section)
+ parameter = shlex.quote(parameter)
+ value = shlex.quote(str(value))
result = __salt__["cmd.run_all"](
"openstack-config --set {} {} {} {}".format(
@@ -109,12 +92,12 @@ def get(filename, section, parameter):
"""
- filename = _quote(filename)
- section = _quote(section)
- parameter = _quote(parameter)
+ filename = shlex.quote(filename)
+ section = shlex.quote(section)
+ parameter = shlex.quote(parameter)
result = __salt__["cmd.run_all"](
- "openstack-config --get {} {} {}".format(filename, section, parameter),
+ f"openstack-config --get {filename} {section} {parameter}",
python_shell=False,
)
@@ -145,12 +128,12 @@ def delete(filename, section, parameter):
salt-call openstack_config.delete /etc/keystone/keystone.conf sql connection
"""
- filename = _quote(filename)
- section = _quote(section)
- parameter = _quote(parameter)
+ filename = shlex.quote(filename)
+ section = shlex.quote(section)
+ parameter = shlex.quote(parameter)
result = __salt__["cmd.run_all"](
- "openstack-config --del {} {} {}".format(filename, section, parameter),
+ f"openstack-config --del {filename} {section} {parameter}",
python_shell=False,
)
diff --git a/salt/modules/postgres.py b/salt/modules/postgres.py
index 25a72f1063c..f73959a92ed 100644
--- a/salt/modules/postgres.py
+++ b/salt/modules/postgres.py
@@ -46,8 +46,8 @@ import hmac
import io
import logging
import os
-import pipes
import re
+import shlex
import tempfile
import salt.utils.files
@@ -136,7 +136,7 @@ def __virtual__():
for util in utils:
if not salt.utils.path.which(util):
if not _find_pg_binary(util):
- return (False, "{} was not found".format(util))
+ return (False, f"{util} was not found")
return True
@@ -241,14 +241,14 @@ def _run_initdb(
raise CommandExecutionError("initdb executable not found.")
cmd = [
_INITDB_BIN,
- "--pgdata={}".format(name),
- "--username={}".format(user),
- "--auth={}".format(auth),
- "--encoding={}".format(encoding),
+ f"--pgdata={name}",
+ f"--username={user}",
+ f"--auth={auth}",
+ f"--encoding={encoding}",
]
if locale is not None:
- cmd.append("--locale={}".format(locale))
+ cmd.append(f"--locale={locale}")
# intentionally use short option, as the long option name has been
# renamed from "xlogdir" to "waldir" in PostgreSQL 10
@@ -262,9 +262,9 @@ def _run_initdb(
if password is not None:
pgpassfile = salt.utils.files.mkstemp(text=True)
with salt.utils.files.fopen(pgpassfile, "w") as fp_:
- fp_.write(salt.utils.stringutils.to_str("{}".format(password)))
+ fp_.write(salt.utils.stringutils.to_str(f"{password}"))
__salt__["file.chown"](pgpassfile, runas, "")
- cmd.extend(["--pwfile={}".format(pgpassfile)])
+ cmd.extend([f"--pwfile={pgpassfile}"])
kwargs = dict(
runas=runas,
@@ -273,7 +273,7 @@ def _run_initdb(
"postgres.timeout", default=_DEFAULT_COMMAND_TIMEOUT_SECS
),
)
- cmdstr = " ".join([pipes.quote(c) for c in cmd])
+ cmdstr = " ".join([shlex.quote(c) for c in cmd])
ret = __salt__["cmd.run_all"](cmdstr, python_shell=False, **kwargs)
if ret.get("retcode", 0) != 0:
@@ -582,9 +582,7 @@ def _quote_ddl_value(value, quote="'"):
if value is None:
return None
if quote in value: # detect trivial sqli
- raise SaltInvocationError(
- "Unsupported character {} in value: {}".format(quote, value)
- )
+ raise SaltInvocationError(f"Unsupported character {quote} in value: {value}")
return "{quote}{value}{quote}".format(quote=quote, value=value)
@@ -617,7 +615,7 @@ def db_create(
"""
# Base query to create a database
- query = 'CREATE DATABASE "{}"'.format(name)
+ query = f'CREATE DATABASE "{name}"'
# "With"-options to create a database
with_args = salt.utils.odict.OrderedDict(
@@ -685,11 +683,9 @@ def db_alter(
else:
queries = []
if owner:
- queries.append('ALTER DATABASE "{}" OWNER TO "{}"'.format(name, owner))
+ queries.append(f'ALTER DATABASE "{name}" OWNER TO "{owner}"')
if tablespace:
- queries.append(
- 'ALTER DATABASE "{}" SET TABLESPACE "{}"'.format(name, tablespace)
- )
+ queries.append(f'ALTER DATABASE "{name}" SET TABLESPACE "{tablespace}"')
for query in queries:
ret = _psql_prepare_and_run(
["-c", query],
@@ -726,10 +722,10 @@ def db_remove(
salt '*' postgres.db_remove 'dbname'
"""
for query in [
- 'REVOKE CONNECT ON DATABASE "{db}" FROM public;'.format(db=name),
+ f'REVOKE CONNECT ON DATABASE "{name}" FROM public;',
"SELECT pid, pg_terminate_backend(pid) FROM pg_stat_activity WHERE datname ="
" '{db}' AND pid <> pg_backend_pid();".format(db=name),
- 'DROP DATABASE "{db}";'.format(db=name),
+ f'DROP DATABASE "{name}";',
]:
ret = _psql_prepare_and_run(
["-c", query],
@@ -741,7 +737,7 @@ def db_remove(
password=password,
)
if ret["retcode"] != 0:
- raise Exception("Failed: ret={}".format(ret))
+ raise Exception(f"Failed: ret={ret}")
return True
@@ -846,10 +842,10 @@ def tablespace_create(
owner_query = ""
options_query = ""
if owner:
- owner_query = 'OWNER "{}"'.format(owner)
+ owner_query = f'OWNER "{owner}"'
# should come out looking like: 'OWNER postgres'
if options:
- optionstext = ["{} = {}".format(k, v) for k, v in options.items()]
+ optionstext = [f"{k} = {v}" for k, v in options.items()]
options_query = "WITH ( {} )".format(", ".join(optionstext))
# should come out looking like: 'WITH ( opt1 = 1.0, opt2 = 4.0 )'
query = "CREATE TABLESPACE \"{}\" {} LOCATION '{}' {}".format(
@@ -902,9 +898,9 @@ def tablespace_alter(
queries = []
if new_name:
- queries.append('ALTER TABLESPACE "{}" RENAME TO "{}"'.format(name, new_name))
+ queries.append(f'ALTER TABLESPACE "{name}" RENAME TO "{new_name}"')
if new_owner:
- queries.append('ALTER TABLESPACE "{}" OWNER TO "{}"'.format(name, new_owner))
+ queries.append(f'ALTER TABLESPACE "{name}" OWNER TO "{new_owner}"')
if set_option:
queries.append(
'ALTER TABLESPACE "{}" SET ({} = {})'.format(
@@ -912,7 +908,7 @@ def tablespace_alter(
)
)
if reset_option:
- queries.append('ALTER TABLESPACE "{}" RESET ({})'.format(name, reset_option))
+ queries.append(f'ALTER TABLESPACE "{name}" RESET ({reset_option})')
for query in queries:
ret = _psql_prepare_and_run(
@@ -950,7 +946,7 @@ def tablespace_remove(
.. versionadded:: 2015.8.0
"""
- query = 'DROP TABLESPACE "{}"'.format(name)
+ query = f'DROP TABLESPACE "{name}"'
ret = _psql_prepare_and_run(
["-c", query],
user=user,
@@ -1158,11 +1154,11 @@ def _add_role_flag(string, test, flag, cond=None, prefix="NO", addtxt="", skip=F
cond = test
if test is not None:
if cond:
- string = "{} {}".format(string, flag)
+ string = f"{string} {flag}"
else:
- string = "{0} {2}{1}".format(string, flag, prefix)
+ string = f"{string} {prefix}{flag}"
if addtxt:
- string = "{} {}".format(string, addtxt)
+ string = f"{string} {addtxt}"
return string
@@ -1224,7 +1220,7 @@ def _verify_password(role, password, verifier, method):
def _md5_password(role, password):
return "md5{}".format(
hashlib.md5( # nosec
- salt.utils.stringutils.to_bytes("{}{}".format(password, role))
+ salt.utils.stringutils.to_bytes(f"{password}{role}")
).hexdigest()
)
@@ -1343,7 +1339,7 @@ def _role_cmd_args(
if isinstance(groups, list):
groups = ",".join(groups)
for group in groups.split(","):
- sub_cmd = '{}; GRANT "{}" TO "{}"'.format(sub_cmd, group, name)
+ sub_cmd = f'{sub_cmd}; GRANT "{group}" TO "{name}"'
return sub_cmd
@@ -1380,7 +1376,7 @@ def _role_create(
log.info("%s '%s' already exists", typ_.capitalize(), name)
return False
- sub_cmd = 'CREATE ROLE "{}" WITH'.format(name)
+ sub_cmd = f'CREATE ROLE "{name}" WITH'
sub_cmd = "{} {}".format(
sub_cmd,
_role_cmd_args(
@@ -1506,7 +1502,7 @@ def _role_update(
log.info("%s '%s' could not be found", typ_.capitalize(), name)
return False
- sub_cmd = 'ALTER ROLE "{}" WITH'.format(name)
+ sub_cmd = f'ALTER ROLE "{name}" WITH'
sub_cmd = "{} {}".format(
sub_cmd,
_role_cmd_args(
@@ -1613,7 +1609,7 @@ def _role_remove(
return False
# user exists, proceed
- sub_cmd = 'DROP ROLE "{}"'.format(name)
+ sub_cmd = f'DROP ROLE "{name}"'
_psql_prepare_and_run(
["-c", sub_cmd],
runas=runas,
@@ -1995,14 +1991,14 @@ def create_extension(
args = ["CREATE EXTENSION"]
if if_not_exists:
args.append("IF NOT EXISTS")
- args.append('"{}"'.format(name))
+ args.append(f'"{name}"')
sargs = []
if schema:
- sargs.append('SCHEMA "{}"'.format(schema))
+ sargs.append(f'SCHEMA "{schema}"')
if ext_version:
- sargs.append("VERSION {}".format(ext_version))
+ sargs.append(f"VERSION {ext_version}")
if from_version:
- sargs.append("FROM {}".format(from_version))
+ sargs.append(f"FROM {from_version}")
if sargs:
args.append("WITH")
args.extend(sargs)
@@ -2011,13 +2007,9 @@ def create_extension(
else:
args = []
if schema and _EXTENSION_TO_MOVE in mtdata:
- args.append(
- 'ALTER EXTENSION "{}" SET SCHEMA "{}";'.format(name, schema)
- )
+ args.append(f'ALTER EXTENSION "{name}" SET SCHEMA "{schema}";')
if ext_version and _EXTENSION_TO_UPGRADE in mtdata:
- args.append(
- 'ALTER EXTENSION "{}" UPDATE TO {};'.format(name, ext_version)
- )
+ args.append(f'ALTER EXTENSION "{name}" UPDATE TO {ext_version};')
cmd = " ".join(args).strip()
if cmd:
_psql_prepare_and_run(
@@ -2227,7 +2219,7 @@ def owner_to(
sqlfile = tempfile.NamedTemporaryFile()
sqlfile.write("begin;\n")
- sqlfile.write('alter database "{}" owner to "{}";\n'.format(dbname, ownername))
+ sqlfile.write(f'alter database "{dbname}" owner to "{ownername}";\n')
queries = (
# schemas
@@ -2335,9 +2327,9 @@ def schema_create(
log.info("'%s' already exists in '%s'", name, dbname)
return False
- sub_cmd = 'CREATE SCHEMA "{}"'.format(name)
+ sub_cmd = f'CREATE SCHEMA "{name}"'
if owner is not None:
- sub_cmd = '{} AUTHORIZATION "{}"'.format(sub_cmd, owner)
+ sub_cmd = f'{sub_cmd} AUTHORIZATION "{owner}"'
ret = _psql_prepare_and_run(
["-c", sub_cmd],
@@ -2401,7 +2393,7 @@ def schema_remove(
return False
# schema exists, proceed
- sub_cmd = 'DROP SCHEMA "{}"'.format(name)
+ sub_cmd = f'DROP SCHEMA "{name}"'
_psql_prepare_and_run(
["-c", sub_cmd],
runas=user,
@@ -2721,7 +2713,7 @@ def language_create(
log.info("Language %s already exists in %s", name, maintenance_db)
return False
- query = "CREATE LANGUAGE {}".format(name)
+ query = f"CREATE LANGUAGE {name}"
ret = _psql_prepare_and_run(
["-c", query],
@@ -2776,7 +2768,7 @@ def language_remove(
log.info("Language %s does not exist in %s", name, maintenance_db)
return False
- query = "DROP LANGUAGE {}".format(name)
+ query = f"DROP LANGUAGE {name}"
ret = _psql_prepare_and_run(
["-c", query],
@@ -3035,9 +3027,7 @@ def _validate_privileges(object_type, privs, privileges):
_perms.append("ALL")
if object_type not in _PRIVILEGES_OBJECTS:
- raise SaltInvocationError(
- "Invalid object_type: {} provided".format(object_type)
- )
+ raise SaltInvocationError(f"Invalid object_type: {object_type} provided")
if not set(privs).issubset(set(_perms)):
raise SaltInvocationError(
@@ -3145,9 +3135,7 @@ def privileges_list(
query = _make_privileges_list_query(name, object_type, prepend)
if object_type not in _PRIVILEGES_OBJECTS:
- raise SaltInvocationError(
- "Invalid object_type: {} provided".format(object_type)
- )
+ raise SaltInvocationError(f"Invalid object_type: {object_type} provided")
rows = psql_query(
query,
@@ -3439,15 +3427,15 @@ def privileges_grant(
_grants = ",".join(_privs)
if object_type in ["table", "sequence"]:
- on_part = '{}."{}"'.format(prepend, object_name)
+ on_part = f'{prepend}."{object_name}"'
elif object_type == "function":
- on_part = "{}".format(object_name)
+ on_part = f"{object_name}"
else:
- on_part = '"{}"'.format(object_name)
+ on_part = f'"{object_name}"'
if grant_option:
if object_type == "group":
- query = 'GRANT {} TO "{}" WITH ADMIN OPTION'.format(object_name, name)
+ query = f'GRANT {object_name} TO "{name}" WITH ADMIN OPTION'
elif object_type in ("table", "sequence") and object_name.upper() == "ALL":
query = 'GRANT {} ON ALL {}S IN SCHEMA {} TO "{}" WITH GRANT OPTION'.format(
_grants, object_type.upper(), prepend, name
@@ -3458,7 +3446,7 @@ def privileges_grant(
)
else:
if object_type == "group":
- query = 'GRANT {} TO "{}"'.format(object_name, name)
+ query = f'GRANT {object_name} TO "{name}"'
elif object_type in ("table", "sequence") and object_name.upper() == "ALL":
query = 'GRANT {} ON ALL {}S IN SCHEMA {} TO "{}"'.format(
_grants, object_type.upper(), prepend, name
@@ -3587,12 +3575,12 @@ def privileges_revoke(
_grants = ",".join(_privs)
if object_type in ["table", "sequence"]:
- on_part = "{}.{}".format(prepend, object_name)
+ on_part = f"{prepend}.{object_name}"
else:
on_part = object_name
if object_type == "group":
- query = "REVOKE {} FROM {}".format(object_name, name)
+ query = f"REVOKE {object_name} FROM {name}"
else:
query = "REVOKE {} ON {} {} FROM {}".format(
_grants, object_type.upper(), on_part, name
diff --git a/salt/utils/cloud.py b/salt/utils/cloud.py
index 9edf006c299..b7208dc4a64 100644
--- a/salt/utils/cloud.py
+++ b/salt/utils/cloud.py
@@ -10,8 +10,8 @@ import hashlib
import logging
import multiprocessing
import os
-import pipes
import re
+import shlex
import shutil
import socket
import stat
@@ -199,7 +199,7 @@ def __ssh_gateway_arguments(kwargs):
"-oUserKnownHostsFile=/dev/null",
"-oControlPath=none",
str(ssh_gateway_key),
- "{}@{}".format(ssh_gateway_user, ssh_gateway),
+ f"{ssh_gateway_user}@{ssh_gateway}",
"-p",
str(ssh_gateway_port),
str(ssh_gateway_command),
@@ -228,18 +228,18 @@ def os_script(os_, vm_=None, opts=None, minion=""):
# The user provided an absolute path to the deploy script, let's use it
return __render_script(os_, vm_, opts, minion)
- if os.path.isabs("{}.sh".format(os_)):
+ if os.path.isabs(f"{os_}.sh"):
# The user provided an absolute path to the deploy script, although no
# extension was provided. Let's use it anyway.
- return __render_script("{}.sh".format(os_), vm_, opts, minion)
+ return __render_script(f"{os_}.sh", vm_, opts, minion)
for search_path in opts["deploy_scripts_search_path"]:
if os.path.isfile(os.path.join(search_path, os_)):
return __render_script(os.path.join(search_path, os_), vm_, opts, minion)
- if os.path.isfile(os.path.join(search_path, "{}.sh".format(os_))):
+ if os.path.isfile(os.path.join(search_path, f"{os_}.sh")):
return __render_script(
- os.path.join(search_path, "{}.sh".format(os_)), vm_, opts, minion
+ os.path.join(search_path, f"{os_}.sh"), vm_, opts, minion
)
# No deploy script was found, return an empty string
return ""
@@ -416,7 +416,7 @@ def bootstrap(vm_, opts=None):
)
if key_filename is not None and not os.path.isfile(key_filename):
raise SaltCloudConfigError(
- "The defined ssh_keyfile '{}' does not exist".format(key_filename)
+ f"The defined ssh_keyfile '{key_filename}' does not exist"
)
has_ssh_agent = False
if (
@@ -782,8 +782,8 @@ def wait_for_port(
# Don't add new hosts to the host key database
"-oStrictHostKeyChecking=no",
# make sure ssh can time out on connection lose
- "-oServerAliveInterval={}".format(server_alive_interval),
- "-oServerAliveCountMax={}".format(server_alive_count_max),
+ f"-oServerAliveInterval={server_alive_interval}",
+ f"-oServerAliveCountMax={server_alive_count_max}",
# Set hosts key database path to /dev/null, i.e., non-existing
"-oUserKnownHostsFile=/dev/null",
# Don't re-use the SSH connection. Less failures.
@@ -808,21 +808,21 @@ def wait_for_port(
]
)
# Netcat command testing remote port
- command = "nc -z -w5 -q0 {} {}".format(host, port)
+ command = f"nc -z -w5 -q0 {host} {port}"
# SSH command
pcmd = "ssh {} {}@{} -p {} {}".format(
" ".join(ssh_args),
gateway["ssh_gateway_user"],
ssh_gateway,
ssh_gateway_port,
- pipes.quote("date"),
+ shlex.quote("date"),
)
cmd = "ssh {} {}@{} -p {} {}".format(
" ".join(ssh_args),
gateway["ssh_gateway_user"],
ssh_gateway,
ssh_gateway_port,
- pipes.quote(command),
+ shlex.quote(command),
)
log.debug("SSH command: '%s'", cmd)
@@ -893,7 +893,7 @@ class Client:
service_name=None,
):
self.service_name = service_name
- self._exe_file = "{}.exe".format(self.service_name)
+ self._exe_file = f"{self.service_name}.exe"
self._client = PsExecClient(server, username, password, port, encrypt)
self._client._service = ScmrService(self.service_name, self._client.session)
@@ -943,7 +943,7 @@ class Client:
# delete the PAExec executable
smb_tree = TreeConnect(
self._client.session,
- r"\\{}\ADMIN$".format(self._client.connection.server_name),
+ rf"\\{self._client.connection.server_name}\ADMIN$",
)
log.info("Connecting to SMB Tree %s", smb_tree.share_name)
smb_tree.connect()
@@ -968,10 +968,10 @@ def run_winexe_command(cmd, args, host, username, password, port=445):
"""
Run a command remotely via the winexe executable
"""
- creds = "-U '{}%{}' //{}".format(username, password, host)
- logging_creds = "-U '{}%XXX-REDACTED-XXX' //{}".format(username, host)
- cmd = "winexe {} {} {}".format(creds, cmd, args)
- logging_cmd = "winexe {} {} {}".format(logging_creds, cmd, args)
+ creds = f"-U '{username}%{password}' //{host}"
+ logging_creds = f"-U '{username}%XXX-REDACTED-XXX' //{host}"
+ cmd = f"winexe {creds} {cmd} {args}"
+ logging_cmd = f"winexe {logging_creds} {cmd} {args}"
return win_cmd(cmd, logging_command=logging_cmd)
@@ -979,7 +979,7 @@ def run_psexec_command(cmd, args, host, username, password, port=445):
"""
Run a command remotely using the psexec protocol
"""
- service_name = "PS-Exec-{}".format(uuid.uuid4())
+ service_name = f"PS-Exec-{uuid.uuid4()}"
with Client(
host, username, password, port=port, encrypt=False, service_name=service_name
) as client:
@@ -1098,7 +1098,7 @@ def validate_windows_cred_winexe(
"""
Check if the windows credentials are valid
"""
- cmd = "winexe -U '{}%{}' //{} \"hostname\"".format(username, password, host)
+ cmd = f"winexe -U '{username}%{password}' //{host} \"hostname\""
logging_cmd = "winexe -U '{}%XXX-REDACTED-XXX' //{} \"hostname\"".format(
username, host
)
@@ -1230,7 +1230,7 @@ def deploy_windows(
winrm_port=5986,
winrm_use_ssl=True,
winrm_verify_ssl=True,
- **kwargs
+ **kwargs,
):
"""
Copy the install files to a remote Windows box, and execute them
@@ -1289,20 +1289,20 @@ def deploy_windows(
salt.utils.smb.mkdirs("salttemp", conn=smb_conn)
root_dir = "ProgramData/Salt Project/Salt"
- salt.utils.smb.mkdirs("{}/conf/pki/minion".format(root_dir), conn=smb_conn)
+ salt.utils.smb.mkdirs(f"{root_dir}/conf/pki/minion", conn=smb_conn)
root_dir = "ProgramData\\Salt Project\\Salt"
if minion_pub:
salt.utils.smb.put_str(
minion_pub,
- "{}\\conf\\pki\\minion\\minion.pub".format(root_dir),
+ f"{root_dir}\\conf\\pki\\minion\\minion.pub",
conn=smb_conn,
)
if minion_pem:
salt.utils.smb.put_str(
minion_pem,
- "{}\\conf\\pki\\minion\\minion.pem".format(root_dir),
+ f"{root_dir}\\conf\\pki\\minion\\minion.pem",
conn=smb_conn,
)
@@ -1314,7 +1314,7 @@ def deploy_windows(
try:
salt.utils.smb.put_file(
master_sign_pub_file,
- "{}\\conf\\pki\\minion\\master_sign.pub".format(root_dir),
+ f"{root_dir}\\conf\\pki\\minion\\master_sign.pub",
conn=smb_conn,
)
except Exception as e: # pylint: disable=broad-except
@@ -1332,26 +1332,27 @@ def deploy_windows(
installer = comps[-1]
salt.utils.smb.put_file(
win_installer,
- "salttemp\\{}".format(installer),
+ f"salttemp\\{installer}",
"C$",
conn=smb_conn,
)
+ cmd = f"c:\\salttemp\\{installer}"
+ args = [
+ "/S",
+ f"/master={_format_master_param(master)}",
+ f"/minion-name={name}",
+ ]
+
if use_winrm:
- winrm_cmd(
- winrm_session,
- "c:\\salttemp\\{}".format(installer),
- ["/S", "/master={}".format(master), "/minion-name={}".format(name)],
- )
+ winrm_cmd(winrm_session, cmd, args)
else:
- cmd = "c:\\salttemp\\{}".format(installer)
- args = "/S /master={} /minion-name={}".format(master, name)
stdout, stderr, ret_code = run_psexec_command(
- cmd, args, host, username, password
+ cmd, " ".join(args), host, username, password
)
if ret_code != 0:
- raise Exception("Fail installer {}".format(ret_code))
+ raise Exception(f"Fail installer {ret_code}")
# Copy over minion_conf
if minion_conf:
@@ -1367,7 +1368,7 @@ def deploy_windows(
if minion_grains:
salt.utils.smb.put_str(
salt_config_to_yaml(minion_grains, line_break="\r\n"),
- "{}\\conf\\grains".format(root_dir),
+ f"{root_dir}\\conf\\grains",
conn=smb_conn,
)
# Add special windows minion configuration
@@ -1384,7 +1385,7 @@ def deploy_windows(
minion_conf = dict(minion_conf, **windows_minion_conf)
salt.utils.smb.put_str(
salt_config_to_yaml(minion_conf, line_break="\r\n"),
- "{}\\conf\\minion".format(root_dir),
+ f"{root_dir}\\conf\\minion",
conn=smb_conn,
)
# Delete C:\salttmp\ and installer file
@@ -1394,7 +1395,7 @@ def deploy_windows(
winrm_cmd(winrm_session, "rmdir", ["/Q", "/S", "C:\\salttemp\\"])
else:
salt.utils.smb.delete_file(
- "salttemp\\{}".format(installer), "C$", conn=smb_conn
+ f"salttemp\\{installer}", "C$", conn=smb_conn
)
salt.utils.smb.delete_directory("salttemp", "C$", conn=smb_conn)
# Shell out to psexec to ensure salt-minion service started
@@ -1418,8 +1419,8 @@ def deploy_windows(
# Fire deploy action
fire_event(
"event",
- "{} has been deployed at {}".format(name, host),
- "salt/cloud/{}/deploy_windows".format(name),
+ f"{name} has been deployed at {host}",
+ f"salt/cloud/{name}/deploy_windows",
args={"name": name},
sock_dir=opts.get("sock_dir", os.path.join(__opts__["sock_dir"], "master")),
transport=opts.get("transport", "zeromq"),
@@ -1469,7 +1470,7 @@ def deploy_script(
master_sign_pub_file=None,
cloud_grains=None,
force_minion_config=False,
- **kwargs
+ **kwargs,
):
"""
Copy a deploy script to a remote server, execute it, and remove it
@@ -1485,7 +1486,7 @@ def deploy_script(
)
if key_filename is not None and not os.path.isfile(key_filename):
raise SaltCloudConfigError(
- "The defined key_filename '{}' does not exist".format(key_filename)
+ f"The defined key_filename '{key_filename}' does not exist"
)
gateway = None
@@ -1532,35 +1533,28 @@ def deploy_script(
ssh_kwargs["password"] = password
if root_cmd(
- "test -e '{}'".format(tmp_dir),
- tty,
- sudo,
- allow_failure=True,
- **ssh_kwargs
+ f"test -e '{tmp_dir}'", tty, sudo, allow_failure=True, **ssh_kwargs
):
ret = root_cmd(
- "sh -c \"( mkdir -p -m 700 '{}' )\"".format(tmp_dir),
+ f"sh -c \"( mkdir -p -m 700 '{tmp_dir}' )\"",
tty,
sudo,
- **ssh_kwargs
+ **ssh_kwargs,
)
if ret:
raise SaltCloudSystemExit(
- "Can't create temporary directory in {} !".format(tmp_dir)
+ f"Can't create temporary directory in {tmp_dir} !"
)
if sudo:
comps = tmp_dir.lstrip("/").rstrip("/").split("/")
if comps:
if len(comps) > 1 or comps[0] != "tmp":
ret = root_cmd(
- 'chown {} "{}"'.format(username, tmp_dir),
- tty,
- sudo,
- **ssh_kwargs
+ f'chown {username} "{tmp_dir}"', tty, sudo, **ssh_kwargs
)
if ret:
raise SaltCloudSystemExit(
- "Cant set {} ownership on {}".format(username, tmp_dir)
+ f"Cant set {username} ownership on {tmp_dir}"
)
if not isinstance(file_map, dict):
@@ -1590,15 +1584,13 @@ def deploy_script(
remote_dir = os.path.dirname(remote_file)
if remote_dir not in remote_dirs:
- root_cmd(
- "mkdir -p '{}'".format(remote_dir), tty, sudo, **ssh_kwargs
- )
+ root_cmd(f"mkdir -p '{remote_dir}'", tty, sudo, **ssh_kwargs)
if ssh_kwargs["username"] != "root":
root_cmd(
"chown {} '{}'".format(ssh_kwargs["username"], remote_dir),
tty,
sudo,
- **ssh_kwargs
+ **ssh_kwargs,
)
remote_dirs.append(remote_dir)
ssh_file(opts, remote_file, kwargs=ssh_kwargs, local_file=local_file)
@@ -1606,21 +1598,21 @@ def deploy_script(
# Minion configuration
if minion_pem:
- ssh_file(opts, "{}/minion.pem".format(tmp_dir), minion_pem, ssh_kwargs)
+ ssh_file(opts, f"{tmp_dir}/minion.pem", minion_pem, ssh_kwargs)
ret = root_cmd(
- "chmod 600 '{}/minion.pem'".format(tmp_dir), tty, sudo, **ssh_kwargs
+ f"chmod 600 '{tmp_dir}/minion.pem'", tty, sudo, **ssh_kwargs
)
if ret:
raise SaltCloudSystemExit(
- "Can't set perms on {}/minion.pem".format(tmp_dir)
+ f"Can't set perms on {tmp_dir}/minion.pem"
)
if minion_pub:
- ssh_file(opts, "{}/minion.pub".format(tmp_dir), minion_pub, ssh_kwargs)
+ ssh_file(opts, f"{tmp_dir}/minion.pub", minion_pub, ssh_kwargs)
if master_sign_pub_file:
ssh_file(
opts,
- "{}/master_sign.pub".format(tmp_dir),
+ f"{tmp_dir}/master_sign.pub",
kwargs=ssh_kwargs,
local_file=master_sign_pub_file,
)
@@ -1638,7 +1630,7 @@ def deploy_script(
if minion_grains:
ssh_file(
opts,
- "{}/grains".format(tmp_dir),
+ f"{tmp_dir}/grains",
salt_config_to_yaml(minion_grains),
ssh_kwargs,
)
@@ -1646,24 +1638,22 @@ def deploy_script(
minion_conf["grains"] = {"salt-cloud": cloud_grains}
ssh_file(
opts,
- "{}/minion".format(tmp_dir),
+ f"{tmp_dir}/minion",
salt_config_to_yaml(minion_conf),
ssh_kwargs,
)
# Master configuration
if master_pem:
- ssh_file(opts, "{}/master.pem".format(tmp_dir), master_pem, ssh_kwargs)
+ ssh_file(opts, f"{tmp_dir}/master.pem", master_pem, ssh_kwargs)
ret = root_cmd(
- "chmod 600 '{}/master.pem'".format(tmp_dir), tty, sudo, **ssh_kwargs
+ f"chmod 600 '{tmp_dir}/master.pem'", tty, sudo, **ssh_kwargs
)
if ret:
- raise SaltCloudSystemExit(
- "Cant set perms on {}/master.pem".format(tmp_dir)
- )
+ raise SaltCloudSystemExit(f"Cant set perms on {tmp_dir}/master.pem")
if master_pub:
- ssh_file(opts, "{}/master.pub".format(tmp_dir), master_pub, ssh_kwargs)
+ ssh_file(opts, f"{tmp_dir}/master.pub", master_pub, ssh_kwargs)
if master_conf:
if not isinstance(master_conf, dict):
@@ -1677,34 +1667,31 @@ def deploy_script(
ssh_file(
opts,
- "{}/master".format(tmp_dir),
+ f"{tmp_dir}/master",
salt_config_to_yaml(master_conf),
ssh_kwargs,
)
# XXX: We need to make these paths configurable
- preseed_minion_keys_tempdir = "{}/preseed-minion-keys".format(tmp_dir)
+ preseed_minion_keys_tempdir = f"{tmp_dir}/preseed-minion-keys"
if preseed_minion_keys is not None:
# Create remote temp dir
ret = root_cmd(
- "mkdir '{}'".format(preseed_minion_keys_tempdir),
- tty,
- sudo,
- **ssh_kwargs
+ f"mkdir '{preseed_minion_keys_tempdir}'", tty, sudo, **ssh_kwargs
)
if ret:
raise SaltCloudSystemExit(
- "Cant create {}".format(preseed_minion_keys_tempdir)
+ f"Cant create {preseed_minion_keys_tempdir}"
)
ret = root_cmd(
- "chmod 700 '{}'".format(preseed_minion_keys_tempdir),
+ f"chmod 700 '{preseed_minion_keys_tempdir}'",
tty,
sudo,
- **ssh_kwargs
+ **ssh_kwargs,
)
if ret:
raise SaltCloudSystemExit(
- "Can't set perms on {}".format(preseed_minion_keys_tempdir)
+ f"Can't set perms on {preseed_minion_keys_tempdir}"
)
if ssh_kwargs["username"] != "root":
root_cmd(
@@ -1713,7 +1700,7 @@ def deploy_script(
),
tty,
sudo,
- **ssh_kwargs
+ **ssh_kwargs,
)
# Copy pre-seed minion keys
@@ -1723,10 +1710,10 @@ def deploy_script(
if ssh_kwargs["username"] != "root":
root_cmd(
- "chown -R root '{}'".format(preseed_minion_keys_tempdir),
+ f"chown -R root '{preseed_minion_keys_tempdir}'",
tty,
sudo,
- **ssh_kwargs
+ **ssh_kwargs,
)
if ret:
raise SaltCloudSystemExit(
@@ -1740,25 +1727,21 @@ def deploy_script(
for command in preflight_cmds:
cmd_ret = root_cmd(command, tty, sudo, **ssh_kwargs)
if cmd_ret:
- raise SaltCloudSystemExit(
- "Pre-flight command failed: '{}'".format(command)
- )
+ raise SaltCloudSystemExit(f"Pre-flight command failed: '{command}'")
# The actual deploy script
if script:
# got strange escaping issues with sudoer, going onto a
# subshell fixes that
- ssh_file(opts, "{}/deploy.sh".format(tmp_dir), script, ssh_kwargs)
+ ssh_file(opts, f"{tmp_dir}/deploy.sh", script, ssh_kwargs)
ret = root_cmd(
- "sh -c \"( chmod +x '{}/deploy.sh' )\";exit $?".format(tmp_dir),
+ f"sh -c \"( chmod +x '{tmp_dir}/deploy.sh' )\";exit $?",
tty,
sudo,
- **ssh_kwargs
+ **ssh_kwargs,
)
if ret:
- raise SaltCloudSystemExit(
- "Can't set perms on {}/deploy.sh".format(tmp_dir)
- )
+ raise SaltCloudSystemExit(f"Can't set perms on {tmp_dir}/deploy.sh")
time_used = time.mktime(time.localtime()) - time.mktime(starttime)
newtimeout = timeout - time_used
@@ -1774,7 +1757,7 @@ def deploy_script(
kwargs=dict(
name=name, sock_dir=sock_dir, timeout=newtimeout, queue=queue
),
- name="DeployScriptCheckAuth({})".format(name),
+ name=f"DeployScriptCheckAuth({name})",
)
log.debug("Starting new process to wait for salt-minion")
process.start()
@@ -1782,7 +1765,7 @@ def deploy_script(
# Run the deploy script
if script:
if "bootstrap-salt" in script:
- deploy_command += " -c '{}'".format(tmp_dir)
+ deploy_command += f" -c '{tmp_dir}'"
if force_minion_config:
deploy_command += " -F"
if make_syndic is True:
@@ -1794,9 +1777,9 @@ def deploy_script(
if keep_tmp is True:
deploy_command += " -K"
if preseed_minion_keys is not None:
- deploy_command += " -k '{}'".format(preseed_minion_keys_tempdir)
+ deploy_command += f" -k '{preseed_minion_keys_tempdir}'"
if script_args:
- deploy_command += " {}".format(script_args)
+ deploy_command += f" {script_args}"
if script_env:
if not isinstance(script_env, dict):
@@ -1815,15 +1798,15 @@ def deploy_script(
# Upload our environ setter wrapper
ssh_file(
opts,
- "{}/environ-deploy-wrapper.sh".format(tmp_dir),
+ f"{tmp_dir}/environ-deploy-wrapper.sh",
"\n".join(environ_script_contents),
ssh_kwargs,
)
root_cmd(
- "chmod +x '{}/environ-deploy-wrapper.sh'".format(tmp_dir),
+ f"chmod +x '{tmp_dir}/environ-deploy-wrapper.sh'",
tty,
sudo,
- **ssh_kwargs
+ **ssh_kwargs,
)
# The deploy command is now our wrapper
deploy_command = "'{}/environ-deploy-wrapper.sh'".format(
@@ -1831,22 +1814,20 @@ def deploy_script(
)
if root_cmd(deploy_command, tty, sudo, **ssh_kwargs) != 0:
raise SaltCloudSystemExit(
- "Executing the command '{}' failed".format(deploy_command)
+ f"Executing the command '{deploy_command}' failed"
)
log.debug("Executed command '%s'", deploy_command)
# Remove the deploy script
if not keep_tmp:
- root_cmd(
- "rm -f '{}/deploy.sh'".format(tmp_dir), tty, sudo, **ssh_kwargs
- )
+ root_cmd(f"rm -f '{tmp_dir}/deploy.sh'", tty, sudo, **ssh_kwargs)
log.debug("Removed %s/deploy.sh", tmp_dir)
if script_env:
root_cmd(
- "rm -f '{}/environ-deploy-wrapper.sh'".format(tmp_dir),
+ f"rm -f '{tmp_dir}/environ-deploy-wrapper.sh'",
tty,
sudo,
- **ssh_kwargs
+ **ssh_kwargs,
)
log.debug("Removed %s/environ-deploy-wrapper.sh", tmp_dir)
@@ -1855,57 +1836,40 @@ def deploy_script(
else:
# Remove minion configuration
if minion_pub:
- root_cmd(
- "rm -f '{}/minion.pub'".format(tmp_dir), tty, sudo, **ssh_kwargs
- )
+ root_cmd(f"rm -f '{tmp_dir}/minion.pub'", tty, sudo, **ssh_kwargs)
log.debug("Removed %s/minion.pub", tmp_dir)
if minion_pem:
- root_cmd(
- "rm -f '{}/minion.pem'".format(tmp_dir), tty, sudo, **ssh_kwargs
- )
+ root_cmd(f"rm -f '{tmp_dir}/minion.pem'", tty, sudo, **ssh_kwargs)
log.debug("Removed %s/minion.pem", tmp_dir)
if minion_conf:
- root_cmd(
- "rm -f '{}/grains'".format(tmp_dir), tty, sudo, **ssh_kwargs
- )
+ root_cmd(f"rm -f '{tmp_dir}/grains'", tty, sudo, **ssh_kwargs)
log.debug("Removed %s/grains", tmp_dir)
- root_cmd(
- "rm -f '{}/minion'".format(tmp_dir), tty, sudo, **ssh_kwargs
- )
+ root_cmd(f"rm -f '{tmp_dir}/minion'", tty, sudo, **ssh_kwargs)
log.debug("Removed %s/minion", tmp_dir)
if master_sign_pub_file:
root_cmd(
- "rm -f {}/master_sign.pub".format(tmp_dir),
- tty,
- sudo,
- **ssh_kwargs
+ f"rm -f {tmp_dir}/master_sign.pub", tty, sudo, **ssh_kwargs
)
log.debug("Removed %s/master_sign.pub", tmp_dir)
# Remove master configuration
if master_pub:
- root_cmd(
- "rm -f '{}/master.pub'".format(tmp_dir), tty, sudo, **ssh_kwargs
- )
+ root_cmd(f"rm -f '{tmp_dir}/master.pub'", tty, sudo, **ssh_kwargs)
log.debug("Removed %s/master.pub", tmp_dir)
if master_pem:
- root_cmd(
- "rm -f '{}/master.pem'".format(tmp_dir), tty, sudo, **ssh_kwargs
- )
+ root_cmd(f"rm -f '{tmp_dir}/master.pem'", tty, sudo, **ssh_kwargs)
log.debug("Removed %s/master.pem", tmp_dir)
if master_conf:
- root_cmd(
- "rm -f '{}/master'".format(tmp_dir), tty, sudo, **ssh_kwargs
- )
+ root_cmd(f"rm -f '{tmp_dir}/master'", tty, sudo, **ssh_kwargs)
log.debug("Removed %s/master", tmp_dir)
# Remove pre-seed keys directory
if preseed_minion_keys is not None:
root_cmd(
- "rm -rf '{}'".format(preseed_minion_keys_tempdir),
+ f"rm -rf '{preseed_minion_keys_tempdir}'",
tty,
sudo,
- **ssh_kwargs
+ **ssh_kwargs,
)
log.debug("Removed %s", preseed_minion_keys_tempdir)
@@ -1920,15 +1884,13 @@ def deploy_script(
# for line in output:
# print(line)
log.info("Executing %s on the salt-minion", start_action)
- root_cmd(
- "salt-call {}".format(start_action), tty, sudo, **ssh_kwargs
- )
+ root_cmd(f"salt-call {start_action}", tty, sudo, **ssh_kwargs)
log.info("Finished executing %s on the salt-minion", start_action)
# Fire deploy action
fire_event(
"event",
- "{} has been deployed at {}".format(name, host),
- "salt/cloud/{}/deploy_script".format(name),
+ f"{name} has been deployed at {host}",
+ f"salt/cloud/{name}/deploy_script",
args={"name": name, "host": host},
sock_dir=opts.get(
"sock_dir", os.path.join(__opts__["sock_dir"], "master")
@@ -1961,7 +1923,7 @@ def run_inline_script(
tty=None,
opts=None,
tmp_dir="/tmp/.saltcloud-inline_script",
- **kwargs
+ **kwargs,
):
"""
Run the inline script commands, one by one
@@ -2018,11 +1980,11 @@ def run_inline_script(
# TODO: check edge cases (e.g. ssh gateways, salt deploy disabled, etc.)
if (
root_cmd(
- 'test -e \\"{}\\"'.format(tmp_dir),
+ f'test -e \\"{tmp_dir}\\"',
tty,
sudo,
allow_failure=True,
- **ssh_kwargs
+ **ssh_kwargs,
)
and inline_script
):
@@ -2030,11 +1992,11 @@ def run_inline_script(
for cmd_line in inline_script:
log.info("Executing inline command: %s", cmd_line)
ret = root_cmd(
- 'sh -c "( {} )"'.format(cmd_line),
+ f'sh -c "( {cmd_line} )"',
tty,
sudo,
allow_failure=True,
- **ssh_kwargs
+ **ssh_kwargs,
)
if ret:
log.info("[%s] Output: %s", cmd_line, ret)
@@ -2138,7 +2100,7 @@ def _exec_ssh_cmd(cmd, error_msg=None, allow_failure=False, **kwargs):
time.sleep(0.5)
if proc.exitstatus != 0 and allow_failure is False:
raise SaltCloudSystemExit(
- "Command '{}' failed. Exit code: {}".format(cmd, proc.exitstatus)
+ f"Command '{cmd}' failed. Exit code: {proc.exitstatus}"
)
return proc.exitstatus
except salt.utils.vt.TerminalException as err:
@@ -2241,7 +2203,7 @@ def scp_file(dest_path, contents=None, kwargs=None, local_file=None):
cmd,
error_msg="Failed to upload file '{0}': {1}\n{2}",
password_retries=3,
- **kwargs
+ **kwargs,
)
finally:
if contents is not None:
@@ -2359,7 +2321,7 @@ def sftp_file(dest_path, contents=None, kwargs=None, local_file=None):
cmd,
error_msg="Failed to upload file '{0}': {1}\n{2}",
password_retries=3,
- **kwargs
+ **kwargs,
)
finally:
if contents is not None:
@@ -2419,11 +2381,11 @@ def root_cmd(command, tty, sudo, allow_failure=False, **kwargs):
if sudo:
if sudo_password is None:
- command = "sudo {}".format(command)
+ command = f"sudo {command}"
logging_command = command
else:
- logging_command = 'sudo -S "XXX-REDACTED-XXX" {}'.format(command)
- command = "sudo -S {}".format(command)
+ logging_command = f'sudo -S "XXX-REDACTED-XXX" {command}'
+ command = f"sudo -S {command}"
log.debug("Using sudo to run command %s", logging_command)
@@ -2442,9 +2404,9 @@ def root_cmd(command, tty, sudo, allow_failure=False, **kwargs):
ssh_args.extend(
[
# Don't add new hosts to the host key database
- "-oStrictHostKeyChecking={}".format(host_key_checking),
+ f"-oStrictHostKeyChecking={host_key_checking}",
# Set hosts key database path to /dev/null, i.e., non-existing
- "-oUserKnownHostsFile={}".format(known_hosts_file),
+ f"-oUserKnownHostsFile={known_hosts_file}",
# Don't re-use the SSH connection. Less failures.
"-oControlPath=none",
]
@@ -2477,12 +2439,12 @@ def root_cmd(command, tty, sudo, allow_failure=False, **kwargs):
cmd = "ssh {0} {1[username]}@{1[hostname]} ".format(" ".join(ssh_args), kwargs)
logging_command = cmd + logging_command
- cmd = cmd + pipes.quote(command)
+ cmd = cmd + shlex.quote(command)
hard_timeout = kwargs.get("hard_timeout")
if hard_timeout is not None:
- logging_command = "timeout {} {}".format(hard_timeout, logging_command)
- cmd = "timeout {} {}".format(hard_timeout, cmd)
+ logging_command = f"timeout {hard_timeout} {logging_command}"
+ cmd = f"timeout {hard_timeout} {cmd}"
log.debug("SSH command: '%s'", logging_command)
@@ -2504,7 +2466,7 @@ def check_auth(name, sock_dir=None, queue=None, timeout=300):
ret = event.get_event(full=True)
if ret is None:
continue
- if ret["tag"] == "salt/minion/{}/start".format(name):
+ if ret["tag"] == f"salt/minion/{name}/start":
queue.put(name)
newtimeout = 0
log.debug("Minion %s is ready to receive commands", name)
@@ -2550,7 +2512,7 @@ def check_name(name, safe_chars):
"""
Check whether the specified name contains invalid characters
"""
- regexp = re.compile("[^{}]".format(safe_chars))
+ regexp = re.compile(f"[^{safe_chars}]")
if regexp.search(name):
raise SaltCloudException(
"{} contains characters not supported by this cloud provider. "
@@ -2844,7 +2806,7 @@ def request_minion_cachedir(
"provider": provider,
}
- fname = "{}.p".format(minion_id)
+ fname = f"{minion_id}.p"
path = os.path.join(base, "requested", fname)
with salt.utils.files.fopen(path, "wb") as fh_:
salt.utils.msgpack.dump(data, fh_, encoding=MSGPACK_ENCODING)
@@ -2875,7 +2837,7 @@ def change_minion_cachedir(
if base is None:
base = __opts__["cachedir"]
- fname = "{}.p".format(minion_id)
+ fname = f"{minion_id}.p"
path = os.path.join(base, cachedir, fname)
with salt.utils.files.fopen(path, "r") as fh_:
@@ -2898,7 +2860,7 @@ def activate_minion_cachedir(minion_id, base=None):
if base is None:
base = __opts__["cachedir"]
- fname = "{}.p".format(minion_id)
+ fname = f"{minion_id}.p"
src = os.path.join(base, "requested", fname)
dst = os.path.join(base, "active")
shutil.move(src, dst)
@@ -2920,7 +2882,7 @@ def delete_minion_cachedir(minion_id, provider, opts, base=None):
base = __opts__["cachedir"]
driver = next(iter(__opts__["providers"][provider].keys()))
- fname = "{}.p".format(minion_id)
+ fname = f"{minion_id}.p"
for cachedir in "requested", "active":
path = os.path.join(base, cachedir, driver, provider, fname)
log.debug("path: %s", path)
@@ -3013,7 +2975,7 @@ def update_bootstrap(config, url=None):
# in last case, assuming we got a script content
else:
script_content = url
- script_name = "{}.sh".format(hashlib.sha1(script_content).hexdigest())
+ script_name = f"{hashlib.sha1(script_content).hexdigest()}.sh"
if not script_content:
raise ValueError("No content in bootstrap script !")
@@ -3107,7 +3069,7 @@ def cache_node_list(nodes, provider, opts):
for node in nodes:
diff_node_cache(prov_dir, node, nodes[node], opts)
- path = os.path.join(prov_dir, "{}.p".format(node))
+ path = os.path.join(prov_dir, f"{node}.p")
with salt.utils.files.fopen(path, "wb") as fh_:
salt.utils.msgpack.dump(nodes[node], fh_, encoding=MSGPACK_ENCODING)
@@ -3162,7 +3124,7 @@ def missing_node_cache(prov_dir, node_list, provider, opts):
fire_event(
"event",
"cached node missing from provider",
- "salt/cloud/{}/cache_node_missing".format(node),
+ f"salt/cloud/{node}/cache_node_missing",
args={"missing node": node},
sock_dir=opts.get(
"sock_dir", os.path.join(__opts__["sock_dir"], "master")
@@ -3190,7 +3152,7 @@ def diff_node_cache(prov_dir, node, new_data, opts):
if node is None:
return
- path = "{}.p".format(os.path.join(prov_dir, node))
+ path = f"{os.path.join(prov_dir, node)}.p"
if not os.path.exists(path):
event_data = _strip_cache_events(new_data, opts)
@@ -3198,7 +3160,7 @@ def diff_node_cache(prov_dir, node, new_data, opts):
fire_event(
"event",
"new node found",
- "salt/cloud/{}/cache_node_new".format(node),
+ f"salt/cloud/{node}/cache_node_new",
args={"new_data": event_data},
sock_dir=opts.get("sock_dir", os.path.join(__opts__["sock_dir"], "master")),
transport=opts.get("transport", "zeromq"),
@@ -3222,7 +3184,7 @@ def diff_node_cache(prov_dir, node, new_data, opts):
fire_event(
"event",
"node data differs",
- "salt/cloud/{}/cache_node_diff".format(node),
+ f"salt/cloud/{node}/cache_node_diff",
args={
"new_data": _strip_cache_events(new_data, opts),
"cache_data": _strip_cache_events(cache_data, opts),
@@ -3266,7 +3228,7 @@ def _salt_cloud_force_ascii(exc):
errors.
"""
if not isinstance(exc, (UnicodeEncodeError, UnicodeTranslateError)):
- raise TypeError("Can't handle {}".format(exc))
+ raise TypeError(f"Can't handle {exc}")
unicode_trans = {
# Convert non-breaking space to space
@@ -3326,7 +3288,7 @@ def store_password_in_keyring(credential_id, username, password=None):
# pylint: enable=import-error
if password is None:
- prompt = "Please enter password for {}: ".format(credential_id)
+ prompt = f"Please enter password for {credential_id}: "
try:
password = getpass.getpass(prompt)
except EOFError:
diff --git a/salt/utils/http.py b/salt/utils/http.py
index 91c5cbf08ed..26f2e85c2ee 100644
--- a/salt/utils/http.py
+++ b/salt/utils/http.py
@@ -5,7 +5,7 @@ and the like, but also useful for basic HTTP testing.
.. versionadded:: 2015.5.0
"""
-import cgi
+import email.message
import gzip
import http.client
import http.cookiejar
@@ -84,7 +84,7 @@ except ImportError:
HAS_CERTIFI = False
log = logging.getLogger(__name__)
-USERAGENT = "Salt/{}".format(salt.version.__version__)
+USERAGENT = f"Salt/{salt.version.__version__}"
def __decompressContent(coding, pgctnt):
@@ -170,7 +170,7 @@ def query(
formdata_fieldname=None,
formdata_filename=None,
decode_body=True,
- **kwargs
+ **kwargs,
):
"""
Query a resource, and decode the return data
@@ -295,7 +295,7 @@ def query(
auth = (username, password)
if agent == USERAGENT:
- agent = "{} http.query()".format(agent)
+ agent = f"{agent} http.query()"
header_dict["User-agent"] = agent
if backend == "requests":
@@ -360,14 +360,14 @@ def query(
url,
params=params,
files={formdata_fieldname: (formdata_filename, io.StringIO(data))},
- **req_kwargs
+ **req_kwargs,
)
else:
result = sess.request(method, url, params=params, data=data, **req_kwargs)
result.raise_for_status()
if stream is True:
# fake a HTTP response header
- header_callback("HTTP/1.0 {} MESSAGE".format(result.status_code))
+ header_callback(f"HTTP/1.0 {result.status_code} MESSAGE")
# fake streaming the content
streaming_callback(result.content)
return {
@@ -483,15 +483,12 @@ def query(
result_headers = dict(result.info())
result_text = result.read()
if "Content-Type" in result_headers:
- res_content_type, res_params = cgi.parse_header(
- result_headers["Content-Type"]
- )
- if (
- res_content_type.startswith("text/")
- and "charset" in res_params
- and not isinstance(result_text, str)
- ):
- result_text = result_text.decode(res_params["charset"])
+ msg = email.message.EmailMessage()
+ msg.add_header("Content-Type", result_headers["Content-Type"])
+ if msg.get_content_type().startswith("text/"):
+ content_charset = msg.get_content_charset()
+ if content_charset and not isinstance(result_text, str):
+ result_text = result_text.decode(content_charset)
if isinstance(result_text, bytes) and decode_body:
result_text = result_text.decode("utf-8")
ret["body"] = result_text
@@ -636,15 +633,12 @@ def query(
result_headers = result.headers
result_text = result.body
if "Content-Type" in result_headers:
- res_content_type, res_params = cgi.parse_header(
- result_headers["Content-Type"]
- )
- if (
- res_content_type.startswith("text/")
- and "charset" in res_params
- and not isinstance(result_text, str)
- ):
- result_text = result_text.decode(res_params["charset"])
+ msg = email.message.EmailMessage()
+ msg.add_header("Content-Type", result_headers["Content-Type"])
+ if msg.get_content_type().startswith("text/"):
+ content_charset = msg.get_content_charset()
+ if content_charset and not isinstance(result_text, str):
+ result_text = result_text.decode(content_charset)
if isinstance(result_text, bytes) and decode_body:
result_text = result_text.decode("utf-8")
ret["body"] = result_text
@@ -1038,12 +1032,12 @@ def _sanitize_url_components(comp_list, field):
"""
if not comp_list:
return ""
- elif comp_list[0].startswith("{}=".format(field)):
- ret = "{}=XXXXXXXXXX&".format(field)
+ elif comp_list[0].startswith(f"{field}="):
+ ret = f"{field}=XXXXXXXXXX&"
comp_list.remove(comp_list[0])
return ret + _sanitize_url_components(comp_list, field)
else:
- ret = "{}&".format(comp_list[0])
+ ret = f"{comp_list[0]}&"
comp_list.remove(comp_list[0])
return ret + _sanitize_url_components(comp_list, field)
diff --git a/salt/utils/jinja.py b/salt/utils/jinja.py
index a6a8a279605..d90957a0087 100644
--- a/salt/utils/jinja.py
+++ b/salt/utils/jinja.py
@@ -2,13 +2,12 @@
Jinja loading utils to enable a more powerful backend for jinja templates
"""
-
import itertools
import logging
import os.path
-import pipes
import pprint
import re
+import shlex
import time
import uuid
import warnings
@@ -242,11 +241,11 @@ class PrintableDict(OrderedDict):
if isinstance(value, str):
# keeps quotes around strings
# pylint: disable=repr-flag-used-in-string
- output.append("{!r}: {!r}".format(key, value))
+ output.append(f"{key!r}: {value!r}")
# pylint: enable=repr-flag-used-in-string
else:
# let default output
- output.append("{!r}: {!s}".format(key, value))
+ output.append(f"{key!r}: {value!s}")
return "{" + ", ".join(output) + "}"
def __repr__(self): # pylint: disable=W0221
@@ -255,7 +254,7 @@ class PrintableDict(OrderedDict):
# Raw string formatter required here because this is a repr
# function.
# pylint: disable=repr-flag-used-in-string
- output.append("{!r}: {!r}".format(key, value))
+ output.append(f"{key!r}: {value!r}")
# pylint: enable=repr-flag-used-in-string
return "{" + ", ".join(output) + "}"
@@ -441,7 +440,7 @@ def quote(txt):
'my_text'
"""
- return pipes.quote(txt)
+ return shlex.quote(txt)
@jinja_filter()
@@ -1095,13 +1094,13 @@ class SerializerExtension(Extension):
# to the stringified version of the exception.
msg += str(exc)
else:
- msg += "{}\n".format(problem)
+ msg += f"{problem}\n"
msg += salt.utils.stringutils.get_context(
buf, line, marker=" <======================"
)
raise TemplateRuntimeError(msg)
except AttributeError:
- raise TemplateRuntimeError("Unable to load yaml from {}".format(value))
+ raise TemplateRuntimeError(f"Unable to load yaml from {value}")
def load_json(self, value):
if isinstance(value, TemplateModule):
@@ -1109,7 +1108,7 @@ class SerializerExtension(Extension):
try:
return salt.utils.json.loads(value)
except (ValueError, TypeError, AttributeError):
- raise TemplateRuntimeError("Unable to load json from {}".format(value))
+ raise TemplateRuntimeError(f"Unable to load json from {value}")
def load_text(self, value):
if isinstance(value, TemplateModule):
@@ -1144,7 +1143,7 @@ class SerializerExtension(Extension):
return self._parse_profile_block(parser, label, "profile block", body, lineno)
def _create_profile_id(self, parser):
- return "_salt_profile_{}".format(parser.free_identifier().name)
+ return f"_salt_profile_{parser.free_identifier().name}"
def _profile_start(self, label, source):
return (label, source, time.time())
@@ -1186,7 +1185,7 @@ class SerializerExtension(Extension):
filter_name = parser.stream.current.value
lineno = next(parser.stream).lineno
if filter_name not in self.environment.filters:
- parser.fail("Unable to parse {}".format(filter_name), lineno)
+ parser.fail(f"Unable to parse {filter_name}", lineno)
parser.stream.expect("name:as")
target = parser.parse_assign_target()
@@ -1225,7 +1224,7 @@ class SerializerExtension(Extension):
nodes.Name(target, "store").set_lineno(lineno),
nodes.Filter(
nodes.Name(target, "load").set_lineno(lineno),
- "load_{}".format(converter),
+ f"load_{converter}",
[],
[],
None,
@@ -1234,7 +1233,7 @@ class SerializerExtension(Extension):
).set_lineno(lineno),
]
return self._parse_profile_block(
- parser, import_node.template, "import_{}".format(converter), body, lineno
+ parser, import_node.template, f"import_{converter}", body, lineno
)
def dict_to_sls_yaml_params(self, value, flow_style=False):
diff --git a/salt/utils/locales.py b/salt/utils/locales.py
index 8017958d5de..a380ddbe7a2 100644
--- a/salt/utils/locales.py
+++ b/salt/utils/locales.py
@@ -1,8 +1,7 @@
"""
the locale utils used by salt
"""
-
-
+import locale
import sys
from salt.utils.decorators import memoize as real_memoize
@@ -83,3 +82,39 @@ def normalize_locale(loc):
comps["codeset"] = comps["codeset"].lower().replace("-", "")
comps["charmap"] = ""
return join_locale(comps)
+
+
+def getdefaultlocale(envvars=("LC_ALL", "LC_CTYPE", "LANG", "LANGUAGE")):
+ """
+ This function was backported from Py3.11 which started triggering a
+ deprecation warning about it's removal in 3.13.
+ """
+ try:
+ # check if it's supported by the _locale module
+ import _locale
+
+ code, encoding = _locale._getdefaultlocale()
+ except (ImportError, AttributeError):
+ pass
+ else:
+ # make sure the code/encoding values are valid
+ if sys.platform == "win32" and code and code[:2] == "0x":
+ # map windows language identifier to language name
+ code = locale.windows_locale.get(int(code, 0))
+ # ...add other platform-specific processing here, if
+ # necessary...
+ return code, encoding
+
+ # fall back on POSIX behaviour
+ import os
+
+ lookup = os.environ.get
+ for variable in envvars:
+ localename = lookup(variable, None)
+ if localename:
+ if variable == "LANGUAGE":
+ localename = localename.split(":")[0]
+ break
+ else:
+ localename = "C"
+ return locale._parse_localename(localename)
diff --git a/tests/integration/states/test_ssh_auth.py b/tests/integration/states/test_ssh_auth.py
index 660c3f62d6a..46ffc9b4115 100644
--- a/tests/integration/states/test_ssh_auth.py
+++ b/tests/integration/states/test_ssh_auth.py
@@ -24,6 +24,20 @@ class SSHAuthStateTests(ModuleCase, SaltReturnAssertsMixin):
user_ssh_dir = os.path.join(userdetails["home"], ".ssh")
authorized_keys_file = os.path.join(user_ssh_dir, "authorized_keys")
+ key1 = (
+ # Explicit no ending line break
+ "ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAAgQC3dd5ACsvJhnIOrn6bSOkX5"
+ "KyVDpTYsVAaJj3AmEo6Fr5cHXJFJoJS+Ld8K5vCscPzuXashdYUdrhL1E5Liz"
+ "bza+zneQ5AkJ7sn2NXymD6Bbra+infO4NgnQXbGMp/NyY65jbQGqJeQ081iEV"
+ f"YbDP2zXp6fmrqqmFCaakZfGRbVw== root"
+ )
+ key2 = (
+ "AAAAB3NzaC1yc2EAAAADAQABAAAAgQC7h77HyBPCUDONCs5bI/PrrPwyYJegl0"
+ "f9YWLaBofVYOUl/uSv1ux8zjIoLVs4kguY1ihtIoK2kho4YsjNtIaAd6twdua9"
+ "oqCg2g/54cIK/8WbIjwnb3LFRgyTG5DFuj+7526EdJycAZvhSzIZYui3RUj4Vp"
+ "eMoF7mcB6TIK2/2w=="
+ )
+
ret = self.run_state(
"file.managed",
name=authorized_keys_file,
@@ -31,23 +45,22 @@ class SSHAuthStateTests(ModuleCase, SaltReturnAssertsMixin):
makedirs=True,
contents_newline=False,
# Explicit no ending line break
- contents="ssh-rsa AAAAB3NzaC1kc3MAAACBAL0sQ9fJ5bYTEyY== root",
+ contents=key1,
)
ret = self.run_state(
"ssh_auth.present",
- name="AAAAB3NzaC1kcQ9J5bYTEyZ==",
+ name=key2,
enc="ssh-rsa",
user=username,
comment=username,
)
self.assertSaltTrueReturn(ret)
- self.assertSaltStateChangesEqual(ret, {"AAAAB3NzaC1kcQ9J5bYTEyZ==": "New"})
+ self.assertSaltStateChangesEqual(ret, {key2: "New"})
with salt.utils.files.fopen(authorized_keys_file, "r") as fhr:
self.assertEqual(
fhr.read(),
- "ssh-rsa AAAAB3NzaC1kc3MAAACBAL0sQ9fJ5bYTEyY== root\n"
- "ssh-rsa AAAAB3NzaC1kcQ9J5bYTEyZ== {}\n".format(username),
+ f"{key1}\nssh-rsa {key2} {username}\n",
)
@pytest.mark.destructive_test
@@ -60,39 +73,48 @@ class SSHAuthStateTests(ModuleCase, SaltReturnAssertsMixin):
authorized_keys_file = os.path.join(user_ssh_dir, "authorized_keys")
key_fname = "issue_10198.id_rsa.pub"
+ key_contents = (
+ "ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAAgQC3dd5ACsvJhnIOrn6bSOkX5"
+ "KyVDpTYsVAaJj3AmEo6Fr5cHXJFJoJS+Ld8K5vCscPzuXashdYUdrhL1E5Liz"
+ "bza+zneQ5AkJ7sn2NXymD6Bbra+infO4NgnQXbGMp/NyY65jbQGqJeQ081iEV"
+ f"YbDP2zXp6fmrqqmFCaakZfGRbVw== {username}\n"
+ )
# Create the keyfile that we expect to get back on the state call
with salt.utils.files.fopen(
os.path.join(RUNTIME_VARS.TMP_PRODENV_STATE_TREE, key_fname), "w"
) as kfh:
- kfh.write("ssh-rsa AAAAB3NzaC1kcQ9J5bYTEyZ== {}\n".format(username))
+ kfh.write(key_contents)
# Create a bogus key file on base environment
with salt.utils.files.fopen(
os.path.join(RUNTIME_VARS.TMP_STATE_TREE, key_fname), "w"
) as kfh:
- kfh.write("ssh-rsa BAAAB3NzaC1kcQ9J5bYTEyZ== {}\n".format(username))
+ kfh.write(
+ "ssh-rsa A!AAB3NzaC1yc2EAAAADAQABAAAAgQC3dd5ACsvJhnIOrn6bSOkX5"
+ "KyVDpTYsVAaJj3AmEo6Fr5cHXJFJoJS+Ld8K5vCscPzuXashdYUdrhL1E5Liz"
+ "bza+zneQ5AkJ7sn2NXymD6Bbra+infO4NgnQXbGMp/NyY65jbQGqJeQ081iEV"
+ f"YbDP2zXp6fmrqqmFCaakZfGRbVw== {username}\n"
+ )
ret = self.run_state(
"ssh_auth.present",
name="Setup Keys",
- source="salt://{}?saltenv=prod".format(key_fname),
+ source=f"salt://{key_fname}?saltenv=prod",
enc="ssh-rsa",
user=username,
comment=username,
)
self.assertSaltTrueReturn(ret)
with salt.utils.files.fopen(authorized_keys_file, "r") as fhr:
- self.assertEqual(
- fhr.read(), "ssh-rsa AAAAB3NzaC1kcQ9J5bYTEyZ== {}\n".format(username)
- )
+ self.assertEqual(fhr.read(), key_contents)
os.unlink(authorized_keys_file)
ret = self.run_state(
"ssh_auth.present",
name="Setup Keys",
- source="salt://{}".format(key_fname),
+ source=f"salt://{key_fname}",
enc="ssh-rsa",
user=username,
comment=username,
@@ -100,6 +122,4 @@ class SSHAuthStateTests(ModuleCase, SaltReturnAssertsMixin):
)
self.assertSaltTrueReturn(ret)
with salt.utils.files.fopen(authorized_keys_file, "r") as fhr:
- self.assertEqual(
- fhr.read(), "ssh-rsa AAAAB3NzaC1kcQ9J5bYTEyZ== {}\n".format(username)
- )
+ self.assertEqual(fhr.read(), key_contents)
diff --git a/tests/pytests/unit/modules/state/test_state.py b/tests/pytests/unit/modules/state/test_state.py
index 7c42646bcf7..cff66defa9d 100644
--- a/tests/pytests/unit/modules/state/test_state.py
+++ b/tests/pytests/unit/modules/state/test_state.py
@@ -610,7 +610,7 @@ def test_show_states_missing_sls():
chunks_mock = MagicMock(side_effect=[msg])
mock = MagicMock(side_effect=["A", None])
with patch.object(state, "_check_queue", mock), patch(
- "salt.state.HighState.compile_low_chunks", chunks_mock
+ "salt.modules.state.salt.state.HighState.compile_low_chunks", chunks_mock
):
assert state.show_low_sls("foo") == "A"
assert state.show_states("foo") == [msg[0]]
diff --git a/tests/unit/states/test_module.py b/tests/unit/states/test_module.py
index a705bd30285..4853c24ca07 100644
--- a/tests/unit/states/test_module.py
+++ b/tests/unit/states/test_module.py
@@ -4,7 +4,7 @@
import logging
-from inspect import ArgSpec
+from inspect import FullArgSpec
import salt.states.module as module
from tests.support.mixins import LoaderModuleMockMixin
@@ -117,11 +117,25 @@ class ModuleStateTest(TestCase, LoaderModuleMockMixin):
@classmethod
def setUpClass(cls):
- cls.aspec = ArgSpec(
- args=["hello", "world"], varargs=None, keywords=None, defaults=False
+ cls.aspec = FullArgSpec(
+ args=["hello", "world"],
+ varargs=None,
+ varkw=None,
+ defaults=False,
+ kwonlyargs=None,
+ kwonlydefaults=None,
+ annotations=None,
)
- cls.bspec = ArgSpec(args=[], varargs="names", keywords="kwargs", defaults=None)
+ cls.bspec = FullArgSpec(
+ args=[],
+ varargs="names",
+ varkw=None,
+ defaults=None,
+ kwonlyargs="kwargs",
+ kwonlydefaults=None,
+ annotations=None,
+ )
@classmethod
def tearDownClass(cls):
@@ -137,8 +151,8 @@ class ModuleStateTest(TestCase, LoaderModuleMockMixin):
module.__opts__, {"use_superseded": ["module.run"]}
):
ret = module.run(**{CMD: None})
- if ret["comment"] != "Unavailable function: {}.".format(CMD) or ret["result"]:
- self.fail("module.run did not fail as expected: {}".format(ret))
+ if ret["comment"] != f"Unavailable function: {CMD}." or ret["result"]:
+ self.fail(f"module.run did not fail as expected: {ret}")
def test_run_module_not_available_testmode(self):
"""
@@ -151,10 +165,10 @@ class ModuleStateTest(TestCase, LoaderModuleMockMixin):
):
ret = module.run(**{CMD: None})
if (
- ret["comment"] != "Unavailable function: {}.".format(CMD)
+ ret["comment"] != f"Unavailable function: {CMD}."
or ret["result"] is not False
):
- self.fail("module.run did not fail as expected: {}".format(ret))
+ self.fail(f"module.run did not fail as expected: {ret}")
def test_run_module_noop(self):
"""
@@ -166,7 +180,7 @@ class ModuleStateTest(TestCase, LoaderModuleMockMixin):
):
ret = module.run()
if ret["comment"] != "No function provided." or ret["result"] is not False:
- self.fail("module.run did not fail as expected: {}".format(ret))
+ self.fail(f"module.run did not fail as expected: {ret}")
def test_module_run_hidden_varargs(self):
"""
@@ -189,10 +203,10 @@ class ModuleStateTest(TestCase, LoaderModuleMockMixin):
):
ret = module.run(**{CMD: None})
if (
- ret["comment"] != "Function {} to be executed.".format(CMD)
+ ret["comment"] != f"Function {CMD} to be executed."
or ret["result"] is not None
):
- self.fail("module.run failed: {}".format(ret))
+ self.fail(f"module.run failed: {ret}")
def test_run_missing_arg(self):
"""
@@ -203,9 +217,7 @@ class ModuleStateTest(TestCase, LoaderModuleMockMixin):
module.__opts__, {"use_superseded": ["module.run"]}
):
ret = module.run(**{CMD: None})
- self.assertEqual(
- ret["comment"], "'{}' failed: Missing arguments: name".format(CMD)
- )
+ self.assertEqual(ret["comment"], f"'{CMD}' failed: Missing arguments: name")
def test_run_correct_arg(self):
"""
@@ -216,8 +228,8 @@ class ModuleStateTest(TestCase, LoaderModuleMockMixin):
module.__opts__, {"use_superseded": ["module.run"]}
):
ret = module.run(**{CMD: ["Fred"]})
- if ret["comment"] != "{}: Success".format(CMD) or not ret["result"]:
- self.fail("module.run failed: {}".format(ret))
+ if ret["comment"] != f"{CMD}: Success" or not ret["result"]:
+ self.fail(f"module.run failed: {ret}")
def test_run_state_apply_result_false(self):
"""
@@ -294,9 +306,7 @@ class ModuleStateTest(TestCase, LoaderModuleMockMixin):
):
ret = module.run(**{CMD: ["bla", {"example": "bla"}]})
self.assertFalse(ret["result"])
- self.assertEqual(
- ret["comment"], "'{}' failed: Missing arguments: arg2".format(CMD)
- )
+ self.assertEqual(ret["comment"], f"'{CMD}' failed: Missing arguments: arg2")
def test_run_42270_kwargs_to_args(self):
"""
@@ -390,9 +400,7 @@ class ModuleStateTest(TestCase, LoaderModuleMockMixin):
with patch.dict(module.__salt__, {}, clear=True):
ret = module._legacy_run(CMD)
self.assertFalse(ret["result"])
- self.assertEqual(
- ret["comment"], "Module function {} is not available".format(CMD)
- )
+ self.assertEqual(ret["comment"], f"Module function {CMD} is not available")
def test_module_run_test_true(self):
"""
@@ -400,9 +408,7 @@ class ModuleStateTest(TestCase, LoaderModuleMockMixin):
"""
with patch.dict(module.__opts__, {"test": True}):
ret = module._legacy_run(CMD)
- self.assertEqual(
- ret["comment"], "Module function {} is set to execute".format(CMD)
- )
+ self.assertEqual(ret["comment"], f"Module function {CMD} is set to execute")
def test_module_run_missing_arg(self):
"""
diff --git a/tests/unit/test_master.py b/tests/unit/test_master.py
index b454882f06c..96fe2a54595 100644
--- a/tests/unit/test_master.py
+++ b/tests/unit/test_master.py
@@ -56,6 +56,7 @@ class TransportMethodsTest(TestCase):
"__format__",
"__ge__",
"__getattribute__",
+ "__getstate__",
"__gt__",
"__hash__",
"__init__",
@@ -71,9 +72,9 @@ class TransportMethodsTest(TestCase):
"__sizeof__",
"__str__",
"__subclasshook__",
+ "destroy",
"get_method",
"run_func",
- "destroy",
]
for name in dir(aes_funcs):
if name in aes_funcs.expose_methods:
@@ -108,6 +109,7 @@ class TransportMethodsTest(TestCase):
"__format__",
"__ge__",
"__getattribute__",
+ "__getstate__",
"__gt__",
"__hash__",
"__init__",
@@ -128,9 +130,9 @@ class TransportMethodsTest(TestCase):
"_prep_pub",
"_send_pub",
"_send_ssh_pub",
- "get_method",
- "destroy",
"connect",
+ "destroy",
+ "get_method",
]
for name in dir(clear_funcs):
if name in clear_funcs.expose_methods:
--
2.44.0