, &c)
:keyword devicetree: device tree to look up devices in (optional)
:type devicetree: :class:`blivet.DeviceTree`
:returns: names of matching devices
:rtype: list of str
parse methods will not have access to a devicetree, while execute
methods will. The devicetree is superior in that it can resolve md
array names and in that it reflects scheduled device removals, but for
normal local disks udev.resolve_devspec should suffice.
"""
full_spec = spec
if not full_spec.startswith("/dev/"):
full_spec = os.path.normpath("/dev/" + full_spec)
# the regular case
matches = udev.resolve_glob(full_spec)
# Use spec here instead of full_spec to preserve the spec and let the
# called code decide whether to treat the spec as a path instead of a name.
if devicetree is None:
dev = udev.resolve_devspec(spec)
else:
dev = getattr(devicetree.resolveDevice(spec), "name", None)
# udev.resolve_devspec returns None if there's no match, but we don't
# want that ending up in the list.
if dev and dev not in matches:
matches.append(dev)
return matches
def lookupAlias(devicetree, alias):
for dev in devicetree.devices:
if getattr(dev, "req_name", None) == alias:
return dev
return None
# Remove any existing formatting on a device, but do not remove the partition
# itself. This sets up an existing device to be used in a --onpart option.
def removeExistingFormat(device, storage):
deps = storage.deviceDeps(device)
while deps:
leaves = [d for d in deps if d.isleaf]
for leaf in leaves:
storage.destroyDevice(leaf)
deps.remove(leaf)
storage.devicetree.registerAction(ActionDestroyFormat(device))
def getAvailableDiskSpace(storage):
"""
Get overall disk space available on disks we may use.
:param storage: blivet.Blivet instance
:return: overall disk space available
:rtype: :class:`blivet.size.Size`
"""
free_space = storage.freeSpaceSnapshot
# blivet creates a new free space dict to instead of modifying the old one,
# so there is no worry about the dictionary changing during iteration.
return sum(disk_free for disk_free, fs_free in free_space.values())
def refreshAutoSwapSize(storage):
"""
Refresh size of the auto partitioning request for swap device according to
the current state of the storage configuration.
:param storage: blivet.Blivet instance
"""
for request in storage.autoPartitionRequests:
if request.fstype == "swap":
disk_space = getAvailableDiskSpace(storage)
request.size = autopart.swapSuggestion(disk_space=disk_space)
break
###
### SUBCLASSES OF PYKICKSTART COMMAND HANDLERS
###
class Authconfig(commands.authconfig.FC3_Authconfig):
def __init__(self, *args, **kwargs):
commands.authconfig.FC3_Authconfig.__init__(self, *args, **kwargs)
self.packages = []
def setup(self):
if self.seen:
self.packages = ["authconfig"]
def execute(self, *args):
cmd = "/usr/sbin/authconfig"
if not os.path.lexists(iutil.getSysroot()+cmd):
if flags.automatedInstall and self.seen:
msg = _("%s is missing. Cannot setup authentication.") % cmd
raise KickstartError(msg)
else:
return
args = ["--update", "--nostart"] + shlex.split(self.authconfig)
if not flags.automatedInstall and \
(os.path.exists(iutil.getSysroot() + "/lib64/security/pam_fprintd.so") or \
os.path.exists(iutil.getSysroot() + "/lib/security/pam_fprintd.so")):
args += ["--enablefingerprint"]
try:
iutil.execInSysroot(cmd, args)
except RuntimeError as msg:
log.error("Error running %s %s: %s", cmd, args, msg)
class AutoPart(commands.autopart.F21_AutoPart):
def __init__(self, writePriority=100, *args, **kwargs):
if 'encrypted' not in kwargs:
kwargs['encrypted'] = True
super(AutoPart, self).__init__(writePriority=writePriority, *args, **kwargs)
def parse(self, args):
retval = commands.autopart.F21_AutoPart.parse(self, args)
if self.fstype:
fmt = blivet.formats.getFormat(self.fstype)
if not fmt or fmt.type is None:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("autopart fstype of %s is invalid.") % self.fstype))
return retval
def execute(self, storage, ksdata, instClass):
from blivet.autopart import doAutoPartition
from pyanaconda.storage_utils import sanity_check
if not self.autopart:
return
if self.fstype:
try:
storage.setDefaultFSType(self.fstype)
storage.setDefaultBootFSType(self.fstype)
except ValueError:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("Settings default fstype to %s failed.") % self.fstype))
# sets up default autopartitioning. use clearpart separately
# if you want it
instClass.setDefaultPartitioning(storage)
storage.doAutoPart = True
if self.encrypted:
storage.encryptedAutoPart = True
storage.encryptionPassphrase = self.passphrase
storage.encryptionCipher = self.cipher
storage.autoPartEscrowCert = getEscrowCertificate(storage.escrowCertificates, self.escrowcert)
storage.autoPartAddBackupPassphrase = self.backuppassphrase
if self.type is not None:
storage.autoPartType = self.type
doAutoPartition(storage, ksdata, min_luks_entropy=MIN_CREATE_ENTROPY)
errors = sanity_check(storage)
if errors:
raise PartitioningError("autopart failed:\n" + "\n".join(str(error) for error in errors))
class Bootloader(commands.bootloader.F21_Bootloader):
def __init__(self, *args, **kwargs):
commands.bootloader.F21_Bootloader.__init__(self, *args, **kwargs)
self.location = "mbr"
def parse(self, args):
commands.bootloader.F21_Bootloader.parse(self, args)
if self.location == "partition" and isinstance(get_bootloader(), GRUB2):
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("GRUB2 does not support installation to a partition.")))
if self.isCrypted and isinstance(get_bootloader(), GRUB2):
if not self.password.startswith("grub.pbkdf2."):
raise KickstartValueError(formatErrorMsg(self.lineno,
msg="GRUB2 encrypted password must be in grub.pbkdf2 format."))
return self
def execute(self, storage, ksdata, instClass):
if flags.imageInstall and blivet.arch.isS390():
self.location = "none"
if self.location == "none":
location = None
elif self.location == "partition":
location = "boot"
else:
location = self.location
if not location:
storage.bootloader.skip_bootloader = True
return
if self.appendLine:
args = self.appendLine.split()
storage.bootloader.boot_args.update(args)
if self.password:
if self.isCrypted:
storage.bootloader.encrypted_password = self.password
else:
storage.bootloader.password = self.password
if location:
storage.bootloader.set_preferred_stage1_type(location)
if self.timeout is not None:
storage.bootloader.timeout = self.timeout
# Throw out drives specified that don't exist or cannot be used (iSCSI
# device on an s390 machine)
disk_names = [d.name for d in storage.disks
if not d.format.hidden and not d.protected and
(not blivet.arch.isS390() or not isinstance(d, blivet.devices.iScsiDiskDevice))]
diskSet = set(disk_names)
for drive in self.driveorder[:]:
matches = set(deviceMatches(drive, devicetree=storage.devicetree))
if matches.isdisjoint(diskSet):
log.warning("requested drive %s in boot drive order doesn't exist or cannot be used", drive)
self.driveorder.remove(drive)
storage.bootloader.disk_order = self.driveorder
if self.bootDrive:
matches = set(deviceMatches(self.bootDrive,
devicetree=storage.devicetree))
if len(matches) > 1:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("More than one match found for given boot drive \"%s\".") % self.bootDrive))
elif matches.isdisjoint(diskSet):
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("Requested boot drive \"%s\" doesn't exist or cannot be used.") % self.bootDrive))
else:
self.bootDrive = disk_names[0]
drive = storage.devicetree.resolveDevice(self.bootDrive)
storage.bootloader.stage1_disk = drive
if self.leavebootorder:
flags.leavebootorder = True
if self.nombr:
flags.nombr = True
class BTRFS(commands.btrfs.F23_BTRFS):
def execute(self, storage, ksdata, instClass):
for b in self.btrfsList:
b.execute(storage, ksdata, instClass)
class BTRFSData(commands.btrfs.F23_BTRFSData):
def execute(self, storage, ksdata, instClass):
devicetree = storage.devicetree
storage.doAutoPart = False
members = []
# Get a list of all the devices that make up this volume.
for member in self.devices:
dev = devicetree.resolveDevice(member)
if not dev:
# if using --onpart, use original device
member_name = ksdata.onPart.get(member, member)
dev = devicetree.resolveDevice(member_name) or lookupAlias(devicetree, member)
if dev and dev.format.type == "luks":
try:
dev = devicetree.getChildren(dev)[0]
except IndexError:
dev = None
if dev and dev.format.type != "btrfs":
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("Btrfs partition \"%(device)s\" has a format of \"%(format)s\", but should have a format of \"btrfs\".") %
{"device": member, "format": dev.format.type}))
if not dev:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("Tried to use undefined partition \"%s\" in Btrfs volume specification.") % member))
members.append(dev)
if self.subvol:
name = self.name
elif self.label:
name = self.label
else:
name = None
if len(members) == 0 and not self.preexist:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("Btrfs volume defined without any member devices. Either specify member devices or use --useexisting.")))
# allow creating btrfs vols/subvols without specifying mountpoint
if self.mountpoint in ("none", "None"):
self.mountpoint = ""
# Sanity check mountpoint
if self.mountpoint != "" and self.mountpoint[0] != '/':
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("The mount point \"%s\" is not valid. It must start with a /.") % self.mountpoint))
# If a previous device has claimed this mount point, delete the
# old one.
try:
if self.mountpoint:
device = storage.mountpoints[self.mountpoint]
storage.destroyDevice(device)
except KeyError:
pass
if self.preexist:
device = devicetree.resolveDevice(self.name)
if not device:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("Btrfs volume \"%s\" specified with --useexisting does not exist.") % self.name))
device.format.mountpoint = self.mountpoint
else:
try:
request = storage.newBTRFS(name=name,
subvol=self.subvol,
mountpoint=self.mountpoint,
metaDataLevel=self.metaDataLevel,
dataLevel=self.dataLevel,
parents=members,
createOptions=self.mkfsopts)
except BTRFSValueError as e:
raise KickstartValueError(formatErrorMsg(self.lineno, msg=str(e)))
storage.createDevice(request)
class Realm(commands.realm.F19_Realm):
def __init__(self, *args):
commands.realm.F19_Realm.__init__(self, *args)
self.packages = []
self.discovered = ""
def setup(self):
if not self.join_realm:
return
try:
argv = ["discover", "--verbose"] + \
self.discover_options + [self.join_realm]
output = iutil.execWithCapture("realm", argv, filter_stderr=True)
except OSError:
# TODO: A lousy way of propagating what will usually be
# 'no such realm'
# The error message is logged by iutil
return
# Now parse the output for the required software. First line is the
# realm name, and following lines are information as "name: value"
self.packages = ["realmd"]
self.discovered = ""
lines = output.split("\n")
if not lines:
return
self.discovered = lines.pop(0).strip()
log.info("Realm discovered: %s", self.discovered)
for line in lines:
parts = line.split(":", 1)
if len(parts) == 2 and parts[0].strip() == "required-package":
self.packages.append(parts[1].strip())
log.info("Realm %s needs packages %s",
self.discovered, ", ".join(self.packages))
def execute(self, *args):
if not self.discovered:
return
for arg in self.join_args:
if arg.startswith("--no-password") or arg.startswith("--one-time-password"):
pw_args = []
break
else:
# no explicit password arg using implicit --no-password
pw_args = ["--no-password"]
argv = ["join", "--install", iutil.getSysroot(), "--verbose"] + \
pw_args + self.join_args
rc = -1
try:
rc = iutil.execWithRedirect("realm", argv)
except OSError:
pass
if rc == 0:
log.info("Joined realm %s", self.join_realm)
class ClearPart(commands.clearpart.F21_ClearPart):
def parse(self, args):
retval = commands.clearpart.F21_ClearPart.parse(self, args)
if self.type is None:
self.type = CLEARPART_TYPE_NONE
if self.disklabel and self.disklabel not in platform.diskLabelTypes:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("Disklabel \"%s\" given in clearpart command is not "
"supported on this platform.") % self.disklabel))
# Do any glob expansion now, since we need to have the real list of
# disks available before the execute methods run.
drives = []
for spec in self.drives:
matched = deviceMatches(spec)
if matched:
drives.extend(matched)
else:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("Disk \"%s\" given in clearpart command does not exist.") % spec))
self.drives = drives
# Do any glob expansion now, since we need to have the real list of
# devices available before the execute methods run.
devices = []
for spec in self.devices:
matched = deviceMatches(spec)
if matched:
devices.extend(matched)
else:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("Device \"%s\" given in clearpart device list does not exist.") % spec))
self.devices = devices
return retval
def execute(self, storage, ksdata, instClass):
storage.config.clearPartType = self.type
storage.config.clearPartDisks = self.drives
storage.config.clearPartDevices = self.devices
if self.initAll:
storage.config.initializeDisks = self.initAll
if self.disklabel:
if not platform.setDefaultDiskLabelType(self.disklabel):
log.warn("%s is not a supported disklabel type on this platform. "
"Using default disklabel %s instead.", self.disklabel, platform.defaultDiskLabelType)
storage.clearPartitions()
class Fcoe(commands.fcoe.F13_Fcoe):
def parse(self, args):
fc = commands.fcoe.F13_Fcoe.parse(self, args)
if fc.nic not in nm.nm_devices():
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("NIC \"%s\" given in fcoe command does not exist.") % fc.nic))
if fc.nic in (info[0] for info in blivet.fcoe.fcoe().nics):
log.info("Kickstart fcoe device %s already added from EDD, ignoring", fc.nic)
else:
msg = blivet.fcoe.fcoe().addSan(nic=fc.nic, dcb=fc.dcb, auto_vlan=True)
if not msg:
msg = "Succeeded."
blivet.fcoe.fcoe().added_nics.append(fc.nic)
log.info("adding FCoE SAN on %s: %s", fc.nic, msg)
return fc
class Firewall(commands.firewall.F20_Firewall):
def __init__(self, *args, **kwargs):
commands.firewall.F20_Firewall.__init__(self, *args, **kwargs)
self.packages = []
def setup(self):
if self.seen:
self.packages = ["firewalld"]
def execute(self, storage, ksdata, instClass):
args = []
# enabled is None if neither --enable or --disable is passed
# default to enabled if nothing has been set.
if self.enabled == False:
args += ["--disabled"]
else:
args += ["--enabled"]
if "ssh" not in self.services and "ssh" not in self.remove_services \
and "22:tcp" not in self.ports:
args += ["--service=ssh"]
for dev in self.trusts:
args += ["--trust=%s" % (dev,)]
for port in self.ports:
args += ["--port=%s" % (port,)]
for remove_service in self.remove_services:
args += ["--remove-service=%s" % (remove_service,)]
for service in self.services:
args += ["--service=%s" % (service,)]
cmd = "/usr/bin/firewall-offline-cmd"
if not os.path.exists(iutil.getSysroot()+cmd):
if self.enabled:
msg = _("%s is missing. Cannot setup firewall.") % (cmd,)
raise KickstartError(msg)
else:
iutil.execInSysroot(cmd, args)
class Firstboot(commands.firstboot.FC3_Firstboot):
def setup(self, *args):
# firstboot should be disabled by default after kickstart installations
if flags.automatedInstall and not self.seen:
self.firstboot = FIRSTBOOT_SKIP
def execute(self, *args):
action = "enable"
services = ["initial-setup-graphical.service",
"initial-setup-text.service"]
# find if the unit files for the Initial Setup services are installed
services = [name for name in services if os.path.exists(os.path.join(iutil.getSysroot(), "lib/systemd/system/", name))]
if services and self.firstboot == FIRSTBOOT_RECONFIG:
# write the reconfig trigger file
f = open(os.path.join(iutil.getSysroot(), "/etc/reconfigSys"), "w+")
f.close()
if self.firstboot == FIRSTBOOT_SKIP:
action = "disable"
# enable/disable all installed Initial Setup services
for service in services:
iutil.execInSysroot("systemctl", [action, service])
class Group(commands.group.F12_Group):
def execute(self, storage, ksdata, instClass, users):
for grp in self.groupList:
kwargs = grp.__dict__
kwargs.update({"root": iutil.getSysroot()})
users.createGroup(grp.name, **kwargs)
class IgnoreDisk(commands.ignoredisk.RHEL6_IgnoreDisk):
def parse(self, args):
retval = commands.ignoredisk.RHEL6_IgnoreDisk.parse(self, args)
# See comment in ClearPart.parse
drives = []
for spec in self.ignoredisk:
matched = deviceMatches(spec)
if matched:
drives.extend(matched)
else:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("Disk \"%s\" given in ignoredisk command does not exist.") % spec))
self.ignoredisk = drives
drives = []
for spec in self.onlyuse:
matched = deviceMatches(spec)
if matched:
drives.extend(matched)
else:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("Disk \"%s\" given in ignoredisk command does not exist.") % spec))
self.onlyuse = drives
return retval
class Iscsi(commands.iscsi.F17_Iscsi):
def parse(self, args):
tg = commands.iscsi.F17_Iscsi.parse(self, args)
if tg.iface:
if not network.wait_for_network_devices([tg.iface]):
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("Network interface \"%(nic)s\" required by iSCSI \"%(iscsiTarget)s\" target is not up.") %
{"nic": tg.iface, "iscsiTarget": tg.target}))
mode = blivet.iscsi.iscsi().mode
if mode == "none":
if tg.iface:
blivet.iscsi.iscsi().create_interfaces(nm.nm_activated_devices())
elif ((mode == "bind" and not tg.iface)
or (mode == "default" and tg.iface)):
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("iscsi --iface must be specified (binding used) either for all targets or for none")))
try:
blivet.iscsi.iscsi().addTarget(tg.ipaddr, tg.port, tg.user,
tg.password, tg.user_in,
tg.password_in,
target=tg.target,
iface=tg.iface)
log.info("added iscsi target %s at %s via %s", tg.target,
tg.ipaddr,
tg.iface)
except (IOError, ValueError) as e:
raise KickstartValueError(formatErrorMsg(self.lineno, msg=str(e)))
return tg
class IscsiName(commands.iscsiname.FC6_IscsiName):
def parse(self, args):
retval = commands.iscsiname.FC6_IscsiName.parse(self, args)
blivet.iscsi.iscsi().initiator = self.iscsiname
return retval
class Lang(commands.lang.F19_Lang):
def execute(self, *args, **kwargs):
localization.write_language_configuration(self, iutil.getSysroot())
# no overrides needed here
Eula = commands.eula.F20_Eula
class LogVol(commands.logvol.F23_LogVol):
def execute(self, storage, ksdata, instClass):
for l in self.lvList:
l.execute(storage, ksdata, instClass)
if self.lvList:
growLVM(storage)
class LogVolData(commands.logvol.F23_LogVolData):
def execute(self, storage, ksdata, instClass):
devicetree = storage.devicetree
storage.doAutoPart = False
# FIXME: we should be running sanityCheck on partitioning that is not ks
# autopart, but that's likely too invasive for #873135 at this moment
if self.mountpoint == "/boot" and blivet.arch.isS390():
raise KickstartValueError(formatErrorMsg(self.lineno, msg="/boot can not be of type 'lvmlv' on s390x"))
# we might have truncated or otherwise changed the specified vg name
vgname = ksdata.onPart.get(self.vgname, self.vgname)
size = None
if self.percent:
size = Size(0)
if self.mountpoint == "swap":
ty = "swap"
self.mountpoint = ""
if self.recommended or self.hibernation:
disk_space = getAvailableDiskSpace(storage)
size = autopart.swapSuggestion(hibernation=self.hibernation, disk_space=disk_space)
self.grow = False
else:
if self.fstype != "":
ty = self.fstype
else:
ty = storage.defaultFSType
if size is None and not self.preexist:
if not self.size:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg="Size can not be decided on from kickstart nor obtained from device."))
try:
size = Size("%d MiB" % self.size)
except ValueError:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg="The size \"%s\" is invalid." % self.size))
if self.thin_pool:
self.mountpoint = ""
ty = None
# Sanity check mountpoint
if self.mountpoint != "" and self.mountpoint[0] != '/':
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("The mount point \"%s\" is not valid. It must start with a /.") % self.mountpoint))
# Check that the VG this LV is a member of has already been specified.
vg = devicetree.getDeviceByName(vgname)
if not vg:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("No volume group exists with the name \"%s\". Specify volume groups before logical volumes.") % self.vgname))
pool = None
if self.thin_volume:
pool = devicetree.getDeviceByName("%s-%s" % (vg.name, self.pool_name))
if not pool:
err = formatErrorMsg(self.lineno,
msg=_("No thin pool exists with the name \"%s\". Specify thin pools before thin volumes.") % self.pool_name)
raise KickstartValueError(err)
# If this specifies an existing request that we should not format,
# quit here after setting up enough information to mount it later.
if not self.format:
if not self.name:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("logvol --noformat must also use the --name= option.")))
dev = devicetree.getDeviceByName("%s-%s" % (vg.name, self.name))
if not dev:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("Logical volume \"%s\" given in logvol command does not exist.") % self.name))
if self.resize:
size = dev.raw_device.alignTargetSize(size)
if size < dev.currentSize:
# shrink
try:
devicetree.registerAction(ActionResizeFormat(dev, size))
devicetree.registerAction(ActionResizeDevice(dev, size))
except ValueError:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("Target size \"%(size)s\" for device \"%(device)s\" is invalid.") %
{"size": self.size, "device": dev.name}))
else:
# grow
try:
devicetree.registerAction(ActionResizeDevice(dev, size))
devicetree.registerAction(ActionResizeFormat(dev, size))
except ValueError:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("Target size \"%(size)s\" for device \"%(device)s\" is invalid.") %
{"size": self.size, "device": dev.name}))
dev.format.mountpoint = self.mountpoint
dev.format.mountopts = self.fsopts
if ty == "swap":
storage.addFstabSwap(dev)
return
# Make sure this LV name is not already used in the requested VG.
if not self.preexist:
tmp = devicetree.getDeviceByName("%s-%s" % (vg.name, self.name))
if tmp:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("Logical volume name \"%(logvol)s\" is already in use in volume group \"%(volgroup)s\".") %
{"logvol": self.name, "volgroup": vg.name}))
if not self.percent and size and not self.grow and size < vg.peSize:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("Logical volume size \"%(logvolSize)s\" must be larger than the volume group extent size of \"%(extentSize)s\".") %
{"logvolSize": size, "extentSize": vg.peSize}))
# Now get a format to hold a lot of these extra values.
fmt = getFormat(ty,
mountpoint=self.mountpoint,
label=self.label,
fsprofile=self.fsprofile,
createOptions=self.mkfsopts,
mountopts=self.fsopts)
if not fmt.type and not self.thin_pool:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("The \"%s\" file system type is not supported.") % ty))
add_fstab_swap = None
# If we were given a pre-existing LV to create a filesystem on, we need
# to verify it and its VG exists and then schedule a new format action
# to take place there. Also, we only support a subset of all the
# options on pre-existing LVs.
if self.preexist:
device = devicetree.getDeviceByName("%s-%s" % (vg.name, self.name))
if not device:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("Logical volume \"%s\" given in logvol command does not exist.") % self.name))
removeExistingFormat(device, storage)
if self.resize:
size = device.raw_device.alignTargetSize(size)
try:
devicetree.registerAction(ActionResizeDevice(device, size))
except ValueError:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("Target size \"%(size)s\" for device \"%(device)s\" is invalid.") %
{"size": self.size, "device": device.name}))
devicetree.registerAction(ActionCreateFormat(device, fmt))
if ty == "swap":
add_fstab_swap = device
else:
# If a previous device has claimed this mount point, delete the
# old one.
try:
if self.mountpoint:
device = storage.mountpoints[self.mountpoint]
storage.destroyDevice(device)
except KeyError:
pass
if self.thin_volume:
parents = [pool]
else:
parents = [vg]
pool_args = {}
if self.thin_pool:
if self.profile:
matching = (p for p in KNOWN_THPOOL_PROFILES if p.name == self.profile)
profile = next(matching, None)
if profile:
pool_args["profile"] = profile
else:
log.warning("No matching profile for %s found in LVM configuration", self.profile)
if self.metadata_size:
pool_args["metadatasize"] = Size("%d MiB" % self.metadata_size)
if self.chunk_size:
pool_args["chunksize"] = Size("%d KiB" % self.chunk_size)
if self.maxSizeMB:
try:
maxsize = Size("%d MiB" % self.maxSizeMB)
except ValueError:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg="The maximum size \"%s\" is invalid." % self.maxSizeMB))
else:
maxsize = None
if self.cache_size and self.cache_pvs:
pv_devices = [lookupAlias(devicetree, pv) for pv in self.cache_pvs]
cache_size = Size("%d MiB" % self.cache_size)
cache_mode = self.cache_mode or None
cache_request = LVMCacheRequest(cache_size, pv_devices, cache_mode)
else:
cache_request = None
try:
request = storage.newLV(fmt=fmt,
name=self.name,
parents=parents,
size=size,
thin_pool=self.thin_pool,
thin_volume=self.thin_volume,
grow=self.grow,
maxsize=maxsize,
percent=self.percent,
cacheRequest=cache_request,
**pool_args)
except (StorageError, ValueError) as e:
raise KickstartValueError(formatErrorMsg(self.lineno, msg=str(e)))
storage.createDevice(request)
if ty == "swap":
add_fstab_swap = request
if self.encrypted:
if self.passphrase and not storage.encryptionPassphrase:
storage.encryptionPassphrase = self.passphrase
# try to use the global passphrase if available
# XXX: we require the LV/part with --passphrase to be processed
# before this one to setup the storage.encryptionPassphrase
self.passphrase = self.passphrase or storage.encryptionPassphrase
cert = getEscrowCertificate(storage.escrowCertificates, self.escrowcert)
if self.preexist:
luksformat = fmt
device.format = getFormat("luks", passphrase=self.passphrase, device=device.path,
cipher=self.cipher,
escrow_cert=cert,
add_backup_passphrase=self.backuppassphrase)
luksdev = LUKSDevice("luks%d" % storage.nextID,
fmt=luksformat,
parents=device)
else:
luksformat = request.format
request.format = getFormat("luks", passphrase=self.passphrase,
cipher=self.cipher,
escrow_cert=cert,
add_backup_passphrase=self.backuppassphrase,
min_luks_entropy=MIN_CREATE_ENTROPY)
luksdev = LUKSDevice("luks%d" % storage.nextID,
fmt=luksformat,
parents=request)
if ty == "swap":
# swap is on the LUKS device not on the LUKS' parent device,
# override the info here
add_fstab_swap = luksdev
storage.createDevice(luksdev)
if add_fstab_swap:
storage.addFstabSwap(add_fstab_swap)
class Logging(commands.logging.FC6_Logging):
def execute(self, *args):
if logger.loglevel == DEFAULT_LEVEL:
# not set from the command line
level = logLevelMap[self.level]
logger.loglevel = level
setHandlersLevel(log, level)
setHandlersLevel(storage_log, level)
if logger.remote_syslog == None and len(self.host) > 0:
# not set from the command line, ok to use kickstart
remote_server = self.host
if self.port:
remote_server = "%s:%s" %(self.host, self.port)
logger.updateRemote(remote_server)
class Network(commands.network.F22_Network):
def __init__(self, *args, **kwargs):
commands.network.F22_Network.__init__(self, *args, **kwargs)
self.packages = []
def setup(self):
if network.is_using_team_device():
self.packages = ["teamd"]
def execute(self, storage, ksdata, instClass):
network.write_network_config(storage, ksdata, instClass, iutil.getSysroot())
class MultiPath(commands.multipath.FC6_MultiPath):
def parse(self, args):
raise NotImplementedError(_("The %s kickstart command is not currently supported.") % "multipath")
class DmRaid(commands.dmraid.FC6_DmRaid):
def parse(self, args):
raise NotImplementedError(_("The %s kickstart command is not currently supported.") % "dmraid")
class Partition(commands.partition.F23_Partition):
def execute(self, storage, ksdata, instClass):
for p in self.partitions:
p.execute(storage, ksdata, instClass)
if self.partitions:
doPartitioning(storage)
class PartitionData(commands.partition.F23_PartData):
def execute(self, storage, ksdata, instClass):
devicetree = storage.devicetree
kwargs = {}
storage.doAutoPart = False
if self.onbiosdisk != "":
# eddDict is only modified during storage.reset(), so don't do that
# while executing storage.
for (disk, biosdisk) in storage.eddDict.items():
if "%x" % biosdisk == self.onbiosdisk:
self.disk = disk
break
if not self.disk:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("No disk found for specified BIOS disk \"%s\".") % self.onbiosdisk))
size = None
if self.mountpoint == "swap":
ty = "swap"
self.mountpoint = ""
if self.recommended or self.hibernation:
disk_space = getAvailableDiskSpace(storage)
size = autopart.swapSuggestion(hibernation=self.hibernation, disk_space=disk_space)
self.grow = False
# if people want to specify no mountpoint for some reason, let them
# this is really needed for pSeries boot partitions :(
elif self.mountpoint == "None":
self.mountpoint = ""
if self.fstype:
ty = self.fstype
else:
ty = storage.defaultFSType
elif self.mountpoint == 'appleboot':
ty = "appleboot"
self.mountpoint = ""
elif self.mountpoint == 'prepboot':
ty = "prepboot"
self.mountpoint = ""
elif self.mountpoint == 'biosboot':
ty = "biosboot"
self.mountpoint = ""
elif self.mountpoint.startswith("raid."):
ty = "mdmember"
kwargs["name"] = self.mountpoint
self.mountpoint = ""
if devicetree.getDeviceByName(kwargs["name"]):
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("RAID partition \"%s\" is defined multiple times.") % kwargs["name"]))
if self.onPart:
ksdata.onPart[kwargs["name"]] = self.onPart
elif self.mountpoint.startswith("pv."):
ty = "lvmpv"
kwargs["name"] = self.mountpoint
self.mountpoint = ""
if devicetree.getDeviceByName(kwargs["name"]):
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("PV partition \"%s\" is defined multiple times.") % kwargs["name"]))
if self.onPart:
ksdata.onPart[kwargs["name"]] = self.onPart
elif self.mountpoint.startswith("btrfs."):
ty = "btrfs"
kwargs["name"] = self.mountpoint
self.mountpoint = ""
if devicetree.getDeviceByName(kwargs["name"]):
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("Btrfs partition \"%s\" is defined multiple times.") % kwargs["name"]))
if self.onPart:
ksdata.onPart[kwargs["name"]] = self.onPart
elif self.mountpoint == "/boot/efi":
if blivet.arch.isMactel():
ty = "macefi"
else:
ty = "EFI System Partition"
self.fsopts = "defaults,uid=0,gid=0,umask=077,shortname=winnt"
else:
if self.fstype != "":
ty = self.fstype
elif self.mountpoint == "/boot":
ty = storage.defaultBootFSType
else:
ty = storage.defaultFSType
if not size and self.size:
try:
size = Size("%d MiB" % self.size)
except ValueError:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("The size \"%s\" is invalid.") % self.size))
# If this specified an existing request that we should not format,
# quit here after setting up enough information to mount it later.
if not self.format:
if not self.onPart:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("part --noformat must also use the --onpart option.")))
dev = devicetree.resolveDevice(self.onPart)
if not dev:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("Partition \"%s\" given in part command does not exist.") % self.onPart))
if self.resize:
size = dev.raw_device.alignTargetSize(size)
if size < dev.currentSize:
# shrink
try:
devicetree.registerAction(ActionResizeFormat(dev, size))
devicetree.registerAction(ActionResizeDevice(dev, size))
except ValueError:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("Target size \"%(size)s\" for device \"%(device)s\" is invalid.") %
{"size": self.size, "device": dev.name}))
else:
# grow
try:
devicetree.registerAction(ActionResizeDevice(dev, size))
devicetree.registerAction(ActionResizeFormat(dev, size))
except ValueError:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("Target size \"%(size)s\" for device \"%(device)s\" is invalid.") %
{"size": self.size, "device": dev.name}))
dev.format.mountpoint = self.mountpoint
dev.format.mountopts = self.fsopts
if ty == "swap":
storage.addFstabSwap(dev)
return
# Now get a format to hold a lot of these extra values.
kwargs["fmt"] = getFormat(ty,
mountpoint=self.mountpoint,
label=self.label,
fsprofile=self.fsprofile,
mountopts=self.fsopts,
createOptions=self.mkfsopts,
size=size)
if not kwargs["fmt"].type:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("The \"%s\" file system type is not supported.") % ty))
# If we were given a specific disk to create the partition on, verify
# that it exists first. If it doesn't exist, see if it exists with
# mapper/ on the front. If that doesn't exist either, it's an error.
if self.disk:
disk = devicetree.resolveDevice(self.disk)
# if this is a multipath member promote it to the real mpath
if disk and disk.format.type == "multipath_member":
mpath_device = storage.devicetree.getChildren(disk)[0]
storage_log.info("kickstart: part: promoting %s to %s",
disk.name, mpath_device.name)
disk = mpath_device
if not disk:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("Disk \"%s\" given in part command does not exist.") % self.disk))
if not disk.partitionable:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("Cannot install to unpartitionable device \"%s\".") % self.disk))
should_clear = storage.shouldClear(disk)
if disk and (disk.partitioned or should_clear):
kwargs["parents"] = [disk]
elif disk:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("Disk \"%s\" in part command is not partitioned.") % self.disk))
if not kwargs["parents"]:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("Disk \"%s\" given in part command does not exist.") % self.disk))
kwargs["grow"] = self.grow
kwargs["size"] = size
if self.maxSizeMB:
try:
maxsize = Size("%d MiB" % self.maxSizeMB)
except ValueError:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("The maximum size \"%s\" is invalid.") % self.maxSizeMB))
else:
maxsize = None
kwargs["maxsize"] = maxsize
kwargs["primary"] = self.primOnly
add_fstab_swap = None
# If we were given a pre-existing partition to create a filesystem on,
# we need to verify it exists and then schedule a new format action to
# take place there. Also, we only support a subset of all the options
# on pre-existing partitions.
if self.onPart:
device = devicetree.resolveDevice(self.onPart)
if not device:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("Partition \"%s\" given in part command does not exist.") % self.onPart))
removeExistingFormat(device, storage)
if self.resize:
size = device.raw_device.alignTargetSize(size)
try:
devicetree.registerAction(ActionResizeDevice(device, size))
except ValueError:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("Target size \"%(size)s\" for device \"%(device)s\" is invalid.") %
{"size": self.size, "device": device.name}))
devicetree.registerAction(ActionCreateFormat(device, kwargs["fmt"]))
if ty == "swap":
add_fstab_swap = device
# tmpfs mounts are not disks and don't occupy a disk partition,
# so handle them here
elif self.fstype == "tmpfs":
try:
request = storage.newTmpFS(**kwargs)
except (StorageError, ValueError) as e:
raise KickstartValueError(formatErrorMsg(self.lineno, msg=str(e)))
storage.createDevice(request)
else:
# If a previous device has claimed this mount point, delete the
# old one.
try:
if self.mountpoint:
device = storage.mountpoints[self.mountpoint]
storage.destroyDevice(device)
except KeyError:
pass
try:
request = storage.newPartition(**kwargs)
except (StorageError, ValueError) as e:
raise KickstartValueError(formatErrorMsg(self.lineno, msg=str(e)))
storage.createDevice(request)
if ty == "swap":
add_fstab_swap = request
if self.encrypted:
if self.passphrase and not storage.encryptionPassphrase:
storage.encryptionPassphrase = self.passphrase
# try to use the global passphrase if available
# XXX: we require the LV/part with --passphrase to be processed
# before this one to setup the storage.encryptionPassphrase
self.passphrase = self.passphrase or storage.encryptionPassphrase
cert = getEscrowCertificate(storage.escrowCertificates, self.escrowcert)
if self.onPart:
luksformat = kwargs["fmt"]
device.format = getFormat("luks", passphrase=self.passphrase, device=device.path,
cipher=self.cipher,
escrow_cert=cert,
add_backup_passphrase=self.backuppassphrase,
min_luks_entropy=MIN_CREATE_ENTROPY)
luksdev = LUKSDevice("luks%d" % storage.nextID,
fmt=luksformat,
parents=device)
else:
luksformat = request.format
request.format = getFormat("luks", passphrase=self.passphrase,
cipher=self.cipher,
escrow_cert=cert,
add_backup_passphrase=self.backuppassphrase,
min_luks_entropy=MIN_CREATE_ENTROPY)
luksdev = LUKSDevice("luks%d" % storage.nextID,
fmt=luksformat,
parents=request)
if ty == "swap":
# swap is on the LUKS device not on the LUKS' parent device,
# override the info here
add_fstab_swap = luksdev
storage.createDevice(luksdev)
if add_fstab_swap:
storage.addFstabSwap(add_fstab_swap)
class Raid(commands.raid.F23_Raid):
def execute(self, storage, ksdata, instClass):
for r in self.raidList:
r.execute(storage, ksdata, instClass)
class RaidData(commands.raid.F23_RaidData):
def execute(self, storage, ksdata, instClass):
raidmems = []
devicetree = storage.devicetree
devicename = self.device
if self.preexist:
device = devicetree.resolveDevice(devicename)
if device:
devicename = device.name
kwargs = {}
storage.doAutoPart = False
if self.mountpoint == "swap":
ty = "swap"
self.mountpoint = ""
elif self.mountpoint.startswith("pv."):
ty = "lvmpv"
kwargs["name"] = self.mountpoint
ksdata.onPart[kwargs["name"]] = devicename
if devicetree.getDeviceByName(kwargs["name"]):
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("PV partition \"%s\" is defined multiple times.") % kwargs["name"]))
self.mountpoint = ""
elif self.mountpoint.startswith("btrfs."):
ty = "btrfs"
kwargs["name"] = self.mountpoint
ksdata.onPart[kwargs["name"]] = devicename
if devicetree.getDeviceByName(kwargs["name"]):
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("Btrfs partition \"%s\" is defined multiple times.") % kwargs["name"]))
self.mountpoint = ""
else:
if self.fstype != "":
ty = self.fstype
elif self.mountpoint == "/boot" and \
"mdarray" in storage.bootloader.stage2_device_types:
ty = storage.defaultBootFSType
else:
ty = storage.defaultFSType
# Sanity check mountpoint
if self.mountpoint != "" and self.mountpoint[0] != '/':
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("The mount point \"%s\" is not valid. It must start with a /.") % self.mountpoint))
# If this specifies an existing request that we should not format,
# quit here after setting up enough information to mount it later.
if not self.format:
if not devicename:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("raid --noformat must also use the --device option.")))
dev = devicetree.getDeviceByName(devicename)
if not dev:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("RAID device \"%s\" given in raid command does not exist.") % devicename))
dev.format.mountpoint = self.mountpoint
dev.format.mountopts = self.fsopts
if ty == "swap":
storage.addFstabSwap(dev)
return
# Get a list of all the RAID members.
for member in self.members:
dev = devicetree.resolveDevice(member)
if not dev:
# if member is using --onpart, use original device
mem = ksdata.onPart.get(member, member)
dev = devicetree.resolveDevice(mem) or lookupAlias(devicetree, member)
if dev and dev.format.type == "luks":
try:
dev = devicetree.getChildren(dev)[0]
except IndexError:
dev = None
if dev and dev.format.type != "mdmember":
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("RAID device \"%(device)s\" has a format of \"%(format)s\", but should have a format of \"mdmember\".") %
{"device": member, "format": dev.format.type}))
if not dev:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("Tried to use undefined partition \"%s\" in RAID specification.") % member))
raidmems.append(dev)
# Now get a format to hold a lot of these extra values.
kwargs["fmt"] = getFormat(ty,
label=self.label,
fsprofile=self.fsprofile,
mountpoint=self.mountpoint,
mountopts=self.fsopts,
createOptions=self.mkfsopts)
if not kwargs["fmt"].type:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("The \"%s\" file system type is not supported.") % ty))
kwargs["name"] = devicename
kwargs["level"] = self.level
kwargs["parents"] = raidmems
kwargs["memberDevices"] = len(raidmems) - self.spares
kwargs["totalDevices"] = len(raidmems)
# If we were given a pre-existing RAID to create a filesystem on,
# we need to verify it exists and then schedule a new format action
# to take place there. Also, we only support a subset of all the
# options on pre-existing RAIDs.
if self.preexist:
device = devicetree.getDeviceByName(devicename)
if not device:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("RAID volume \"%s\" specified with --useexisting does not exist.") % devicename))
removeExistingFormat(device, storage)
devicetree.registerAction(ActionCreateFormat(device, kwargs["fmt"]))
if ty == "swap":
storage.addFstabSwap(device)
else:
if devicename and devicename in (a.name for a in storage.mdarrays):
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("The RAID volume name \"%s\" is already in use.") % devicename))
# If a previous device has claimed this mount point, delete the
# old one.
try:
if self.mountpoint:
device = storage.mountpoints[self.mountpoint]
storage.destroyDevice(device)
except KeyError:
pass
try:
request = storage.newMDArray(**kwargs)
except (StorageError, ValueError) as e:
raise KickstartValueError(formatErrorMsg(self.lineno, msg=str(e)))
storage.createDevice(request)
if ty == "swap":
storage.addFstabSwap(request)
if self.encrypted:
if self.passphrase and not storage.encryptionPassphrase:
storage.encryptionPassphrase = self.passphrase
cert = getEscrowCertificate(storage.escrowCertificates, self.escrowcert)
if self.preexist:
luksformat = kwargs["fmt"]
device.format = getFormat("luks", passphrase=self.passphrase, device=device.path,
cipher=self.cipher,
escrow_cert=cert,
add_backup_passphrase=self.backuppassphrase)
luksdev = LUKSDevice("luks%d" % storage.nextID,
fmt=luksformat,
parents=device)
else:
luksformat = request.format
request.format = getFormat("luks", passphrase=self.passphrase,
cipher=self.cipher,
escrow_cert=cert,
add_backup_passphrase=self.backuppassphrase)
luksdev = LUKSDevice("luks%d" % storage.nextID,
fmt=luksformat,
parents=request)
storage.createDevice(luksdev)
class RepoData(commands.repo.F21_RepoData):
def __init__(self, *args, **kwargs):
""" Add enabled kwarg
:param enabled: The repo has been enabled
:type enabled: bool
"""
self.enabled = kwargs.pop("enabled", True)
self.repo_id = kwargs.pop("repo_id", None)
commands.repo.F21_RepoData.__init__(self, *args, **kwargs)
class ReqPart(commands.reqpart.F23_ReqPart):
def execute(self, storage, ksdata, instClass):
from blivet.autopart import doReqPartition
if not self.reqpart:
return
reqs = platform.setPlatformBootloaderReqs()
if self.addBoot:
bootPartitions = platform.setPlatformBootPartition()
# blivet doesn't know this - anaconda sets up the default boot fstype
# in various places in this file, as well as in setDefaultPartitioning
# in the install classes. We need to duplicate that here.
for part in bootPartitions:
if part.mountpoint == "/boot":
part.fstype = storage.defaultBootFSType
reqs += bootPartitions
doReqPartition(storage, reqs)
class RootPw(commands.rootpw.F18_RootPw):
def __init__(self, writePriority=100, *args, **kwargs):
if 'lock' not in kwargs:
kwargs['lock'] = True
super(RootPw, self).__init__(writePriority=writePriority, *args, **kwargs)
def execute(self, storage, ksdata, instClass, users):
if not self.password and not flags.automatedInstall:
self.lock = True
algo = getPassAlgo(ksdata.authconfig.authconfig)
users.setRootPassword(self.password, self.isCrypted, self.lock, algo, iutil.getSysroot())
class SELinux(commands.selinux.FC3_SELinux):
def execute(self, *args):
selinux_states = {SELINUX_DISABLED: "disabled",
SELINUX_ENFORCING: "enforcing",
SELINUX_PERMISSIVE: "permissive"}
if self.selinux is None:
# Use the defaults set by the installed (or not) selinux package
return
elif self.selinux not in selinux_states:
log.error("unknown selinux state: %s", self.selinux)
return
try:
selinux_cfg = SimpleConfigFile(iutil.getSysroot()+"/etc/selinux/config")
selinux_cfg.read()
selinux_cfg.set(("SELINUX", selinux_states[self.selinux]))
selinux_cfg.write()
except IOError as msg:
log.error("Error setting selinux mode: %s", msg)
class Services(commands.services.FC6_Services):
def execute(self, storage, ksdata, instClass):
for svc in self.disabled:
if not svc.endswith(".service"):
svc += ".service"
iutil.execInSysroot("systemctl", ["disable", svc])
for svc in self.enabled:
if not svc.endswith(".service"):
svc += ".service"
iutil.execInSysroot("systemctl", ["enable", svc])
class SshKey(commands.sshkey.F22_SshKey):
def execute(self, storage, ksdata, instClass, users):
for usr in self.sshUserList:
users.setUserSshKey(usr.username, usr.key)
class Timezone(commands.timezone.F23_Timezone):
def __init__(self, *args):
commands.timezone.F23_Timezone.__init__(self, *args)
self._added_chrony = False
self._enabled_chrony = False
self._disabled_chrony = False
def setup(self, ksdata):
if self.nontp:
if iutil.service_running(NTP_SERVICE) and \
can_touch_runtime_system("stop NTP service"):
ret = iutil.stop_service(NTP_SERVICE)
if ret != 0:
log.error("Failed to stop NTP service")
if self._added_chrony and NTP_PACKAGE in ksdata.packages.packageList:
ksdata.packages.packageList.remove(NTP_PACKAGE)
self._added_chrony = False
# Both un-enable and disable chrony, because sometimes it's installed
# off by default (packages) and sometimes not (liveimg).
if self._enabled_chrony and NTP_SERVICE in ksdata.services.enabled:
ksdata.services.enabled.remove(NTP_SERVICE)
self._enabled_chrony = False
if NTP_SERVICE not in ksdata.services.disabled:
ksdata.services.disabled.append(NTP_SERVICE)
self._disabled_chrony = True
else:
if not iutil.service_running(NTP_SERVICE) and \
can_touch_runtime_system("start NTP service"):
ret = iutil.start_service(NTP_SERVICE)
if ret != 0:
log.error("Failed to start NTP service")
if not NTP_PACKAGE in ksdata.packages.packageList:
ksdata.packages.packageList.append(NTP_PACKAGE)
self._added_chrony = True
if self._disabled_chrony and NTP_SERVICE in ksdata.services.disabled:
ksdata.services.disabled.remove(NTP_SERVICE)
self._disabled_chrony = False
if not NTP_SERVICE in ksdata.services.enabled and \
not NTP_SERVICE in ksdata.services.disabled:
ksdata.services.enabled.append(NTP_SERVICE)
self._enabled_chrony = True
def execute(self, *args):
# write out timezone configuration
if not timezone.is_valid_timezone(self.timezone):
# this should never happen, but for pity's sake
log.warning("Timezone %s set in kickstart is not valid, falling "\
"back to default (America/New_York).", self.timezone)
self.timezone = "America/New_York"
timezone.write_timezone_config(self, iutil.getSysroot())
# write out NTP configuration (if set)
chronyd_conf_path = os.path.normpath(iutil.getSysroot() + ntp.NTP_CONFIG_FILE)
if self.ntpservers and os.path.exists(chronyd_conf_path):
pools, servers = ntp.internal_to_pools_and_servers(self.ntpservers)
try:
ntp.save_servers_to_config(pools, servers, conf_file_path=chronyd_conf_path)
except ntp.NTPconfigError as ntperr:
log.warning("Failed to save NTP configuration: %s", ntperr)
class User(commands.user.F19_User):
def execute(self, storage, ksdata, instClass, users):
algo = getPassAlgo(ksdata.authconfig.authconfig)
for usr in self.userList:
kwargs = usr.__dict__
kwargs.update({"algo": algo, "root": iutil.getSysroot()})
# If the user password came from a kickstart and it is blank we
# need to make sure the account is locked, not created with an
# empty password.
if ksdata.user.seen and kwargs.get("password", "") == "":
kwargs["password"] = None
if not users.createUser(usr.name, **kwargs):
log.error("User %s already exists, not creating.", usr.name)
class VolGroup(commands.volgroup.F21_VolGroup):
def execute(self, storage, ksdata, instClass):
for v in self.vgList:
v.execute(storage, ksdata, instClass)
class VolGroupData(commands.volgroup.F21_VolGroupData):
def execute(self, storage, ksdata, instClass):
pvs = []
devicetree = storage.devicetree
storage.doAutoPart = False
# Get a list of all the physical volume devices that make up this VG.
for pv in self.physvols:
dev = devicetree.resolveDevice(pv)
if not dev:
# if pv is using --onpart, use original device
pv_name = ksdata.onPart.get(pv, pv)
dev = devicetree.resolveDevice(pv_name) or lookupAlias(devicetree, pv)
if dev and dev.format.type == "luks":
try:
dev = devicetree.getChildren(dev)[0]
except IndexError:
dev = None
if dev and dev.format.type != "lvmpv":
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("Physical volume \"%(device)s\" has a format of \"%(format)s\", but should have a format of \"lvmpv\".") %
{"device": pv, "format": dev.format.type}))
if not dev:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("Tried to use undefined partition \"%s\" in Volume Group specification") % pv))
pvs.append(dev)
if len(pvs) == 0 and not self.preexist:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("Volume group \"%s\" defined without any physical volumes. Either specify physical volumes or use --useexisting.") % self.vgname))
if self.pesize == 0:
# default PE size requested -- we use blivet's default in KiB
self.pesize = LVM_PE_SIZE.convertTo(KiB)
pesize = Size("%d KiB" % self.pesize)
possible_extents = LVMVolumeGroupDevice.get_supported_pe_sizes()
if pesize not in possible_extents:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("Volume group given physical extent size of \"%(extentSize)s\", but must be one of:\n%(validExtentSizes)s.") %
{"extentSize": pesize, "validExtentSizes": ", ".join(str(e) for e in possible_extents)}))
# If --noformat or --useexisting was given, there's really nothing to do.
if not self.format or self.preexist:
if not self.vgname:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("volgroup --noformat and volgroup --useexisting must also use the --name= option.")))
dev = devicetree.getDeviceByName(self.vgname)
if not dev:
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("Volume group \"%s\" given in volgroup command does not exist.") % self.vgname))
elif self.vgname in (vg.name for vg in storage.vgs):
raise KickstartValueError(formatErrorMsg(self.lineno,
msg=_("The volume group name \"%s\" is already in use.") % self.vgname))
else:
try:
request = storage.newVG(parents=pvs,
name=self.vgname,
peSize=pesize)
except (StorageError, ValueError) as e:
raise KickstartValueError(formatErrorMsg(self.lineno, msg=str(e)))
storage.createDevice(request)
if self.reserved_space:
request.reserved_space = self.reserved_space
elif self.reserved_percent:
request.reserved_percent = self.reserved_percent
# in case we had to truncate or otherwise adjust the specified name
ksdata.onPart[self.vgname] = request.name
class XConfig(commands.xconfig.F14_XConfig):
def execute(self, *args):
desktop = Desktop()
if self.startX:
desktop.runlevel = 5
if self.defaultdesktop:
desktop.desktop = self.defaultdesktop
# now write it out
desktop.write()
class SkipX(commands.skipx.FC3_SkipX):
def execute(self, *args):
if self.skipx:
desktop = Desktop()
desktop.runlevel = 3
desktop.write()
class ZFCP(commands.zfcp.F14_ZFCP):
def parse(self, args):
fcp = commands.zfcp.F14_ZFCP.parse(self, args)
try:
blivet.zfcp.ZFCP().addFCP(fcp.devnum, fcp.wwpn, fcp.fcplun)
except ValueError as e:
log.warning(str(e))
return fcp
class Keyboard(commands.keyboard.F18_Keyboard):
def execute(self, *args):
keyboard.write_keyboard_config(self, iutil.getSysroot())
class Upgrade(commands.upgrade.F20_Upgrade):
# Upgrade is no longer supported. If an upgrade command was included in
# a kickstart, warn the user and exit.
def parse(self, *args):
log.error("The upgrade kickstart command is no longer supported. Upgrade functionality is provided through fedup.")
sys.stderr.write(_("The upgrade kickstart command is no longer supported. Upgrade functionality is provided through fedup."))
iutil.ipmi_report(IPMI_ABORTED)
sys.exit(1)
###
### %anaconda Section
###
class AnacondaSectionHandler(BaseHandler):
"""A handler for only the anaconda ection's commands."""
commandMap = {
"pwpolicy": F22_PwPolicy
}
dataMap = {
"PwPolicyData": F22_PwPolicyData
}
def __init__(self):
BaseHandler.__init__(self, mapping=self.commandMap, dataMapping=self.dataMap)
def __str__(self):
"""Return the %anaconda section"""
retval = ""
# This dictionary should only be modified during __init__, so if it
# changes during iteration something has gone horribly wrong.
lst = sorted(self._writeOrder.keys())
for prio in lst:
for obj in self._writeOrder[prio]:
retval += str(obj)
if retval:
retval = "\n%anaconda\n" + retval + "%end\n"
return retval
class AnacondaSection(Section):
"""A section for anaconda specific commands."""
sectionOpen = "%anaconda"
def __init__(self, *args, **kwargs):
Section.__init__(self, *args, **kwargs)
self.cmdno = 0
def handleLine(self, line):
if not self.handler:
return
self.cmdno += 1
args = shlex.split(line, comments=True)
self.handler.currentCmd = args[0]
self.handler.currentLine = self.cmdno
return self.handler.dispatcher(args, self.cmdno)
def handleHeader(self, lineno, args):
"""Process the arguments to the %anaconda header."""
Section.handleHeader(self, lineno, args)
def finalize(self):
"""Let %anaconda know no additional data will come."""
Section.finalize(self)
###
### HANDLERS
###
# This is just the latest entry from pykickstart.handlers.control with all the
# classes we're overriding in place of the defaults.
commandMap = {
"auth": Authconfig,
"authconfig": Authconfig,
"autopart": AutoPart,
"btrfs": BTRFS,
"bootloader": Bootloader,
"clearpart": ClearPart,
"dmraid": DmRaid,
"eula": Eula,
"fcoe": Fcoe,
"firewall": Firewall,
"firstboot": Firstboot,
"group": Group,
"ignoredisk": IgnoreDisk,
"iscsi": Iscsi,
"iscsiname": IscsiName,
"keyboard": Keyboard,
"lang": Lang,
"logging": Logging,
"logvol": LogVol,
"multipath": MultiPath,
"network": Network,
"part": Partition,
"partition": Partition,
"raid": Raid,
"realm": Realm,
"reqpart": ReqPart,
"rootpw": RootPw,
"selinux": SELinux,
"services": Services,
"sshkey": SshKey,
"skipx": SkipX,
"timezone": Timezone,
"upgrade": Upgrade,
"user": User,
"volgroup": VolGroup,
"xconfig": XConfig,
"zfcp": ZFCP,
}
dataMap = {
"BTRFSData": BTRFSData,
"LogVolData": LogVolData,
"PartData": PartitionData,
"RaidData": RaidData,
"RepoData": RepoData,
"VolGroupData": VolGroupData,
}
superclass = returnClassForVersion()
class AnacondaKSHandler(superclass):
AddonClassType = AddonData
def __init__(self, addon_paths=None, commandUpdates=None, dataUpdates=None):
if addon_paths is None:
addon_paths = []
if commandUpdates is None:
commandUpdates = commandMap
if dataUpdates is None:
dataUpdates = dataMap
superclass.__init__(self, commandUpdates=commandUpdates, dataUpdates=dataUpdates)
self.onPart = {}
# collect all kickstart addons for anaconda to addons dictionary
# which maps addon_id to it's own data structure based on BaseData
# with execute method
addons = {}
# collect all AddonData subclasses from
# for p in addon_paths: //ks/*.(py|so)
# and register them under name
for module_name, path in addon_paths:
addon_id = os.path.basename(os.path.dirname(os.path.abspath(path)))
if not os.path.isdir(path):
continue
classes = collect(module_name, path, lambda cls: issubclass(cls, self.AddonClassType))
if classes:
addons[addon_id] = classes[0](name=addon_id)
# Prepare the final structures for 3rd party addons
self.addons = AddonRegistry(addons)
# The %anaconda section uses its own handler for a limited set of commands
self.anaconda = AnacondaSectionHandler()
def __str__(self):
return superclass.__str__(self) + "\n" + str(self.addons) + str(self.anaconda)
class AnacondaPreParser(KickstartParser):
# A subclass of KickstartParser that only looks for %pre scripts and
# sets them up to be run. All other scripts and commands are ignored.
def __init__(self, handler, followIncludes=True, errorsAreFatal=True,
missingIncludeIsFatal=True):
KickstartParser.__init__(self, handler, missingIncludeIsFatal=False)
def handleCommand(self, lineno, args):
pass
def setupSections(self):
self.registerSection(PreScriptSection(self.handler, dataObj=AnacondaKSScript))
self.registerSection(NullSection(self.handler, sectionOpen="%pre-install"))
self.registerSection(NullSection(self.handler, sectionOpen="%post"))
self.registerSection(NullSection(self.handler, sectionOpen="%traceback"))
self.registerSection(NullSection(self.handler, sectionOpen="%packages"))
self.registerSection(NullSection(self.handler, sectionOpen="%addon"))
self.registerSection(NullSection(self.handler.anaconda, sectionOpen="%anaconda"))
class AnacondaKSParser(KickstartParser):
def __init__(self, handler, followIncludes=True, errorsAreFatal=True,
missingIncludeIsFatal=True, scriptClass=AnacondaKSScript):
self.scriptClass = scriptClass
KickstartParser.__init__(self, handler)
def handleCommand(self, lineno, args):
if not self.handler:
return
return KickstartParser.handleCommand(self, lineno, args)
def setupSections(self):
self.registerSection(PreScriptSection(self.handler, dataObj=self.scriptClass))
self.registerSection(PreInstallScriptSection(self.handler, dataObj=self.scriptClass))
self.registerSection(PostScriptSection(self.handler, dataObj=self.scriptClass))
self.registerSection(TracebackScriptSection(self.handler, dataObj=self.scriptClass))
self.registerSection(PackageSection(self.handler))
self.registerSection(AddonSection(self.handler))
self.registerSection(AnacondaSection(self.handler.anaconda))
def preScriptPass(f):
# The first pass through kickstart file processing - look for %pre scripts
# and run them. This must come in a separate pass in case a script
# generates an included file that has commands for later.
ksparser = AnacondaPreParser(AnacondaKSHandler())
try:
ksparser.readKickstart(f)
except KickstartError as e:
# We do not have an interface here yet, so we cannot use our error
# handling callback.
print(e)
iutil.ipmi_report(IPMI_ABORTED)
sys.exit(1)
# run %pre scripts
runPreScripts(ksparser.handler.scripts)
def parseKickstart(f):
# preprocessing the kickstart file has already been handled in initramfs.
addon_paths = collect_addon_paths(ADDON_PATHS)
handler = AnacondaKSHandler(addon_paths["ks"])
ksparser = AnacondaKSParser(handler)
# We need this so all the /dev/disk/* stuff is set up before parsing.
udev.trigger(subsystem="block", action="change")
# So that drives onlined by these can be used in the ks file
blivet.iscsi.iscsi().startup()
blivet.fcoe.fcoe().startup()
blivet.zfcp.ZFCP().startup()
# Note we do NOT call dasd.startup() here, that does not online drives, but
# only checks if they need formatting, which requires zerombr to be known
try:
ksparser.readKickstart(f)
except KickstartError as e:
# We do not have an interface here yet, so we cannot use our error
# handling callback.
print(e)
iutil.ipmi_report(IPMI_ABORTED)
sys.exit(1)
return handler
def appendPostScripts(ksdata):
scripts = ""
# Read in all the post script snippets to a single big string.
for fn in glob.glob("/usr/share/anaconda/post-scripts/*ks"):
f = open(fn, "r")
scripts += f.read()
f.close()
# Then parse the snippets against the existing ksdata. We can do this
# because pykickstart allows multiple parses to save their data into a
# single data object. Errors parsing the scripts are a bug in anaconda,
# so just raise an exception.
ksparser = AnacondaKSParser(ksdata, scriptClass=AnacondaInternalScript)
ksparser.readKickstartFromString(scripts, reset=False)
def runPostScripts(scripts):
postScripts = [s for s in scripts if s.type == KS_SCRIPT_POST]
if len(postScripts) == 0:
return
log.info("Running kickstart %%post script(s)")
for script in postScripts:
script.run(iutil.getSysroot())
log.info("All kickstart %%post script(s) have been run")
def runPreScripts(scripts):
preScripts = [s for s in scripts if s.type == KS_SCRIPT_PRE]
if len(preScripts) == 0:
return
log.info("Running kickstart %%pre script(s)")
stdoutLog.info(_("Running pre-installation scripts"))
for script in preScripts:
script.run("/")
log.info("All kickstart %%pre script(s) have been run")
def runPreInstallScripts(scripts):
preInstallScripts = [s for s in scripts if s.type == KS_SCRIPT_PREINSTALL]
if len(preInstallScripts) == 0:
return
log.info("Running kickstart %%pre-install script(s)")
for script in preInstallScripts:
script.run("/")
log.info("All kickstart %%pre-install script(s) have been run")
def runTracebackScripts(scripts):
log.info("Running kickstart %%traceback script(s)")
for script in filter(lambda s: s.type == KS_SCRIPT_TRACEBACK, scripts):
script.run("/")
log.info("All kickstart %%traceback script(s) have been run")
def resetCustomStorageData(ksdata):
for command in ["partition", "raid", "volgroup", "logvol", "btrfs"]:
ksdata.resetCommand(command)
ksdata.clearpart.type = CLEARPART_TYPE_NONE
def doKickstartStorage(storage, ksdata, instClass):
""" Setup storage state from the kickstart data """
ksdata.clearpart.execute(storage, ksdata, instClass)
if not any(d for d in storage.disks
if not d.format.hidden and not d.protected):
return
# snapshot free space now so that we know how much we had available
storage.createFreeSpaceSnapshot()
ksdata.bootloader.execute(storage, ksdata, instClass)
ksdata.autopart.execute(storage, ksdata, instClass)
ksdata.reqpart.execute(storage, ksdata, instClass)
ksdata.partition.execute(storage, ksdata, instClass)
ksdata.raid.execute(storage, ksdata, instClass)
ksdata.volgroup.execute(storage, ksdata, instClass)
ksdata.logvol.execute(storage, ksdata, instClass)
ksdata.btrfs.execute(storage, ksdata, instClass)
# also calls ksdata.bootloader.execute
storage.setUpBootLoader()