f466747949
This may have performance impact on some older SSD, but on the other hand, without this option it's pretty easy to fill the whole LVM thin pool even if there is plenty free space in dom0. Note that this doesn't enable it on LUKS layer, this is still disabled by default. Fixes QubesOS/qubes-issues#3226
2276 lines
92 KiB
Python
2276 lines
92 KiB
Python
#
|
|
# kickstart.py: kickstart install support
|
|
#
|
|
# Copyright (C) 1999-2016
|
|
# Red Hat, Inc. All rights reserved.
|
|
#
|
|
# This program is free software; you can redistribute it and/or modify
|
|
# it under the terms of the GNU General Public License as published by
|
|
# the Free Software Foundation; either version 2 of the License, or
|
|
# (at your option) any later version.
|
|
#
|
|
# This program is distributed in the hope that it will be useful,
|
|
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
# GNU General Public License for more details.
|
|
#
|
|
# You should have received a copy of the GNU General Public License
|
|
# along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
#
|
|
|
|
from pyanaconda.errors import ScriptError, errorHandler
|
|
from blivet.deviceaction import ActionCreateFormat, ActionResizeDevice, ActionResizeFormat
|
|
from blivet.devices import LUKSDevice
|
|
from blivet.devices.lvm import LVMVolumeGroupDevice, LVMCacheRequest
|
|
from blivet.devicelibs.lvm import LVM_PE_SIZE, KNOWN_THPOOL_PROFILES
|
|
from blivet.devicelibs.crypto import MIN_CREATE_ENTROPY
|
|
from blivet.formats import get_format
|
|
from blivet.partitioning import do_partitioning
|
|
from blivet.partitioning import grow_lvm
|
|
from blivet.errors import PartitioningError, StorageError, BTRFSValueError
|
|
from blivet.size import Size, KiB
|
|
from blivet import udev
|
|
from blivet import autopart
|
|
from blivet.platform import platform
|
|
import blivet.iscsi
|
|
import blivet.fcoe
|
|
import blivet.zfcp
|
|
import blivet.arch
|
|
|
|
import glob
|
|
from pyanaconda import iutil
|
|
import os
|
|
import os.path
|
|
import tempfile
|
|
from pyanaconda.flags import flags, can_touch_runtime_system
|
|
from pyanaconda.constants import ADDON_PATHS, IPMI_ABORTED, TEXT_ONLY_TARGET, GRAPHICAL_TARGET
|
|
import shlex
|
|
import requests
|
|
import sys
|
|
import pykickstart.commands as commands
|
|
from pyanaconda import keyboard
|
|
from pyanaconda import ntp
|
|
from pyanaconda import timezone
|
|
from pyanaconda.timezone import NTP_PACKAGE, NTP_SERVICE
|
|
from pyanaconda import localization
|
|
from pyanaconda import network
|
|
from pyanaconda import nm
|
|
from pyanaconda.simpleconfig import SimpleConfigFile
|
|
from pyanaconda.users import getPassAlgo
|
|
from pyanaconda.desktop import Desktop
|
|
from pyanaconda.i18n import _
|
|
from pyanaconda.ui.common import collect
|
|
from pyanaconda.addons import AddonSection, AddonData, AddonRegistry, collect_addon_paths
|
|
from pyanaconda.bootloader import GRUB2, get_bootloader
|
|
from pyanaconda.pwpolicy import F22_PwPolicy, F22_PwPolicyData
|
|
from pyanaconda.storage_utils import device_matches
|
|
|
|
from pykickstart.constants import CLEARPART_TYPE_NONE, FIRSTBOOT_SKIP, FIRSTBOOT_RECONFIG, KS_SCRIPT_POST, KS_SCRIPT_PRE, \
|
|
KS_SCRIPT_TRACEBACK, KS_SCRIPT_PREINSTALL, SELINUX_DISABLED, SELINUX_ENFORCING, SELINUX_PERMISSIVE
|
|
from pykickstart.base import BaseHandler
|
|
from pykickstart.errors import formatErrorMsg, KickstartError, KickstartParseError
|
|
from pykickstart.parser import KickstartParser
|
|
from pykickstart.parser import Script as KSScript
|
|
from pykickstart.sections import Section
|
|
from pykickstart.sections import NullSection, PackageSection, PostScriptSection, PreScriptSection, PreInstallScriptSection, \
|
|
OnErrorScriptSection, TracebackScriptSection
|
|
from pykickstart.version import returnClassForVersion
|
|
|
|
import logging
|
|
log = logging.getLogger("anaconda")
|
|
stderrLog = logging.getLogger("anaconda.stderr")
|
|
storage_log = logging.getLogger("blivet")
|
|
stdoutLog = logging.getLogger("anaconda.stdout")
|
|
from pyanaconda.anaconda_log import logger, logLevelMap, setHandlersLevel, DEFAULT_LEVEL
|
|
|
|
class AnacondaKSScript(KSScript):
|
|
""" Execute a kickstart script
|
|
|
|
This will write the script to a file named /tmp/ks-script- before
|
|
execution.
|
|
Output is logged by the program logger, the path specified by --log
|
|
or to /tmp/ks-script-\\*.log
|
|
"""
|
|
def run(self, chroot):
|
|
""" Run the kickstart script
|
|
@param chroot directory path to chroot into before execution
|
|
"""
|
|
if self.inChroot:
|
|
scriptRoot = chroot
|
|
else:
|
|
scriptRoot = "/"
|
|
|
|
(fd, path) = tempfile.mkstemp("", "ks-script-", scriptRoot + "/tmp")
|
|
|
|
os.write(fd, self.script.encode("utf-8"))
|
|
os.close(fd)
|
|
os.chmod(path, 0o700)
|
|
|
|
# Always log stdout/stderr from scripts. Using --log just lets you
|
|
# pick where it goes. The script will also be logged to program.log
|
|
# because of execWithRedirect.
|
|
if self.logfile:
|
|
if self.inChroot:
|
|
messages = "%s/%s" % (scriptRoot, self.logfile)
|
|
else:
|
|
messages = self.logfile
|
|
|
|
d = os.path.dirname(messages)
|
|
if not os.path.exists(d):
|
|
os.makedirs(d)
|
|
else:
|
|
# Always log outside the chroot, we copy those logs into the
|
|
# chroot later.
|
|
messages = "/tmp/%s.log" % os.path.basename(path)
|
|
|
|
with open(messages, "w") as fp:
|
|
rc = iutil.execWithRedirect(self.interp, ["/tmp/%s" % os.path.basename(path)],
|
|
stdout=fp,
|
|
root=scriptRoot)
|
|
|
|
if rc != 0:
|
|
log.error("Error code %s running the kickstart script at line %s", rc, self.lineno)
|
|
if self.errorOnFail:
|
|
err = ""
|
|
with open(messages, "r") as fp:
|
|
err = "".join(fp.readlines())
|
|
|
|
errorHandler.cb(ScriptError(self.lineno, err))
|
|
iutil.ipmi_report(IPMI_ABORTED)
|
|
sys.exit(0)
|
|
|
|
class AnacondaInternalScript(AnacondaKSScript):
|
|
def __init__(self, *args, **kwargs):
|
|
AnacondaKSScript.__init__(self, *args, **kwargs)
|
|
self._hidden = True
|
|
|
|
def __str__(self):
|
|
# Scripts that implement portions of anaconda (copying screenshots and
|
|
# log files, setfilecons, etc.) should not be written to the output
|
|
# kickstart file.
|
|
return ""
|
|
|
|
def getEscrowCertificate(escrowCerts, url):
|
|
if not url:
|
|
return None
|
|
|
|
if url in escrowCerts:
|
|
return escrowCerts[url]
|
|
|
|
needs_net = not url.startswith("/") and not url.startswith("file:")
|
|
if needs_net and not nm.nm_is_connected():
|
|
msg = _("Escrow certificate %s requires the network.") % url
|
|
raise KickstartError(msg)
|
|
|
|
log.info("escrow: downloading %s", url)
|
|
|
|
try:
|
|
request = iutil.requests_session().get(url, verify=True)
|
|
except requests.exceptions.SSLError as e:
|
|
msg = _("SSL error while downloading the escrow certificate:\n\n%s") % e
|
|
raise KickstartError(msg)
|
|
except requests.exceptions.RequestException as e:
|
|
msg = _("The following error was encountered while downloading the escrow certificate:\n\n%s") % e
|
|
raise KickstartError(msg)
|
|
|
|
try:
|
|
escrowCerts[url] = request.content
|
|
finally:
|
|
request.close()
|
|
|
|
return escrowCerts[url]
|
|
|
|
def lookupAlias(devicetree, alias):
|
|
for dev in devicetree.devices:
|
|
if getattr(dev, "req_name", None) == alias:
|
|
return dev
|
|
|
|
return None
|
|
|
|
def getAvailableDiskSpace(storage):
|
|
"""
|
|
Get overall disk space available on disks we may use.
|
|
|
|
:param storage: blivet.Blivet instance
|
|
:return: overall disk space available
|
|
:rtype: :class:`blivet.size.Size`
|
|
|
|
"""
|
|
|
|
free_space = storage.free_space_snapshot
|
|
# blivet creates a new free space dict to instead of modifying the old one,
|
|
# so there is no worry about the dictionary changing during iteration.
|
|
return sum(disk_free for disk_free, fs_free in free_space.values())
|
|
|
|
def refreshAutoSwapSize(storage):
|
|
"""
|
|
Refresh size of the auto partitioning request for swap device according to
|
|
the current state of the storage configuration.
|
|
|
|
:param storage: blivet.Blivet instance
|
|
|
|
"""
|
|
|
|
for request in storage.autopart_requests:
|
|
if request.fstype == "swap":
|
|
disk_space = getAvailableDiskSpace(storage)
|
|
request.size = autopart.swap_suggestion(disk_space=disk_space)
|
|
break
|
|
|
|
###
|
|
### SUBCLASSES OF PYKICKSTART COMMAND HANDLERS
|
|
###
|
|
|
|
class Authconfig(commands.authconfig.FC3_Authconfig):
|
|
def __init__(self, *args, **kwargs):
|
|
commands.authconfig.FC3_Authconfig.__init__(self, *args, **kwargs)
|
|
self.packages = []
|
|
|
|
def setup(self):
|
|
if self.seen:
|
|
self.packages = ["authconfig"]
|
|
|
|
def execute(self, *args):
|
|
cmd = "/usr/sbin/authconfig"
|
|
if not os.path.lexists(iutil.getSysroot()+cmd):
|
|
if flags.automatedInstall and self.seen:
|
|
msg = _("%s is missing. Cannot setup authentication.") % cmd
|
|
raise KickstartError(msg)
|
|
else:
|
|
return
|
|
|
|
args = ["--update", "--nostart"] + shlex.split(self.authconfig)
|
|
|
|
if not flags.automatedInstall and \
|
|
(os.path.exists(iutil.getSysroot() + "/lib64/security/pam_fprintd.so") or \
|
|
os.path.exists(iutil.getSysroot() + "/lib/security/pam_fprintd.so")):
|
|
args += ["--enablefingerprint"]
|
|
|
|
try:
|
|
iutil.execInSysroot(cmd, args)
|
|
except RuntimeError as msg:
|
|
log.error("Error running %s %s: %s", cmd, args, msg)
|
|
|
|
class AutoPart(commands.autopart.F21_AutoPart):
|
|
def __init__(self, writePriority=100, *args, **kwargs):
|
|
if 'encrypted' not in kwargs:
|
|
kwargs['encrypted'] = True
|
|
super(AutoPart, self).__init__(writePriority=writePriority, *args, **kwargs)
|
|
|
|
def parse(self, args):
|
|
retval = commands.autopart.F21_AutoPart.parse(self, args)
|
|
|
|
if self.fstype:
|
|
fmt = blivet.formats.get_format(self.fstype)
|
|
if not fmt or fmt.type is None:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("autopart fstype of %s is invalid.") % self.fstype))
|
|
|
|
return retval
|
|
|
|
def execute(self, storage, ksdata, instClass):
|
|
from blivet.autopart import do_autopart
|
|
from pyanaconda.storage_utils import sanity_check
|
|
|
|
if not self.autopart:
|
|
return
|
|
|
|
if self.fstype:
|
|
try:
|
|
storage.set_default_fstype(self.fstype)
|
|
storage.set_default_boot_fstype(self.fstype)
|
|
except ValueError:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("Settings default fstype to %s failed.") % self.fstype))
|
|
|
|
# sets up default autopartitioning. use clearpart separately
|
|
# if you want it
|
|
instClass.setDefaultPartitioning(storage)
|
|
storage.do_autopart = True
|
|
|
|
if self.encrypted:
|
|
storage.encrypted_autopart = True
|
|
storage.encryption_passphrase = self.passphrase
|
|
storage.encryption_cipher = self.cipher
|
|
storage.autopart_escrow_cert = getEscrowCertificate(storage.escrow_certificates, self.escrowcert)
|
|
storage.autoppart_add_backup_passphrase = self.backuppassphrase
|
|
|
|
if self.type is not None:
|
|
storage.autopart_type = self.type
|
|
|
|
do_autopart(storage, ksdata, min_luks_entropy=MIN_CREATE_ENTROPY)
|
|
errors = sanity_check(storage)
|
|
if errors:
|
|
raise PartitioningError("autopart failed:\n" + "\n".join(str(error) for error in errors))
|
|
|
|
class Bootloader(commands.bootloader.F21_Bootloader):
|
|
def __init__(self, *args, **kwargs):
|
|
commands.bootloader.F21_Bootloader.__init__(self, *args, **kwargs)
|
|
self.location = "mbr"
|
|
self._useBackup = False
|
|
self._origBootDrive = None
|
|
|
|
def parse(self, args):
|
|
commands.bootloader.F21_Bootloader.parse(self, args)
|
|
if self.location == "partition" and isinstance(get_bootloader(), GRUB2):
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("GRUB2 does not support installation to a partition.")))
|
|
|
|
if self.isCrypted and isinstance(get_bootloader(), GRUB2):
|
|
if not self.password.startswith("grub.pbkdf2."):
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg="GRUB2 encrypted password must be in grub.pbkdf2 format."))
|
|
|
|
return self
|
|
|
|
def execute(self, storage, ksdata, instClass, dry_run=False):
|
|
""" Resolve and execute the bootloader installation.
|
|
|
|
:param storage: object storing storage-related information
|
|
(disks, partitioning, bootloader, etc.)
|
|
:type storage: blivet.Blivet
|
|
:param payload: object storing packaging-related information
|
|
:type payload: pyanaconda.packaging.Payload
|
|
:param instclass: distribution-specific information
|
|
:type instclass: pyanaconda.installclass.BaseInstallClass
|
|
:param dry_run: flag if this is only dry run before the partitioning
|
|
will be resolved
|
|
:type dry_run: bool
|
|
"""
|
|
if flags.imageInstall and blivet.arch.is_s390():
|
|
self.location = "none"
|
|
|
|
if dry_run:
|
|
self._origBootDrive = self.bootDrive
|
|
self._useBackup = True
|
|
elif self._useBackup:
|
|
self.bootDrive = self._origBootDrive
|
|
self._useBackup = False
|
|
|
|
if self.location == "none":
|
|
location = None
|
|
elif self.location == "partition":
|
|
location = "boot"
|
|
else:
|
|
location = self.location
|
|
|
|
if not location:
|
|
storage.bootloader.skip_bootloader = True
|
|
return
|
|
|
|
if self.appendLine:
|
|
args = self.appendLine.split()
|
|
storage.bootloader.boot_args.update(args)
|
|
|
|
if self.password:
|
|
if self.isCrypted:
|
|
storage.bootloader.encrypted_password = self.password
|
|
else:
|
|
storage.bootloader.password = self.password
|
|
|
|
if location:
|
|
storage.bootloader.set_preferred_stage1_type(location)
|
|
|
|
if self.timeout is not None:
|
|
storage.bootloader.timeout = self.timeout
|
|
|
|
# Throw out drives specified that don't exist or cannot be used (iSCSI
|
|
# device on an s390 machine)
|
|
disk_names = [d.name for d in storage.disks
|
|
if not d.format.hidden and not d.protected and
|
|
(not blivet.arch.is_s390() or not isinstance(d, blivet.devices.iScsiDiskDevice))]
|
|
diskSet = set(disk_names)
|
|
|
|
valid_disks = []
|
|
# Drive specifications can contain | delimited variant specifications,
|
|
# such as for example: "vd*|hd*|sd*"
|
|
# So use the resolved disk identifiers returned by the device_matches() function in place
|
|
# of the original specification but still remove the specifications that don't match anything
|
|
# from the output kickstart to keep existing --driveorder processing behavior.
|
|
for drive in self.driveorder[:]:
|
|
matches = device_matches(drive, devicetree=storage.devicetree, disks_only=True)
|
|
if set(matches).isdisjoint(diskSet):
|
|
log.warning("requested drive %s in boot drive order doesn't exist or cannot be used",
|
|
drive)
|
|
self.driveorder.remove(drive)
|
|
else:
|
|
valid_disks.extend(matches)
|
|
|
|
storage.bootloader.disk_order = valid_disks
|
|
|
|
# When bootloader doesn't have --boot-drive parameter then use this logic as fallback:
|
|
# 1) If present first valid disk from driveorder parameter
|
|
# 2) If present and usable, use disk where /boot partition is placed
|
|
# 3) Use first disk from Blivet
|
|
if self.bootDrive:
|
|
matches = set(device_matches(self.bootDrive, devicetree=storage.devicetree, disks_only=True))
|
|
if len(matches) > 1:
|
|
raise KickstartParseError(
|
|
formatErrorMsg(self.lineno,
|
|
msg=(_("More than one match found for given boot drive \"%s\".")
|
|
% self.bootDrive)))
|
|
elif matches.isdisjoint(diskSet):
|
|
raise KickstartParseError(
|
|
formatErrorMsg(self.lineno,
|
|
msg=(_("Requested boot drive \"%s\" doesn't exist or cannot be used.")
|
|
% self.bootDrive)))
|
|
# Take valid disk from --driveorder
|
|
elif len(valid_disks) >= 1:
|
|
log.debug("Bootloader: use '%s' first disk from driveorder as boot drive, dry run %s",
|
|
valid_disks[0], dry_run)
|
|
self.bootDrive = valid_disks[0]
|
|
else:
|
|
# Try to find /boot
|
|
#
|
|
# This method is executed two times. Before and after partitioning.
|
|
# In the first run, the result is used for other partitioning but
|
|
# the second will be used.
|
|
try:
|
|
boot_dev = storage.mountpoints["/boot"]
|
|
except KeyError:
|
|
log.debug("Bootloader: /boot partition is not present, dry run %s", dry_run)
|
|
else:
|
|
boot_drive = ""
|
|
# Use disk ancestor
|
|
if boot_dev.disks:
|
|
boot_drive = boot_dev.disks[0].name
|
|
|
|
if boot_drive and boot_drive in disk_names:
|
|
self.bootDrive = boot_drive
|
|
log.debug("Bootloader: use /boot partition's disk '%s' as boot drive, dry run %s",
|
|
boot_drive, dry_run)
|
|
|
|
# Nothing was found use first disk from Blivet
|
|
if not self.bootDrive:
|
|
log.debug("Bootloader: fallback use first disk return from Blivet '%s' as boot drive, dry run %s",
|
|
disk_names[0], dry_run)
|
|
self.bootDrive = disk_names[0]
|
|
|
|
drive = storage.devicetree.resolve_device(self.bootDrive)
|
|
storage.bootloader.stage1_disk = drive
|
|
|
|
if self.leavebootorder:
|
|
flags.leavebootorder = True
|
|
|
|
if self.nombr:
|
|
flags.nombr = True
|
|
|
|
class BTRFS(commands.btrfs.F23_BTRFS):
|
|
def execute(self, storage, ksdata, instClass):
|
|
for b in self.btrfsList:
|
|
b.execute(storage, ksdata, instClass)
|
|
|
|
class BTRFSData(commands.btrfs.F23_BTRFSData):
|
|
def execute(self, storage, ksdata, instClass):
|
|
devicetree = storage.devicetree
|
|
|
|
storage.do_autopart = False
|
|
|
|
members = []
|
|
|
|
# Get a list of all the devices that make up this volume.
|
|
for member in self.devices:
|
|
dev = devicetree.resolve_device(member)
|
|
if not dev:
|
|
# if using --onpart, use original device
|
|
member_name = ksdata.onPart.get(member, member)
|
|
dev = devicetree.resolve_device(member_name) or lookupAlias(devicetree, member)
|
|
|
|
if dev and dev.format.type == "luks":
|
|
try:
|
|
dev = dev.children[0]
|
|
except IndexError:
|
|
dev = None
|
|
|
|
if dev and dev.format.type != "btrfs":
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("Btrfs partition \"%(device)s\" has a format of \"%(format)s\", but should have a format of \"btrfs\".") %
|
|
{"device": member, "format": dev.format.type}))
|
|
|
|
if not dev:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("Tried to use undefined partition \"%s\" in Btrfs volume specification.") % member))
|
|
|
|
members.append(dev)
|
|
|
|
if self.subvol:
|
|
name = self.name
|
|
elif self.label:
|
|
name = self.label
|
|
else:
|
|
name = None
|
|
|
|
if len(members) == 0 and not self.preexist:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("Btrfs volume defined without any member devices. Either specify member devices or use --useexisting.")))
|
|
|
|
# allow creating btrfs vols/subvols without specifying mountpoint
|
|
if self.mountpoint in ("none", "None"):
|
|
self.mountpoint = ""
|
|
|
|
# Sanity check mountpoint
|
|
if self.mountpoint != "" and self.mountpoint[0] != '/':
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("The mount point \"%s\" is not valid. It must start with a /.") % self.mountpoint))
|
|
|
|
# If a previous device has claimed this mount point, delete the
|
|
# old one.
|
|
try:
|
|
if self.mountpoint:
|
|
device = storage.mountpoints[self.mountpoint]
|
|
storage.destroy_device(device)
|
|
except KeyError:
|
|
pass
|
|
|
|
if self.preexist:
|
|
device = devicetree.resolve_device(self.name)
|
|
if not device:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("Btrfs volume \"%s\" specified with --useexisting does not exist.") % self.name))
|
|
|
|
device.format.mountpoint = self.mountpoint
|
|
else:
|
|
try:
|
|
request = storage.new_btrfs(name=name,
|
|
subvol=self.subvol,
|
|
mountpoint=self.mountpoint,
|
|
metadata_level=self.metaDataLevel,
|
|
data_level=self.dataLevel,
|
|
parents=members,
|
|
create_options=self.mkfsopts)
|
|
except BTRFSValueError as e:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno, msg=str(e)))
|
|
|
|
storage.create_device(request)
|
|
|
|
class Realm(commands.realm.F19_Realm):
|
|
def __init__(self, *args):
|
|
commands.realm.F19_Realm.__init__(self, *args)
|
|
self.packages = []
|
|
self.discovered = ""
|
|
|
|
def setup(self):
|
|
if not self.join_realm:
|
|
return
|
|
|
|
try:
|
|
argv = ["discover", "--verbose"] + \
|
|
self.discover_options + [self.join_realm]
|
|
output = iutil.execWithCapture("realm", argv, filter_stderr=True)
|
|
except OSError:
|
|
# TODO: A lousy way of propagating what will usually be
|
|
# 'no such realm'
|
|
# The error message is logged by iutil
|
|
return
|
|
|
|
# Now parse the output for the required software. First line is the
|
|
# realm name, and following lines are information as "name: value"
|
|
self.packages = ["realmd"]
|
|
self.discovered = ""
|
|
|
|
lines = output.split("\n")
|
|
if not lines:
|
|
return
|
|
self.discovered = lines.pop(0).strip()
|
|
log.info("Realm discovered: %s", self.discovered)
|
|
for line in lines:
|
|
parts = line.split(":", 1)
|
|
if len(parts) == 2 and parts[0].strip() == "required-package":
|
|
self.packages.append(parts[1].strip())
|
|
|
|
log.info("Realm %s needs packages %s",
|
|
self.discovered, ", ".join(self.packages))
|
|
|
|
def execute(self, *args):
|
|
if not self.discovered:
|
|
return
|
|
for arg in self.join_args:
|
|
if arg.startswith("--no-password") or arg.startswith("--one-time-password"):
|
|
pw_args = []
|
|
break
|
|
else:
|
|
# no explicit password arg using implicit --no-password
|
|
pw_args = ["--no-password"]
|
|
|
|
argv = ["join", "--install", iutil.getSysroot(), "--verbose"] + \
|
|
pw_args + self.join_args
|
|
rc = -1
|
|
try:
|
|
rc = iutil.execWithRedirect("realm", argv)
|
|
except OSError:
|
|
pass
|
|
|
|
if rc == 0:
|
|
log.info("Joined realm %s", self.join_realm)
|
|
|
|
|
|
class ClearPart(commands.clearpart.F21_ClearPart):
|
|
def parse(self, args):
|
|
retval = commands.clearpart.F21_ClearPart.parse(self, args)
|
|
|
|
if self.type is None:
|
|
self.type = CLEARPART_TYPE_NONE
|
|
|
|
if self.disklabel and self.disklabel not in platform.disklabel_types:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("Disklabel \"%s\" given in clearpart command is not "
|
|
"supported on this platform.") % self.disklabel))
|
|
|
|
# Do any glob expansion now, since we need to have the real list of
|
|
# disks available before the execute methods run.
|
|
drives = []
|
|
for spec in self.drives:
|
|
matched = device_matches(spec, disks_only=True)
|
|
if matched:
|
|
drives.extend(matched)
|
|
else:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("Disk \"%s\" given in clearpart command does not exist.") % spec))
|
|
|
|
self.drives = drives
|
|
|
|
# Do any glob expansion now, since we need to have the real list of
|
|
# devices available before the execute methods run.
|
|
devices = []
|
|
for spec in self.devices:
|
|
matched = device_matches(spec, disks_only=True)
|
|
if matched:
|
|
devices.extend(matched)
|
|
else:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("Device \"%s\" given in clearpart device list does not exist.") % spec))
|
|
|
|
self.devices = devices
|
|
|
|
return retval
|
|
|
|
def execute(self, storage, ksdata, instClass):
|
|
storage.config.clearpart_type = self.type
|
|
storage.config.clearpart_disks = self.drives
|
|
storage.config.clearpart_devices = self.devices
|
|
|
|
if self.initAll:
|
|
storage.config.initialize_disks = self.initAll
|
|
|
|
if self.disklabel:
|
|
if not platform.set_default_disklabel_type(self.disklabel):
|
|
log.warning("%s is not a supported disklabel type on this platform. "
|
|
"Using default disklabel %s instead.", self.disklabel, platform.default_disklabel_type)
|
|
|
|
storage.clear_partitions()
|
|
|
|
class Fcoe(commands.fcoe.F13_Fcoe):
|
|
def parse(self, args):
|
|
fc = commands.fcoe.F13_Fcoe.parse(self, args)
|
|
|
|
if fc.nic not in nm.nm_devices():
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("NIC \"%s\" given in fcoe command does not exist.") % fc.nic))
|
|
|
|
if fc.nic in (info[0] for info in blivet.fcoe.fcoe.nics):
|
|
log.info("Kickstart fcoe device %s already added from EDD, ignoring", fc.nic)
|
|
else:
|
|
msg = blivet.fcoe.fcoe.add_san(nic=fc.nic, dcb=fc.dcb, auto_vlan=True)
|
|
if not msg:
|
|
msg = "Succeeded."
|
|
blivet.fcoe.fcoe.added_nics.append(fc.nic)
|
|
|
|
log.info("adding FCoE SAN on %s: %s", fc.nic, msg)
|
|
|
|
return fc
|
|
|
|
class Firewall(commands.firewall.F20_Firewall):
|
|
def __init__(self, *args, **kwargs):
|
|
commands.firewall.F20_Firewall.__init__(self, *args, **kwargs)
|
|
self.packages = []
|
|
|
|
def setup(self):
|
|
if self.seen:
|
|
self.packages = ["firewalld"]
|
|
|
|
def execute(self, storage, ksdata, instClass):
|
|
args = []
|
|
# enabled is None if neither --enable or --disable is passed
|
|
# default to enabled if nothing has been set.
|
|
if self.enabled == False:
|
|
args += ["--disabled"]
|
|
else:
|
|
args += ["--enabled"]
|
|
|
|
if "ssh" not in self.services and "ssh" not in self.remove_services \
|
|
and "22:tcp" not in self.ports:
|
|
args += ["--service=ssh"]
|
|
|
|
for dev in self.trusts:
|
|
args += ["--trust=%s" % (dev,)]
|
|
|
|
for port in self.ports:
|
|
args += ["--port=%s" % (port,)]
|
|
|
|
for remove_service in self.remove_services:
|
|
args += ["--remove-service=%s" % (remove_service,)]
|
|
|
|
for service in self.services:
|
|
args += ["--service=%s" % (service,)]
|
|
|
|
cmd = "/usr/bin/firewall-offline-cmd"
|
|
if not os.path.exists(iutil.getSysroot()+cmd):
|
|
if self.enabled:
|
|
msg = _("%s is missing. Cannot setup firewall.") % (cmd,)
|
|
raise KickstartError(msg)
|
|
else:
|
|
iutil.execInSysroot(cmd, args)
|
|
|
|
class Firstboot(commands.firstboot.FC3_Firstboot):
|
|
def setup(self, *args):
|
|
# firstboot should be disabled by default after kickstart installations
|
|
if flags.automatedInstall and not self.seen:
|
|
self.firstboot = FIRSTBOOT_SKIP
|
|
|
|
def execute(self, *args):
|
|
action = iutil.enable_service
|
|
unit_name = "initial-setup.service"
|
|
|
|
# find if the unit file for the Initial Setup service is installed
|
|
unit_exists = os.path.exists(os.path.join(iutil.getSysroot(), "lib/systemd/system/", unit_name))
|
|
if unit_exists and self.firstboot == FIRSTBOOT_RECONFIG:
|
|
# write the reconfig trigger file
|
|
f = open(os.path.join(iutil.getSysroot(), "etc/reconfigSys"), "w+")
|
|
f.close()
|
|
|
|
if self.firstboot == FIRSTBOOT_SKIP:
|
|
action = iutil.disable_service
|
|
|
|
# enable/disable the Initial Setup service (if its unit is installed)
|
|
if unit_exists:
|
|
action(unit_name)
|
|
|
|
class Group(commands.group.F12_Group):
|
|
def execute(self, storage, ksdata, instClass, users):
|
|
for grp in self.groupList:
|
|
kwargs = grp.__dict__
|
|
kwargs.update({"root": iutil.getSysroot()})
|
|
try:
|
|
users.createGroup(grp.name, **kwargs)
|
|
except ValueError as e:
|
|
log.warning(str(e))
|
|
|
|
class IgnoreDisk(commands.ignoredisk.RHEL6_IgnoreDisk):
|
|
def parse(self, args):
|
|
retval = commands.ignoredisk.RHEL6_IgnoreDisk.parse(self, args)
|
|
|
|
# See comment in ClearPart.parse
|
|
drives = []
|
|
for spec in self.ignoredisk:
|
|
matched = device_matches(spec, disks_only=True)
|
|
if matched:
|
|
drives.extend(matched)
|
|
else:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("Disk \"%s\" given in ignoredisk command does not exist.") % spec))
|
|
|
|
self.ignoredisk = drives
|
|
|
|
drives = []
|
|
for spec in self.onlyuse:
|
|
matched = device_matches(spec, disks_only=True)
|
|
if matched:
|
|
drives.extend(matched)
|
|
else:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("Disk \"%s\" given in ignoredisk command does not exist.") % spec))
|
|
|
|
self.onlyuse = drives
|
|
|
|
return retval
|
|
|
|
class Iscsi(commands.iscsi.F17_Iscsi):
|
|
def parse(self, args):
|
|
tg = commands.iscsi.F17_Iscsi.parse(self, args)
|
|
|
|
if tg.iface:
|
|
if not network.wait_for_network_devices([tg.iface]):
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("Network interface \"%(nic)s\" required by iSCSI \"%(iscsiTarget)s\" target is not up.") %
|
|
{"nic": tg.iface, "iscsiTarget": tg.target}))
|
|
|
|
mode = blivet.iscsi.iscsi.mode
|
|
if mode == "none":
|
|
if tg.iface:
|
|
blivet.iscsi.iscsi.create_interfaces(nm.nm_activated_devices())
|
|
elif ((mode == "bind" and not tg.iface)
|
|
or (mode == "default" and tg.iface)):
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("iscsi --iface must be specified (binding used) either for all targets or for none")))
|
|
|
|
try:
|
|
blivet.iscsi.iscsi.add_target(tg.ipaddr, tg.port, tg.user,
|
|
tg.password, tg.user_in,
|
|
tg.password_in,
|
|
target=tg.target,
|
|
iface=tg.iface)
|
|
log.info("added iscsi target %s at %s via %s", tg.target,
|
|
tg.ipaddr,
|
|
tg.iface)
|
|
except (IOError, ValueError) as e:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno, msg=str(e)))
|
|
|
|
return tg
|
|
|
|
class IscsiName(commands.iscsiname.FC6_IscsiName):
|
|
def parse(self, args):
|
|
retval = commands.iscsiname.FC6_IscsiName.parse(self, args)
|
|
|
|
blivet.iscsi.iscsi.initiator = self.iscsiname
|
|
return retval
|
|
|
|
class Lang(commands.lang.F19_Lang):
|
|
def execute(self, *args, **kwargs):
|
|
localization.write_language_configuration(self, iutil.getSysroot())
|
|
|
|
# no overrides needed here
|
|
Eula = commands.eula.F20_Eula
|
|
|
|
class LogVol(commands.logvol.F23_LogVol):
|
|
def execute(self, storage, ksdata, instClass):
|
|
for l in self.lvList:
|
|
l.execute(storage, ksdata, instClass)
|
|
|
|
if self.lvList:
|
|
grow_lvm(storage)
|
|
|
|
class LogVolData(commands.logvol.F23_LogVolData):
|
|
def execute(self, storage, ksdata, instClass):
|
|
devicetree = storage.devicetree
|
|
|
|
storage.do_autopart = False
|
|
|
|
# FIXME: we should be running sanityCheck on partitioning that is not ks
|
|
# autopart, but that's likely too invasive for #873135 at this moment
|
|
if self.mountpoint == "/boot" and blivet.arch.is_s390():
|
|
raise KickstartParseError(formatErrorMsg(self.lineno, msg="/boot can not be of type 'lvmlv' on s390x"))
|
|
|
|
# we might have truncated or otherwise changed the specified vg name
|
|
vgname = ksdata.onPart.get(self.vgname, self.vgname)
|
|
|
|
size = None
|
|
|
|
if self.percent:
|
|
size = Size(0)
|
|
|
|
if self.mountpoint == "swap":
|
|
ty = "swap"
|
|
self.mountpoint = ""
|
|
if self.recommended or self.hibernation:
|
|
disk_space = getAvailableDiskSpace(storage)
|
|
size = autopart.swap_suggestion(hibernation=self.hibernation, disk_space=disk_space)
|
|
self.grow = False
|
|
else:
|
|
if self.fstype != "":
|
|
ty = self.fstype
|
|
else:
|
|
ty = storage.default_fstype
|
|
|
|
if size is None and not self.preexist:
|
|
if not self.size:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg="Size can not be decided on from kickstart nor obtained from device."))
|
|
try:
|
|
size = Size("%d MiB" % self.size)
|
|
except ValueError:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg="The size \"%s\" is invalid." % self.size))
|
|
|
|
if self.thin_pool:
|
|
self.mountpoint = ""
|
|
ty = None
|
|
|
|
if self.mountpoint.startswith('/') and not self.fsopts:
|
|
# enable discard for normal filesystems in dom0
|
|
self.fsopts = "defaults,discard"
|
|
|
|
# Sanity check mountpoint
|
|
if self.mountpoint != "" and self.mountpoint[0] != '/':
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("The mount point \"%s\" is not valid. It must start with a /.") % self.mountpoint))
|
|
|
|
# Check that the VG this LV is a member of has already been specified.
|
|
vg = devicetree.get_device_by_name(vgname)
|
|
if not vg:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("No volume group exists with the name \"%s\". Specify volume groups before logical volumes.") % self.vgname))
|
|
|
|
# If cache PVs specified, check that they belong to the same VG this LV is a member of
|
|
if self.cache_pvs:
|
|
pv_devices = (lookupAlias(devicetree, pv) for pv in self.cache_pvs)
|
|
if not all(pv in vg.pvs for pv in pv_devices):
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("Cache PVs must belong to the same VG as the cached LV")))
|
|
|
|
pool = None
|
|
if self.thin_volume:
|
|
pool = devicetree.get_device_by_name("%s-%s" % (vg.name, self.pool_name))
|
|
if not pool:
|
|
err = formatErrorMsg(self.lineno,
|
|
msg=_("No thin pool exists with the name \"%s\". Specify thin pools before thin volumes.") % self.pool_name)
|
|
raise KickstartParseError(err)
|
|
|
|
# If this specifies an existing request that we should not format,
|
|
# quit here after setting up enough information to mount it later.
|
|
if not self.format:
|
|
if not self.name:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("logvol --noformat must also use the --name= option.")))
|
|
|
|
dev = devicetree.get_device_by_name("%s-%s" % (vg.name, self.name))
|
|
if not dev:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("Logical volume \"%s\" given in logvol command does not exist.") % self.name))
|
|
|
|
if self.resize:
|
|
size = dev.raw_device.align_target_size(size)
|
|
if size < dev.currentSize:
|
|
# shrink
|
|
try:
|
|
devicetree.actions.add(ActionResizeFormat(dev, size))
|
|
devicetree.actions.add(ActionResizeDevice(dev, size))
|
|
except ValueError:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("Target size \"%(size)s\" for device \"%(device)s\" is invalid.") %
|
|
{"size": self.size, "device": dev.name}))
|
|
else:
|
|
# grow
|
|
try:
|
|
devicetree.actions.add(ActionResizeDevice(dev, size))
|
|
devicetree.actions.add(ActionResizeFormat(dev, size))
|
|
except ValueError:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("Target size \"%(size)s\" for device \"%(device)s\" is invalid.") %
|
|
{"size": self.size, "device": dev.name}))
|
|
|
|
dev.format.mountpoint = self.mountpoint
|
|
dev.format.mountopts = self.fsopts
|
|
if ty == "swap":
|
|
storage.add_fstab_swap(dev)
|
|
return
|
|
|
|
# Make sure this LV name is not already used in the requested VG.
|
|
if not self.preexist:
|
|
tmp = devicetree.get_device_by_name("%s-%s" % (vg.name, self.name))
|
|
if tmp:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("Logical volume name \"%(logvol)s\" is already in use in volume group \"%(volgroup)s\".") %
|
|
{"logvol": self.name, "volgroup": vg.name}))
|
|
|
|
if not self.percent and size and not self.grow and size < vg.pe_size:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("Logical volume size \"%(logvolSize)s\" must be larger than the volume group extent size of \"%(extentSize)s\".") %
|
|
{"logvolSize": size, "extentSize": vg.pe_size}))
|
|
|
|
# Now get a format to hold a lot of these extra values.
|
|
fmt = get_format(ty,
|
|
mountpoint=self.mountpoint,
|
|
label=self.label,
|
|
fsprofile=self.fsprofile,
|
|
create_options=self.mkfsopts,
|
|
mountopts=self.fsopts)
|
|
if not fmt.type and not self.thin_pool:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("The \"%s\" file system type is not supported.") % ty))
|
|
|
|
add_fstab_swap = None
|
|
# If we were given a pre-existing LV to create a filesystem on, we need
|
|
# to verify it and its VG exists and then schedule a new format action
|
|
# to take place there. Also, we only support a subset of all the
|
|
# options on pre-existing LVs.
|
|
if self.preexist:
|
|
device = devicetree.get_device_by_name("%s-%s" % (vg.name, self.name))
|
|
if not device:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("Logical volume \"%s\" given in logvol command does not exist.") % self.name))
|
|
|
|
storage.devicetree.recursive_remove(device, remove_device=False)
|
|
|
|
if self.resize:
|
|
size = device.raw_device.align_target_size(size)
|
|
try:
|
|
devicetree.actions.add(ActionResizeDevice(device, size))
|
|
except ValueError:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("Target size \"%(size)s\" for device \"%(device)s\" is invalid.") %
|
|
{"size": self.size, "device": device.name}))
|
|
|
|
devicetree.actions.add(ActionCreateFormat(device, fmt))
|
|
if ty == "swap":
|
|
add_fstab_swap = device
|
|
else:
|
|
# If a previous device has claimed this mount point, delete the
|
|
# old one.
|
|
try:
|
|
if self.mountpoint:
|
|
device = storage.mountpoints[self.mountpoint]
|
|
storage.destroy_device(device)
|
|
except KeyError:
|
|
pass
|
|
|
|
if self.thin_volume:
|
|
parents = [pool]
|
|
else:
|
|
parents = [vg]
|
|
|
|
pool_args = {}
|
|
if self.thin_pool:
|
|
if self.profile:
|
|
matching = (p for p in KNOWN_THPOOL_PROFILES if p.name == self.profile)
|
|
profile = next(matching, None)
|
|
if profile:
|
|
pool_args["profile"] = profile
|
|
else:
|
|
log.warning("No matching profile for %s found in LVM configuration", self.profile)
|
|
if self.metadata_size:
|
|
pool_args["metadatasize"] = Size("%d MiB" % self.metadata_size)
|
|
if self.chunk_size:
|
|
pool_args["chunksize"] = Size("%d KiB" % self.chunk_size)
|
|
|
|
if self.maxSizeMB:
|
|
try:
|
|
maxsize = Size("%d MiB" % self.maxSizeMB)
|
|
except ValueError:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg="The maximum size \"%s\" is invalid." % self.maxSizeMB))
|
|
else:
|
|
maxsize = None
|
|
|
|
if self.cache_size and self.cache_pvs:
|
|
pv_devices = [lookupAlias(devicetree, pv) for pv in self.cache_pvs]
|
|
cache_size = Size("%d MiB" % self.cache_size)
|
|
cache_mode = self.cache_mode or None
|
|
cache_request = LVMCacheRequest(cache_size, pv_devices, cache_mode)
|
|
else:
|
|
cache_request = None
|
|
|
|
try:
|
|
request = storage.new_lv(fmt=fmt,
|
|
name=self.name,
|
|
parents=parents,
|
|
size=size,
|
|
thin_pool=self.thin_pool,
|
|
thin_volume=self.thin_volume,
|
|
grow=self.grow,
|
|
maxsize=maxsize,
|
|
percent=self.percent,
|
|
cache_request=cache_request,
|
|
**pool_args)
|
|
except (StorageError, ValueError) as e:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno, msg=str(e)))
|
|
|
|
storage.create_device(request)
|
|
if ty == "swap":
|
|
add_fstab_swap = request
|
|
|
|
if self.encrypted:
|
|
if self.passphrase and not storage.encryption_passphrase:
|
|
storage.encryption_passphrase = self.passphrase
|
|
|
|
# try to use the global passphrase if available
|
|
# XXX: we require the LV/part with --passphrase to be processed
|
|
# before this one to setup the storage.encryption_passphrase
|
|
self.passphrase = self.passphrase or storage.encryption_passphrase
|
|
|
|
cert = getEscrowCertificate(storage.escrow_certificates, self.escrowcert)
|
|
if self.preexist:
|
|
luksformat = fmt
|
|
device.format = get_format("luks", passphrase=self.passphrase, device=device.path,
|
|
cipher=self.cipher,
|
|
escrow_cert=cert,
|
|
add_backup_passphrase=self.backuppassphrase)
|
|
luksdev = LUKSDevice("luks%d" % storage.next_id,
|
|
fmt=luksformat,
|
|
parents=device)
|
|
else:
|
|
luksformat = request.format
|
|
request.format = get_format("luks", passphrase=self.passphrase,
|
|
cipher=self.cipher,
|
|
escrow_cert=cert,
|
|
add_backup_passphrase=self.backuppassphrase,
|
|
min_luks_entropy=MIN_CREATE_ENTROPY)
|
|
luksdev = LUKSDevice("luks%d" % storage.next_id,
|
|
fmt=luksformat,
|
|
parents=request)
|
|
if ty == "swap":
|
|
# swap is on the LUKS device not on the LUKS' parent device,
|
|
# override the info here
|
|
add_fstab_swap = luksdev
|
|
|
|
storage.create_device(luksdev)
|
|
|
|
if add_fstab_swap:
|
|
storage.add_fstab_swap(add_fstab_swap)
|
|
|
|
class Logging(commands.logging.FC6_Logging):
|
|
def execute(self, *args):
|
|
if logger.loglevel == DEFAULT_LEVEL:
|
|
# not set from the command line
|
|
level = logLevelMap[self.level]
|
|
logger.loglevel = level
|
|
setHandlersLevel(log, level)
|
|
setHandlersLevel(storage_log, level)
|
|
|
|
if logger.remote_syslog == None and len(self.host) > 0:
|
|
# not set from the command line, ok to use kickstart
|
|
remote_server = self.host
|
|
if self.port:
|
|
remote_server = "%s:%s" %(self.host, self.port)
|
|
logger.updateRemote(remote_server)
|
|
|
|
class Network(commands.network.F25_Network):
|
|
def __init__(self, *args, **kwargs):
|
|
commands.network.F25_Network.__init__(self, *args, **kwargs)
|
|
self.packages = []
|
|
|
|
def parse(self, args):
|
|
nd = commands.network.F25_Network.parse(self, args)
|
|
setting_only_hostname = nd.hostname and len(args) <= 2
|
|
if not setting_only_hostname:
|
|
if not nd.device:
|
|
ksdevice = flags.cmdline.get('ksdevice')
|
|
if ksdevice:
|
|
log.info('network: setting %s from ksdevice for missing kickstart --device', ksdevice)
|
|
nd.device = ksdevice
|
|
else:
|
|
log.info('network: setting "link" for missing --device specification in kickstart')
|
|
nd.device = "link"
|
|
return nd
|
|
|
|
def setup(self):
|
|
if network.is_using_team_device():
|
|
self.packages = ["teamd"]
|
|
|
|
def execute(self, storage, ksdata, instClass):
|
|
network.write_network_config(storage, ksdata, instClass, iutil.getSysroot())
|
|
|
|
class Partition(commands.partition.F23_Partition):
|
|
def execute(self, storage, ksdata, instClass):
|
|
for p in self.partitions:
|
|
p.execute(storage, ksdata, instClass)
|
|
|
|
if self.partitions:
|
|
do_partitioning(storage)
|
|
|
|
class PartitionData(commands.partition.F23_PartData):
|
|
def execute(self, storage, ksdata, instClass):
|
|
devicetree = storage.devicetree
|
|
kwargs = {}
|
|
|
|
storage.do_autopart = False
|
|
|
|
if self.onbiosdisk != "":
|
|
# edd_dict is only modified during storage.reset(), so don't do that
|
|
# while executing storage.
|
|
for (disk, biosdisk) in storage.edd_dict.items():
|
|
if "%x" % biosdisk == self.onbiosdisk:
|
|
self.disk = disk
|
|
break
|
|
|
|
if not self.disk:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("No disk found for specified BIOS disk \"%s\".") % self.onbiosdisk))
|
|
|
|
size = None
|
|
|
|
if self.mountpoint == "swap":
|
|
ty = "swap"
|
|
self.mountpoint = ""
|
|
if self.recommended or self.hibernation:
|
|
disk_space = getAvailableDiskSpace(storage)
|
|
size = autopart.swap_suggestion(hibernation=self.hibernation, disk_space=disk_space)
|
|
self.grow = False
|
|
# if people want to specify no mountpoint for some reason, let them
|
|
# this is really needed for pSeries boot partitions :(
|
|
elif self.mountpoint == "None":
|
|
self.mountpoint = ""
|
|
if self.fstype:
|
|
ty = self.fstype
|
|
else:
|
|
ty = storage.default_fstype
|
|
elif self.mountpoint == 'appleboot':
|
|
ty = "appleboot"
|
|
self.mountpoint = ""
|
|
elif self.mountpoint == 'prepboot':
|
|
ty = "prepboot"
|
|
self.mountpoint = ""
|
|
elif self.mountpoint == 'biosboot':
|
|
ty = "biosboot"
|
|
self.mountpoint = ""
|
|
elif self.mountpoint.startswith("raid."):
|
|
ty = "mdmember"
|
|
kwargs["name"] = self.mountpoint
|
|
self.mountpoint = ""
|
|
|
|
if devicetree.get_device_by_name(kwargs["name"]):
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("RAID partition \"%s\" is defined multiple times.") % kwargs["name"]))
|
|
|
|
if self.onPart:
|
|
ksdata.onPart[kwargs["name"]] = self.onPart
|
|
elif self.mountpoint.startswith("pv."):
|
|
ty = "lvmpv"
|
|
kwargs["name"] = self.mountpoint
|
|
self.mountpoint = ""
|
|
|
|
if devicetree.get_device_by_name(kwargs["name"]):
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("PV partition \"%s\" is defined multiple times.") % kwargs["name"]))
|
|
|
|
if self.onPart:
|
|
ksdata.onPart[kwargs["name"]] = self.onPart
|
|
elif self.mountpoint.startswith("btrfs."):
|
|
ty = "btrfs"
|
|
kwargs["name"] = self.mountpoint
|
|
self.mountpoint = ""
|
|
|
|
if devicetree.get_device_by_name(kwargs["name"]):
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("Btrfs partition \"%s\" is defined multiple times.") % kwargs["name"]))
|
|
|
|
if self.onPart:
|
|
ksdata.onPart[kwargs["name"]] = self.onPart
|
|
elif self.mountpoint == "/boot/efi":
|
|
if blivet.arch.is_mactel():
|
|
ty = "macefi"
|
|
else:
|
|
ty = "EFI System Partition"
|
|
self.fsopts = "defaults,uid=0,gid=0,umask=077,shortname=winnt"
|
|
else:
|
|
if self.fstype != "":
|
|
ty = self.fstype
|
|
elif self.mountpoint == "/boot":
|
|
ty = storage.default_boot_fstype
|
|
else:
|
|
ty = storage.default_fstype
|
|
|
|
if self.mountpoint.startswith('/') and not self.fsopts:
|
|
# enable discard for normal filesystems in dom0
|
|
self.fsopts = "defaults,discard"
|
|
|
|
if not size and self.size:
|
|
try:
|
|
size = Size("%d MiB" % self.size)
|
|
except ValueError:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("The size \"%s\" is invalid.") % self.size))
|
|
|
|
# If this specified an existing request that we should not format,
|
|
# quit here after setting up enough information to mount it later.
|
|
if not self.format:
|
|
if not self.onPart:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("part --noformat must also use the --onpart option.")))
|
|
|
|
dev = devicetree.resolve_device(self.onPart)
|
|
if not dev:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("Partition \"%s\" given in part command does not exist.") % self.onPart))
|
|
|
|
if self.resize:
|
|
size = dev.raw_device.align_target_size(size)
|
|
if size < dev.currentSize:
|
|
# shrink
|
|
try:
|
|
devicetree.actions.add(ActionResizeFormat(dev, size))
|
|
devicetree.actions.add(ActionResizeDevice(dev, size))
|
|
except ValueError:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("Target size \"%(size)s\" for device \"%(device)s\" is invalid.") %
|
|
{"size": self.size, "device": dev.name}))
|
|
else:
|
|
# grow
|
|
try:
|
|
devicetree.actions.add(ActionResizeDevice(dev, size))
|
|
devicetree.actions.add(ActionResizeFormat(dev, size))
|
|
except ValueError:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("Target size \"%(size)s\" for device \"%(device)s\" is invalid.") %
|
|
{"size": self.size, "device": dev.name}))
|
|
|
|
dev.format.mountpoint = self.mountpoint
|
|
dev.format.mountopts = self.fsopts
|
|
if ty == "swap":
|
|
storage.add_fstab_swap(dev)
|
|
return
|
|
|
|
# Now get a format to hold a lot of these extra values.
|
|
kwargs["fmt"] = get_format(ty,
|
|
mountpoint=self.mountpoint,
|
|
label=self.label,
|
|
fsprofile=self.fsprofile,
|
|
mountopts=self.fsopts,
|
|
create_options=self.mkfsopts,
|
|
size=size)
|
|
if not kwargs["fmt"].type:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("The \"%s\" file system type is not supported.") % ty))
|
|
|
|
# If we were given a specific disk to create the partition on, verify
|
|
# that it exists first. If it doesn't exist, see if it exists with
|
|
# mapper/ on the front. If that doesn't exist either, it's an error.
|
|
if self.disk:
|
|
disk = devicetree.resolve_device(self.disk)
|
|
# if this is a multipath member promote it to the real mpath
|
|
if disk and disk.format.type == "multipath_member":
|
|
mpath_device = disk.children[0]
|
|
storage_log.info("kickstart: part: promoting %s to %s",
|
|
disk.name, mpath_device.name)
|
|
disk = mpath_device
|
|
if not disk:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("Disk \"%s\" given in part command does not exist.") % self.disk))
|
|
if not disk.partitionable:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("Cannot install to unpartitionable device \"%s\".") % self.disk))
|
|
|
|
should_clear = storage.should_clear(disk)
|
|
if disk and (disk.partitioned or should_clear):
|
|
kwargs["parents"] = [disk]
|
|
elif disk:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("Disk \"%s\" in part command is not partitioned.") % self.disk))
|
|
|
|
if not kwargs["parents"]:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("Disk \"%s\" given in part command does not exist.") % self.disk))
|
|
|
|
kwargs["grow"] = self.grow
|
|
kwargs["size"] = size
|
|
if self.maxSizeMB:
|
|
try:
|
|
maxsize = Size("%d MiB" % self.maxSizeMB)
|
|
except ValueError:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("The maximum size \"%s\" is invalid.") % self.maxSizeMB))
|
|
else:
|
|
maxsize = None
|
|
|
|
kwargs["maxsize"] = maxsize
|
|
|
|
kwargs["primary"] = self.primOnly
|
|
|
|
add_fstab_swap = None
|
|
# If we were given a pre-existing partition to create a filesystem on,
|
|
# we need to verify it exists and then schedule a new format action to
|
|
# take place there. Also, we only support a subset of all the options
|
|
# on pre-existing partitions.
|
|
if self.onPart:
|
|
device = devicetree.resolve_device(self.onPart)
|
|
if not device:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("Partition \"%s\" given in part command does not exist.") % self.onPart))
|
|
|
|
storage.devicetree.recursive_remove(device, remove_device=False)
|
|
if self.resize:
|
|
size = device.raw_device.align_target_size(size)
|
|
try:
|
|
devicetree.actions.add(ActionResizeDevice(device, size))
|
|
except ValueError:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("Target size \"%(size)s\" for device \"%(device)s\" is invalid.") %
|
|
{"size": self.size, "device": device.name}))
|
|
|
|
devicetree.actions.add(ActionCreateFormat(device, kwargs["fmt"]))
|
|
if ty == "swap":
|
|
add_fstab_swap = device
|
|
# tmpfs mounts are not disks and don't occupy a disk partition,
|
|
# so handle them here
|
|
elif self.fstype == "tmpfs":
|
|
try:
|
|
request = storage.new_tmp_fs(**kwargs)
|
|
except (StorageError, ValueError) as e:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno, msg=str(e)))
|
|
storage.create_device(request)
|
|
else:
|
|
# If a previous device has claimed this mount point, delete the
|
|
# old one.
|
|
try:
|
|
if self.mountpoint:
|
|
device = storage.mountpoints[self.mountpoint]
|
|
storage.destroy_device(device)
|
|
except KeyError:
|
|
pass
|
|
|
|
try:
|
|
request = storage.new_partition(**kwargs)
|
|
except (StorageError, ValueError) as e:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno, msg=str(e)))
|
|
|
|
storage.create_device(request)
|
|
if ty == "swap":
|
|
add_fstab_swap = request
|
|
|
|
if self.encrypted:
|
|
if self.passphrase and not storage.encryption_passphrase:
|
|
storage.encryption_passphrase = self.passphrase
|
|
|
|
# try to use the global passphrase if available
|
|
# XXX: we require the LV/part with --passphrase to be processed
|
|
# before this one to setup the storage.encryption_passphrase
|
|
self.passphrase = self.passphrase or storage.encryption_passphrase
|
|
|
|
cert = getEscrowCertificate(storage.escrow_certificates, self.escrowcert)
|
|
if self.onPart:
|
|
luksformat = kwargs["fmt"]
|
|
device.format = get_format("luks", passphrase=self.passphrase, device=device.path,
|
|
cipher=self.cipher,
|
|
escrow_cert=cert,
|
|
add_backup_passphrase=self.backuppassphrase,
|
|
min_luks_entropy=MIN_CREATE_ENTROPY)
|
|
luksdev = LUKSDevice("luks%d" % storage.next_id,
|
|
fmt=luksformat,
|
|
parents=device)
|
|
else:
|
|
luksformat = request.format
|
|
request.format = get_format("luks", passphrase=self.passphrase,
|
|
cipher=self.cipher,
|
|
escrow_cert=cert,
|
|
add_backup_passphrase=self.backuppassphrase,
|
|
min_luks_entropy=MIN_CREATE_ENTROPY)
|
|
luksdev = LUKSDevice("luks%d" % storage.next_id,
|
|
fmt=luksformat,
|
|
parents=request)
|
|
|
|
if ty == "swap":
|
|
# swap is on the LUKS device not on the LUKS' parent device,
|
|
# override the info here
|
|
add_fstab_swap = luksdev
|
|
|
|
storage.create_device(luksdev)
|
|
|
|
if add_fstab_swap:
|
|
storage.add_fstab_swap(add_fstab_swap)
|
|
|
|
class Raid(commands.raid.F25_Raid):
|
|
def execute(self, storage, ksdata, instClass):
|
|
for r in self.raidList:
|
|
r.execute(storage, ksdata, instClass)
|
|
|
|
class RaidData(commands.raid.F25_RaidData):
|
|
def execute(self, storage, ksdata, instClass):
|
|
raidmems = []
|
|
devicetree = storage.devicetree
|
|
devicename = self.device
|
|
if self.preexist:
|
|
device = devicetree.resolve_device(devicename)
|
|
if device:
|
|
devicename = device.name
|
|
|
|
kwargs = {}
|
|
|
|
storage.do_autopart = False
|
|
|
|
if self.mountpoint == "swap":
|
|
ty = "swap"
|
|
self.mountpoint = ""
|
|
elif self.mountpoint.startswith("pv."):
|
|
ty = "lvmpv"
|
|
kwargs["name"] = self.mountpoint
|
|
ksdata.onPart[kwargs["name"]] = devicename
|
|
|
|
if devicetree.get_device_by_name(kwargs["name"]):
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("PV partition \"%s\" is defined multiple times.") % kwargs["name"]))
|
|
|
|
self.mountpoint = ""
|
|
elif self.mountpoint.startswith("btrfs."):
|
|
ty = "btrfs"
|
|
kwargs["name"] = self.mountpoint
|
|
ksdata.onPart[kwargs["name"]] = devicename
|
|
|
|
if devicetree.get_device_by_name(kwargs["name"]):
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("Btrfs partition \"%s\" is defined multiple times.") % kwargs["name"]))
|
|
|
|
self.mountpoint = ""
|
|
else:
|
|
if self.fstype != "":
|
|
ty = self.fstype
|
|
elif self.mountpoint == "/boot" and \
|
|
"mdarray" in storage.bootloader.stage2_device_types:
|
|
ty = storage.default_boot_fstype
|
|
else:
|
|
ty = storage.default_fstype
|
|
|
|
if self.mountpoint.startswith('/') and not self.fsopts:
|
|
# enable discard for normal filesystems in dom0
|
|
self.fsopts = "defaults,discard"
|
|
|
|
# Sanity check mountpoint
|
|
if self.mountpoint != "" and self.mountpoint[0] != '/':
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("The mount point \"%s\" is not valid. It must start with a /.") % self.mountpoint))
|
|
|
|
# If this specifies an existing request that we should not format,
|
|
# quit here after setting up enough information to mount it later.
|
|
if not self.format:
|
|
if not devicename:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("raid --noformat must also use the --device option.")))
|
|
|
|
dev = devicetree.get_device_by_name(devicename)
|
|
if not dev:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("RAID device \"%s\" given in raid command does not exist.") % devicename))
|
|
|
|
dev.format.mountpoint = self.mountpoint
|
|
dev.format.mountopts = self.fsopts
|
|
if ty == "swap":
|
|
storage.add_fstab_swap(dev)
|
|
return
|
|
|
|
# Get a list of all the RAID members.
|
|
for member in self.members:
|
|
dev = devicetree.resolve_device(member)
|
|
if not dev:
|
|
# if member is using --onpart, use original device
|
|
mem = ksdata.onPart.get(member, member)
|
|
dev = devicetree.resolve_device(mem) or lookupAlias(devicetree, member)
|
|
if dev and dev.format.type == "luks":
|
|
try:
|
|
dev = dev.children[0]
|
|
except IndexError:
|
|
dev = None
|
|
|
|
if dev and dev.format.type != "mdmember":
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("RAID device \"%(device)s\" has a format of \"%(format)s\", but should have a format of \"mdmember\".") %
|
|
{"device": member, "format": dev.format.type}))
|
|
|
|
if not dev:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("Tried to use undefined partition \"%s\" in RAID specification.") % member))
|
|
|
|
raidmems.append(dev)
|
|
|
|
# Now get a format to hold a lot of these extra values.
|
|
kwargs["fmt"] = get_format(ty,
|
|
label=self.label,
|
|
fsprofile=self.fsprofile,
|
|
mountpoint=self.mountpoint,
|
|
mountopts=self.fsopts,
|
|
create_options=self.mkfsopts)
|
|
if not kwargs["fmt"].type:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("The \"%s\" file system type is not supported.") % ty))
|
|
|
|
kwargs["name"] = devicename
|
|
kwargs["level"] = self.level
|
|
kwargs["parents"] = raidmems
|
|
kwargs["member_devices"] = len(raidmems) - self.spares
|
|
kwargs["total_devices"] = len(raidmems)
|
|
|
|
if self.chunk_size:
|
|
kwargs["chunk_size"] = Size("%d KiB" % self.chunk_size)
|
|
|
|
add_fstab_swap = None
|
|
|
|
# If we were given a pre-existing RAID to create a filesystem on,
|
|
# we need to verify it exists and then schedule a new format action
|
|
# to take place there. Also, we only support a subset of all the
|
|
# options on pre-existing RAIDs.
|
|
if self.preexist:
|
|
device = devicetree.get_device_by_name(devicename)
|
|
if not device:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("RAID volume \"%s\" specified with --useexisting does not exist.") % devicename))
|
|
|
|
storage.devicetree.recursive_remove(device, remove_device=False)
|
|
devicetree.actions.add(ActionCreateFormat(device, kwargs["fmt"]))
|
|
if ty == "swap":
|
|
add_fstab_swap = device
|
|
else:
|
|
if devicename and devicename in (a.name for a in storage.mdarrays):
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("The RAID volume name \"%s\" is already in use.") % devicename))
|
|
|
|
# If a previous device has claimed this mount point, delete the
|
|
# old one.
|
|
try:
|
|
if self.mountpoint:
|
|
device = storage.mountpoints[self.mountpoint]
|
|
storage.destroy_device(device)
|
|
except KeyError:
|
|
pass
|
|
|
|
try:
|
|
request = storage.new_mdarray(**kwargs)
|
|
except (StorageError, ValueError) as e:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno, msg=str(e)))
|
|
|
|
storage.create_device(request)
|
|
if ty == "swap":
|
|
add_fstab_swap = request
|
|
|
|
if self.encrypted:
|
|
if self.passphrase and not storage.encryption_passphrase:
|
|
storage.encryption_passphrase = self.passphrase
|
|
|
|
cert = getEscrowCertificate(storage.escrow_certificates, self.escrowcert)
|
|
if self.preexist:
|
|
luksformat = kwargs["fmt"]
|
|
device.format = get_format("luks", passphrase=self.passphrase, device=device.path,
|
|
cipher=self.cipher,
|
|
escrow_cert=cert,
|
|
add_backup_passphrase=self.backuppassphrase)
|
|
luksdev = LUKSDevice("luks%d" % storage.next_id,
|
|
fmt=luksformat,
|
|
parents=device)
|
|
else:
|
|
luksformat = request.format
|
|
request.format = get_format("luks", passphrase=self.passphrase,
|
|
cipher=self.cipher,
|
|
escrow_cert=cert,
|
|
add_backup_passphrase=self.backuppassphrase)
|
|
luksdev = LUKSDevice("luks%d" % storage.next_id,
|
|
fmt=luksformat,
|
|
parents=request)
|
|
|
|
if ty == "swap":
|
|
# swap is on the LUKS device instead of the parent device,
|
|
# override the device here
|
|
add_fstab_swap = luksdev
|
|
|
|
storage.create_device(luksdev)
|
|
|
|
if add_fstab_swap:
|
|
storage.add_fstab_swap(add_fstab_swap)
|
|
|
|
class RepoData(commands.repo.F21_RepoData):
|
|
def __init__(self, *args, **kwargs):
|
|
""" Add enabled kwarg
|
|
|
|
:param enabled: The repo has been enabled
|
|
:type enabled: bool
|
|
"""
|
|
self.enabled = kwargs.pop("enabled", True)
|
|
self.repo_id = kwargs.pop("repo_id", None)
|
|
|
|
commands.repo.F21_RepoData.__init__(self, *args, **kwargs)
|
|
|
|
class ReqPart(commands.reqpart.F23_ReqPart):
|
|
def execute(self, storage, ksdata, instClass):
|
|
from blivet.autopart import do_reqpart
|
|
|
|
if not self.reqpart:
|
|
return
|
|
|
|
reqs = platform.set_platform_bootloader_reqs()
|
|
if self.addBoot:
|
|
bootPartitions = platform.set_platform_boot_partition()
|
|
|
|
# blivet doesn't know this - anaconda sets up the default boot fstype
|
|
# in various places in this file, as well as in setDefaultPartitioning
|
|
# in the install classes. We need to duplicate that here.
|
|
for part in bootPartitions:
|
|
if part.mountpoint == "/boot":
|
|
part.fstype = storage.default_boot_fstype
|
|
|
|
reqs += bootPartitions
|
|
|
|
do_reqpart(storage, reqs)
|
|
|
|
class RootPw(commands.rootpw.F18_RootPw):
|
|
def __init__(self, writePriority=100, *args, **kwargs):
|
|
if 'lock' not in kwargs:
|
|
kwargs['lock'] = True
|
|
super(RootPw, self).__init__(writePriority=writePriority, *args, **kwargs)
|
|
|
|
def execute(self, storage, ksdata, instClass, users):
|
|
if not self.password and not flags.automatedInstall:
|
|
self.lock = True
|
|
|
|
algo = getPassAlgo(ksdata.authconfig.authconfig)
|
|
users.setRootPassword(self.password, self.isCrypted, self.lock, algo, iutil.getSysroot())
|
|
|
|
class SELinux(commands.selinux.FC3_SELinux):
|
|
def execute(self, *args):
|
|
selinux_states = {SELINUX_DISABLED: "disabled",
|
|
SELINUX_ENFORCING: "enforcing",
|
|
SELINUX_PERMISSIVE: "permissive"}
|
|
|
|
if self.selinux is None:
|
|
# Use the defaults set by the installed (or not) selinux package
|
|
return
|
|
elif self.selinux not in selinux_states:
|
|
log.error("unknown selinux state: %s", self.selinux)
|
|
return
|
|
|
|
try:
|
|
selinux_cfg = SimpleConfigFile(iutil.getSysroot()+"/etc/selinux/config")
|
|
selinux_cfg.read()
|
|
selinux_cfg.set(("SELINUX", selinux_states[self.selinux]))
|
|
selinux_cfg.write()
|
|
except IOError as msg:
|
|
log.error("Error setting selinux mode: %s", msg)
|
|
|
|
class Services(commands.services.FC6_Services):
|
|
def execute(self, storage, ksdata, instClass):
|
|
for svc in self.disabled:
|
|
iutil.disable_service(svc)
|
|
|
|
for svc in self.enabled:
|
|
iutil.enable_service(svc)
|
|
|
|
class SshKey(commands.sshkey.F22_SshKey):
|
|
def execute(self, storage, ksdata, instClass, users):
|
|
for usr in self.sshUserList:
|
|
users.setUserSshKey(usr.username, usr.key)
|
|
|
|
class Timezone(commands.timezone.F25_Timezone):
|
|
def __init__(self, *args):
|
|
commands.timezone.F25_Timezone.__init__(self, *args)
|
|
|
|
self._added_chrony = False
|
|
self._enabled_chrony = False
|
|
self._disabled_chrony = False
|
|
|
|
def setup(self, ksdata):
|
|
### Skip the whole NTP setup in Qubes dom0
|
|
return
|
|
|
|
# do not install and use NTP package
|
|
if self.nontp or NTP_PACKAGE in ksdata.packages.excludedList:
|
|
if iutil.service_running(NTP_SERVICE) and \
|
|
can_touch_runtime_system("stop NTP service"):
|
|
ret = iutil.stop_service(NTP_SERVICE)
|
|
if ret != 0:
|
|
log.error("Failed to stop NTP service")
|
|
|
|
if self._added_chrony and NTP_PACKAGE in ksdata.packages.packageList:
|
|
ksdata.packages.packageList.remove(NTP_PACKAGE)
|
|
self._added_chrony = False
|
|
|
|
# Both un-enable and disable chrony, because sometimes it's installed
|
|
# off by default (packages) and sometimes not (liveimg).
|
|
if self._enabled_chrony and NTP_SERVICE in ksdata.services.enabled:
|
|
ksdata.services.enabled.remove(NTP_SERVICE)
|
|
self._enabled_chrony = False
|
|
|
|
if NTP_SERVICE not in ksdata.services.disabled:
|
|
ksdata.services.disabled.append(NTP_SERVICE)
|
|
self._disabled_chrony = True
|
|
# install and use NTP package
|
|
else:
|
|
if not iutil.service_running(NTP_SERVICE) and \
|
|
can_touch_runtime_system("start NTP service"):
|
|
ret = iutil.start_service(NTP_SERVICE)
|
|
if ret != 0:
|
|
log.error("Failed to start NTP service")
|
|
|
|
if not NTP_PACKAGE in ksdata.packages.packageList:
|
|
ksdata.packages.packageList.append(NTP_PACKAGE)
|
|
self._added_chrony = True
|
|
|
|
if self._disabled_chrony and NTP_SERVICE in ksdata.services.disabled:
|
|
ksdata.services.disabled.remove(NTP_SERVICE)
|
|
self._disabled_chrony = False
|
|
|
|
if not NTP_SERVICE in ksdata.services.enabled and \
|
|
not NTP_SERVICE in ksdata.services.disabled:
|
|
ksdata.services.enabled.append(NTP_SERVICE)
|
|
self._enabled_chrony = True
|
|
|
|
def execute(self, *args):
|
|
# write out timezone configuration
|
|
if not timezone.is_valid_timezone(self.timezone):
|
|
# this should never happen, but for pity's sake
|
|
log.warning("Timezone %s set in kickstart is not valid, falling "\
|
|
"back to default (America/New_York).", self.timezone)
|
|
self.timezone = "America/New_York"
|
|
|
|
timezone.write_timezone_config(self, iutil.getSysroot())
|
|
|
|
# write out NTP configuration (if set) and --nontp is not used
|
|
if not self.nontp and self.ntpservers:
|
|
chronyd_conf_path = os.path.normpath(iutil.getSysroot() + ntp.NTP_CONFIG_FILE)
|
|
pools, servers = ntp.internal_to_pools_and_servers(self.ntpservers)
|
|
if os.path.exists(chronyd_conf_path):
|
|
log.debug("Modifying installed chrony configuration")
|
|
try:
|
|
ntp.save_servers_to_config(pools, servers, conf_file_path=chronyd_conf_path)
|
|
except ntp.NTPconfigError as ntperr:
|
|
log.warning("Failed to save NTP configuration: %s", ntperr)
|
|
# use chrony conf file from installation environment when
|
|
# chrony is not installed (chrony conf file is missing)
|
|
else:
|
|
log.debug("Creating chrony configuration based on the "
|
|
"configuration from installation environment")
|
|
try:
|
|
ntp.save_servers_to_config(pools, servers,
|
|
conf_file_path=ntp.NTP_CONFIG_FILE,
|
|
out_file_path=chronyd_conf_path)
|
|
except ntp.NTPconfigError as ntperr:
|
|
log.warning("Failed to save NTP configuration without chrony package: %s", ntperr)
|
|
|
|
class User(commands.user.F19_User):
|
|
def execute(self, storage, ksdata, instClass, users):
|
|
algo = getPassAlgo(ksdata.authconfig.authconfig)
|
|
|
|
for usr in self.userList:
|
|
kwargs = usr.__dict__
|
|
kwargs.update({"algo": algo, "root": iutil.getSysroot()})
|
|
|
|
# If the user password came from a kickstart and it is blank we
|
|
# need to make sure the account is locked, not created with an
|
|
# empty password.
|
|
if ksdata.user.seen and kwargs.get("password", "") == "":
|
|
kwargs["password"] = None
|
|
try:
|
|
users.createUser(usr.name, **kwargs)
|
|
except ValueError as e:
|
|
log.warning(str(e))
|
|
|
|
class VolGroup(commands.volgroup.F21_VolGroup):
|
|
def execute(self, storage, ksdata, instClass):
|
|
for v in self.vgList:
|
|
v.execute(storage, ksdata, instClass)
|
|
|
|
class VolGroupData(commands.volgroup.F21_VolGroupData):
|
|
def execute(self, storage, ksdata, instClass):
|
|
pvs = []
|
|
|
|
devicetree = storage.devicetree
|
|
|
|
storage.do_autopart = False
|
|
|
|
# Get a list of all the physical volume devices that make up this VG.
|
|
for pv in self.physvols:
|
|
dev = devicetree.resolve_device(pv)
|
|
if not dev:
|
|
# if pv is using --onpart, use original device
|
|
pv_name = ksdata.onPart.get(pv, pv)
|
|
dev = devicetree.resolve_device(pv_name) or lookupAlias(devicetree, pv)
|
|
if dev and dev.format.type == "luks":
|
|
try:
|
|
dev = dev.children[0]
|
|
except IndexError:
|
|
dev = None
|
|
|
|
if dev and dev.format.type != "lvmpv":
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("Physical volume \"%(device)s\" has a format of \"%(format)s\", but should have a format of \"lvmpv\".") %
|
|
{"device": pv, "format": dev.format.type}))
|
|
|
|
if not dev:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("Tried to use undefined partition \"%s\" in Volume Group specification") % pv))
|
|
|
|
pvs.append(dev)
|
|
|
|
if len(pvs) == 0 and not self.preexist:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("Volume group \"%s\" defined without any physical volumes. Either specify physical volumes or use --useexisting.") % self.vgname))
|
|
|
|
if self.pesize == 0:
|
|
# default PE size requested -- we use blivet's default in KiB
|
|
self.pesize = LVM_PE_SIZE.convert_to(KiB)
|
|
|
|
pesize = Size("%d KiB" % self.pesize)
|
|
possible_extents = LVMVolumeGroupDevice.get_supported_pe_sizes()
|
|
if pesize not in possible_extents:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("Volume group given physical extent size of \"%(extentSize)s\", but must be one of:\n%(validExtentSizes)s.") %
|
|
{"extentSize": pesize, "validExtentSizes": ", ".join(str(e) for e in possible_extents)}))
|
|
|
|
# If --noformat or --useexisting was given, there's really nothing to do.
|
|
if not self.format or self.preexist:
|
|
if not self.vgname:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("volgroup --noformat and volgroup --useexisting must also use the --name= option.")))
|
|
|
|
dev = devicetree.get_device_by_name(self.vgname)
|
|
if not dev:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("Volume group \"%s\" given in volgroup command does not exist.") % self.vgname))
|
|
elif self.vgname in (vg.name for vg in storage.vgs):
|
|
raise KickstartParseError(formatErrorMsg(self.lineno,
|
|
msg=_("The volume group name \"%s\" is already in use.") % self.vgname))
|
|
else:
|
|
try:
|
|
request = storage.new_vg(parents=pvs,
|
|
name=self.vgname,
|
|
pe_size=pesize)
|
|
except (StorageError, ValueError) as e:
|
|
raise KickstartParseError(formatErrorMsg(self.lineno, msg=str(e)))
|
|
|
|
storage.create_device(request)
|
|
if self.reserved_space:
|
|
request.reserved_space = self.reserved_space
|
|
elif self.reserved_percent:
|
|
request.reserved_percent = self.reserved_percent
|
|
|
|
# in case we had to truncate or otherwise adjust the specified name
|
|
ksdata.onPart[self.vgname] = request.name
|
|
|
|
class XConfig(commands.xconfig.F14_XConfig):
|
|
def execute(self, *args):
|
|
desktop = Desktop()
|
|
if self.startX:
|
|
desktop.default_target = GRAPHICAL_TARGET
|
|
|
|
if self.defaultdesktop:
|
|
desktop.desktop = self.defaultdesktop
|
|
|
|
# now write it out
|
|
desktop.write()
|
|
|
|
class SkipX(commands.skipx.FC3_SkipX):
|
|
def execute(self, *args):
|
|
if self.skipx:
|
|
desktop = Desktop()
|
|
desktop.default_target = TEXT_ONLY_TARGET
|
|
desktop.write()
|
|
|
|
class ZFCP(commands.zfcp.F14_ZFCP):
|
|
def parse(self, args):
|
|
fcp = commands.zfcp.F14_ZFCP.parse(self, args)
|
|
try:
|
|
blivet.zfcp.zfcp.add_fcp(fcp.devnum, fcp.wwpn, fcp.fcplun)
|
|
except ValueError as e:
|
|
log.warning(str(e))
|
|
|
|
return fcp
|
|
|
|
class Keyboard(commands.keyboard.F18_Keyboard):
|
|
def execute(self, *args):
|
|
keyboard.write_keyboard_config(self, iutil.getSysroot())
|
|
|
|
class Upgrade(commands.upgrade.F20_Upgrade):
|
|
# Upgrade is no longer supported. If an upgrade command was included in
|
|
# a kickstart, warn the user and exit.
|
|
def parse(self, *args):
|
|
log.error("The upgrade kickstart command is no longer supported. Upgrade functionality is provided through fedup.")
|
|
sys.stderr.write(_("The upgrade kickstart command is no longer supported. Upgrade functionality is provided through fedup."))
|
|
iutil.ipmi_report(IPMI_ABORTED)
|
|
sys.exit(1)
|
|
|
|
###
|
|
### %anaconda Section
|
|
###
|
|
|
|
class AnacondaSectionHandler(BaseHandler):
|
|
"""A handler for only the anaconda ection's commands."""
|
|
commandMap = {
|
|
"pwpolicy": F22_PwPolicy
|
|
}
|
|
|
|
dataMap = {
|
|
"PwPolicyData": F22_PwPolicyData
|
|
}
|
|
|
|
def __init__(self):
|
|
BaseHandler.__init__(self, mapping=self.commandMap, dataMapping=self.dataMap)
|
|
|
|
def __str__(self):
|
|
"""Return the %anaconda section"""
|
|
retval = ""
|
|
# This dictionary should only be modified during __init__, so if it
|
|
# changes during iteration something has gone horribly wrong.
|
|
lst = sorted(self._writeOrder.keys())
|
|
for prio in lst:
|
|
for obj in self._writeOrder[prio]:
|
|
retval += str(obj)
|
|
|
|
if retval:
|
|
retval = "\n%anaconda\n" + retval + "%end\n"
|
|
return retval
|
|
|
|
class AnacondaSection(Section):
|
|
"""A section for anaconda specific commands."""
|
|
sectionOpen = "%anaconda"
|
|
|
|
def __init__(self, *args, **kwargs):
|
|
Section.__init__(self, *args, **kwargs)
|
|
self.cmdno = 0
|
|
|
|
def handleLine(self, line):
|
|
if not self.handler:
|
|
return
|
|
|
|
self.cmdno += 1
|
|
args = shlex.split(line, comments=True)
|
|
self.handler.currentCmd = args[0]
|
|
self.handler.currentLine = self.cmdno
|
|
return self.handler.dispatcher(args, self.cmdno)
|
|
|
|
def handleHeader(self, lineno, args):
|
|
"""Process the arguments to the %anaconda header."""
|
|
Section.handleHeader(self, lineno, args)
|
|
|
|
def finalize(self):
|
|
"""Let %anaconda know no additional data will come."""
|
|
Section.finalize(self)
|
|
|
|
###
|
|
### HANDLERS
|
|
###
|
|
|
|
# This is just the latest entry from pykickstart.handlers.control with all the
|
|
# classes we're overriding in place of the defaults.
|
|
commandMap = {
|
|
"auth": Authconfig,
|
|
"authconfig": Authconfig,
|
|
"autopart": AutoPart,
|
|
"btrfs": BTRFS,
|
|
"bootloader": Bootloader,
|
|
"clearpart": ClearPart,
|
|
"eula": Eula,
|
|
"fcoe": Fcoe,
|
|
"firewall": Firewall,
|
|
"firstboot": Firstboot,
|
|
"group": Group,
|
|
"ignoredisk": IgnoreDisk,
|
|
"iscsi": Iscsi,
|
|
"iscsiname": IscsiName,
|
|
"keyboard": Keyboard,
|
|
"lang": Lang,
|
|
"logging": Logging,
|
|
"logvol": LogVol,
|
|
"network": Network,
|
|
"part": Partition,
|
|
"partition": Partition,
|
|
"raid": Raid,
|
|
"realm": Realm,
|
|
"reqpart": ReqPart,
|
|
"rootpw": RootPw,
|
|
"selinux": SELinux,
|
|
"services": Services,
|
|
"sshkey": SshKey,
|
|
"skipx": SkipX,
|
|
"timezone": Timezone,
|
|
"upgrade": Upgrade,
|
|
"user": User,
|
|
"volgroup": VolGroup,
|
|
"xconfig": XConfig,
|
|
"zfcp": ZFCP,
|
|
}
|
|
|
|
dataMap = {
|
|
"BTRFSData": BTRFSData,
|
|
"LogVolData": LogVolData,
|
|
"PartData": PartitionData,
|
|
"RaidData": RaidData,
|
|
"RepoData": RepoData,
|
|
"VolGroupData": VolGroupData,
|
|
}
|
|
|
|
superclass = returnClassForVersion()
|
|
|
|
class AnacondaKSHandler(superclass):
|
|
AddonClassType = AddonData
|
|
|
|
def __init__(self, addon_paths=None, commandUpdates=None, dataUpdates=None):
|
|
if addon_paths is None:
|
|
addon_paths = []
|
|
|
|
if commandUpdates is None:
|
|
commandUpdates = commandMap
|
|
|
|
if dataUpdates is None:
|
|
dataUpdates = dataMap
|
|
|
|
superclass.__init__(self, commandUpdates=commandUpdates, dataUpdates=dataUpdates)
|
|
self.onPart = {}
|
|
|
|
# collect all kickstart addons for anaconda to addons dictionary
|
|
# which maps addon_id to it's own data structure based on BaseData
|
|
# with execute method
|
|
addons = {}
|
|
|
|
# collect all AddonData subclasses from
|
|
# for p in addon_paths: <p>/<plugin id>/ks/*.(py|so)
|
|
# and register them under <plugin id> name
|
|
for module_name, path in addon_paths:
|
|
addon_id = os.path.basename(os.path.dirname(os.path.abspath(path)))
|
|
if not os.path.isdir(path):
|
|
continue
|
|
|
|
classes = collect(module_name, path, lambda cls: issubclass(cls, self.AddonClassType))
|
|
if classes:
|
|
addons[addon_id] = classes[0](name=addon_id)
|
|
|
|
# Prepare the final structures for 3rd party addons
|
|
self.addons = AddonRegistry(addons)
|
|
|
|
# The %anaconda section uses its own handler for a limited set of commands
|
|
self.anaconda = AnacondaSectionHandler()
|
|
|
|
def __str__(self):
|
|
return superclass.__str__(self) + "\n" + str(self.addons) + str(self.anaconda)
|
|
|
|
class AnacondaPreParser(KickstartParser):
|
|
# A subclass of KickstartParser that only looks for %pre scripts and
|
|
# sets them up to be run. All other scripts and commands are ignored.
|
|
def __init__(self, handler, followIncludes=True, errorsAreFatal=True,
|
|
missingIncludeIsFatal=True):
|
|
KickstartParser.__init__(self, handler, missingIncludeIsFatal=False)
|
|
|
|
def handleCommand(self, lineno, args):
|
|
pass
|
|
|
|
def setupSections(self):
|
|
self.registerSection(PreScriptSection(self.handler, dataObj=AnacondaKSScript))
|
|
self.registerSection(NullSection(self.handler, sectionOpen="%pre-install"))
|
|
self.registerSection(NullSection(self.handler, sectionOpen="%post"))
|
|
self.registerSection(NullSection(self.handler, sectionOpen="%onerror"))
|
|
self.registerSection(NullSection(self.handler, sectionOpen="%traceback"))
|
|
self.registerSection(NullSection(self.handler, sectionOpen="%packages"))
|
|
self.registerSection(NullSection(self.handler, sectionOpen="%addon"))
|
|
self.registerSection(NullSection(self.handler.anaconda, sectionOpen="%anaconda"))
|
|
|
|
|
|
class AnacondaKSParser(KickstartParser):
|
|
def __init__(self, handler, followIncludes=True, errorsAreFatal=True,
|
|
missingIncludeIsFatal=True, scriptClass=AnacondaKSScript):
|
|
self.scriptClass = scriptClass
|
|
KickstartParser.__init__(self, handler)
|
|
|
|
def handleCommand(self, lineno, args):
|
|
if not self.handler:
|
|
return
|
|
|
|
return KickstartParser.handleCommand(self, lineno, args)
|
|
|
|
def setupSections(self):
|
|
self.registerSection(PreScriptSection(self.handler, dataObj=self.scriptClass))
|
|
self.registerSection(PreInstallScriptSection(self.handler, dataObj=self.scriptClass))
|
|
self.registerSection(PostScriptSection(self.handler, dataObj=self.scriptClass))
|
|
self.registerSection(TracebackScriptSection(self.handler, dataObj=self.scriptClass))
|
|
self.registerSection(OnErrorScriptSection(self.handler, dataObj=self.scriptClass))
|
|
self.registerSection(PackageSection(self.handler))
|
|
self.registerSection(AddonSection(self.handler))
|
|
self.registerSection(AnacondaSection(self.handler.anaconda))
|
|
|
|
def preScriptPass(f):
|
|
# The first pass through kickstart file processing - look for %pre scripts
|
|
# and run them. This must come in a separate pass in case a script
|
|
# generates an included file that has commands for later.
|
|
ksparser = AnacondaPreParser(AnacondaKSHandler())
|
|
|
|
try:
|
|
ksparser.readKickstart(f)
|
|
except KickstartError as e:
|
|
# We do not have an interface here yet, so we cannot use our error
|
|
# handling callback.
|
|
print(e)
|
|
iutil.ipmi_report(IPMI_ABORTED)
|
|
sys.exit(1)
|
|
|
|
# run %pre scripts
|
|
runPreScripts(ksparser.handler.scripts)
|
|
|
|
def parseKickstart(f):
|
|
# preprocessing the kickstart file has already been handled in initramfs.
|
|
|
|
addon_paths = collect_addon_paths(ADDON_PATHS)
|
|
handler = AnacondaKSHandler(addon_paths["ks"])
|
|
ksparser = AnacondaKSParser(handler)
|
|
|
|
# We need this so all the /dev/disk/* stuff is set up before parsing.
|
|
udev.trigger(subsystem="block", action="change")
|
|
|
|
try:
|
|
ksparser.readKickstart(f)
|
|
except KickstartError as e:
|
|
# We do not have an interface here yet, so we cannot use our error
|
|
# handling callback.
|
|
print(e)
|
|
iutil.ipmi_report(IPMI_ABORTED)
|
|
sys.exit(1)
|
|
|
|
return handler
|
|
|
|
def appendPostScripts(ksdata):
|
|
scripts = ""
|
|
|
|
# Read in all the post script snippets to a single big string.
|
|
for fn in glob.glob("/usr/share/anaconda/post-scripts/*ks"):
|
|
f = open(fn, "r")
|
|
scripts += f.read()
|
|
f.close()
|
|
|
|
# Then parse the snippets against the existing ksdata. We can do this
|
|
# because pykickstart allows multiple parses to save their data into a
|
|
# single data object. Errors parsing the scripts are a bug in anaconda,
|
|
# so just raise an exception.
|
|
ksparser = AnacondaKSParser(ksdata, scriptClass=AnacondaInternalScript)
|
|
ksparser.readKickstartFromString(scripts, reset=False)
|
|
|
|
def runPostScripts(scripts):
|
|
postScripts = [s for s in scripts if s.type == KS_SCRIPT_POST]
|
|
|
|
if len(postScripts) == 0:
|
|
return
|
|
|
|
log.info("Running kickstart %%post script(s)")
|
|
for script in postScripts:
|
|
script.run(iutil.getSysroot())
|
|
log.info("All kickstart %%post script(s) have been run")
|
|
|
|
def runPreScripts(scripts):
|
|
preScripts = [s for s in scripts if s.type == KS_SCRIPT_PRE]
|
|
|
|
if len(preScripts) == 0:
|
|
return
|
|
|
|
log.info("Running kickstart %%pre script(s)")
|
|
stdoutLog.info(_("Running pre-installation scripts"))
|
|
|
|
for script in preScripts:
|
|
script.run("/")
|
|
|
|
log.info("All kickstart %%pre script(s) have been run")
|
|
|
|
def runPreInstallScripts(scripts):
|
|
preInstallScripts = [s for s in scripts if s.type == KS_SCRIPT_PREINSTALL]
|
|
|
|
if len(preInstallScripts) == 0:
|
|
return
|
|
|
|
log.info("Running kickstart %%pre-install script(s)")
|
|
|
|
for script in preInstallScripts:
|
|
script.run("/")
|
|
|
|
log.info("All kickstart %%pre-install script(s) have been run")
|
|
|
|
def runTracebackScripts(scripts):
|
|
log.info("Running kickstart %%traceback script(s)")
|
|
for script in filter(lambda s: s.type == KS_SCRIPT_TRACEBACK, scripts):
|
|
script.run("/")
|
|
log.info("All kickstart %%traceback script(s) have been run")
|
|
|
|
def resetCustomStorageData(ksdata):
|
|
for command in ["partition", "raid", "volgroup", "logvol", "btrfs"]:
|
|
ksdata.resetCommand(command)
|
|
|
|
ksdata.clearpart.type = CLEARPART_TYPE_NONE
|
|
|
|
def doKickstartStorage(storage, ksdata, instClass):
|
|
""" Setup storage state from the kickstart data """
|
|
ksdata.clearpart.execute(storage, ksdata, instClass)
|
|
if not any(d for d in storage.disks
|
|
if not d.format.hidden and not d.protected):
|
|
return
|
|
|
|
# snapshot free space now so that we know how much we had available
|
|
storage.create_free_space_snapshot()
|
|
|
|
ksdata.bootloader.execute(storage, ksdata, instClass, dry_run=True)
|
|
ksdata.autopart.execute(storage, ksdata, instClass)
|
|
ksdata.reqpart.execute(storage, ksdata, instClass)
|
|
ksdata.partition.execute(storage, ksdata, instClass)
|
|
ksdata.raid.execute(storage, ksdata, instClass)
|
|
ksdata.volgroup.execute(storage, ksdata, instClass)
|
|
ksdata.logvol.execute(storage, ksdata, instClass)
|
|
ksdata.btrfs.execute(storage, ksdata, instClass)
|
|
# also calls ksdata.bootloader.execute
|
|
storage.set_up_bootloader()
|