+$endif$
+$-------------------------------------------------------------------------------
+
+$endif$
+$--<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<< END IF: title
+$--------------------------------------------------------------------------> toc
+$if(toc)$
+
+
+
+$endif$
+$-------------------------------------------------------------------------> body
+$body$
+$----------------------------------------------------------------> include-after
+$for(include-after)$
+$include-after$
+$endfor$
+$-------------------------------------------------------------------------------
+
+
+
diff --git a/CHANGELOG.md b/CHANGELOG.md
index 370ad78e..9a2df55f 100644
--- a/CHANGELOG.md
+++ b/CHANGELOG.md
@@ -1,6 +1,19 @@
Changelog
=========
+[1.12.2] - 2023-07-19
+--------------------
+
+### Bug Fixes
+
+- fix: RAID volume pre cleanup (#169)
+- fix: facts being gathered unnecessarily (#374)
+
+### Other Changes
+
+- ci: ansible-lint - ignore var-naming[no-role-prefix] (#371)
+- ci: ansible-test ignores file for ansible-core 2.15 (#372)
+
[1.12.1] - 2023-07-08
--------------------
diff --git a/README.md b/README.md
index a8e16759..e07dc3bf 100644
--- a/README.md
+++ b/README.md
@@ -5,6 +5,7 @@
This role allows users to configure local storage with minimal input.
As of now, the role supports managing file systems and mount entries on
+
- unpartitioned disks
- lvm (unpartitioned whole-disk physical volumes only)
@@ -17,7 +18,7 @@ See below
The role requires the `mount` module from `ansible.posix`. If you are using
`ansible-core`, you must install the `ansible.posix` collection.
-```
+```bash
ansible-galaxy collection install -vv -r meta/collection-requirements.yml
```
@@ -34,243 +35,294 @@ parameters will inherit whatever setting the pool or volume already has.
This means that to change/override role defaults in an existing pool or volume,
you must explicitly specify the new values/settings in the role variables.
-#### `storage_pools`
+### `storage_pools`
+
The `storage_pools` variable is a list of pools to manage. Each pool contains a
nested list of `volume` dicts as described below, as well as the following
keys:
-##### `name`
-This specifies the name of the pool to manage/create as a string. (One
-example of a pool is an LVM volume group.)
+- `name`
+
+ This specifies the name of the pool to manage/create as a string. (One
+ example of a pool is an LVM volume group.)
+
+- `type`
+
+ This specifies the type of pool to manage.
+ Valid values for `type`: `lvm`.
+
+- `disks`
+
+ A list which specifies the set of disks to use as backing storage for the pool.
+ Supported identifiers include: device node (like `/dev/sda` or `/dev/mapper/mpathb`),
+ device node basename (like `sda` or `mpathb`), /dev/disk/ symlink
+ (like `/dev/disk/by-id/wwn-0x5000c5005bc37f3f`).
+
+ For LVM pools this can be also used to add and remove disks to/from an existing pool.
+ Disks in the list that are not used by the pool will be added to the pool.
+ Disks that are currently used by the pool but not present in the list will be removed
+ from the pool only if `storage_safe_mode` is set to `false`.
+
+- `raid_level`
-##### `type`
-This specifies the type of pool to manage.
-Valid values for `type`: `lvm`.
+ When used with `type: lvm` it manages a volume group with a mdraid array of given level
+ on it. Input `disks` are in this case used as RAID members.
+ Accepted values are: `linear`, `raid0`, `raid1`, `raid4`, `raid5`, `raid6`, `raid10`
-##### `disks`
-A list which specifies the set of disks to use as backing storage for the pool.
-Supported identifiers include: device node (like `/dev/sda` or `/dev/mapper/mpathb`),
-device node basename (like `sda` or `mpathb`), /dev/disk/ symlink
-(like `/dev/disk/by-id/wwn-0x5000c5005bc37f3f`).
+- `volumes`
-For LVM pools this can be also used to add and remove disks to/from an existing pool.
-Disks in the list that are not used by the pool will be added to the pool.
-Disks that are currently used by the pool but not present in the list will be removed
-from the pool only if `storage_safe_mode` is set to `false`.
+ This is a list of volumes that belong to the current pool. It follows the
+ same pattern as the `storage_volumes` variable, explained below.
-##### `raid_level`
-When used with `type: lvm` it manages a volume group with a mdraid array of given level
-on it. Input `disks` are in this case used as RAID members.
-Accepted values are: `linear`, `raid0`, `raid1`, `raid4`, `raid5`, `raid6`, `raid10`
+- `encryption`
-##### `volumes`
-This is a list of volumes that belong to the current pool. It follows the
-same pattern as the `storage_volumes` variable, explained below.
+ This specifies whether the pool will be encrypted using LUKS.
+ __WARNING__: Toggling encryption for a pool is a destructive operation, meaning
+ the pool itself will be removed as part of the process of
+ adding/removing the encryption layer.
-##### `encryption`
-This specifies whether the pool will be encrypted using LUKS.
-__WARNING__: Toggling encryption for a pool is a destructive operation, meaning
- the pool itself will be removed as part of the process of
- adding/removing the encryption layer.
+- `encryption_password`
-##### `encryption_password`
-This string specifies a password or passphrase used to unlock/open the LUKS volume(s).
+ This string specifies a password or passphrase used to unlock/open the LUKS volume(s).
-##### `encryption_key`
-This string specifies the full path to the key file on the managed nodes used to unlock
-the LUKS volume(s). It is the responsibility of the user of this role to securely copy
-this file to the managed nodes, or otherwise ensure that the file is on the managed
-nodes.
+- `encryption_key`
-##### `encryption_cipher`
-This string specifies a non-default cipher to be used by LUKS.
+ This string specifies the full path to the key file on the managed nodes used to unlock
+ the LUKS volume(s). It is the responsibility of the user of this role to securely copy
+ this file to the managed nodes, or otherwise ensure that the file is on the managed
+ nodes.
-##### `encryption_key_size`
-This integer specifies the LUKS key size (in bytes).
+- `encryption_cipher`
-##### `encryption_luks_version`
-This integer specifies the LUKS version to use.
+ ifies a non-default cipher to be used by LUKS.
+- `encryption_key_size`
+
+ s the LUKS key size (in bytes).
+
+- `encryption_luks_version`
+
+ This integer specifies the LUKS version to use.
+
+### `storage_volumes`
-#### `storage_volumes`
The `storage_volumes` variable is a list of volumes to manage. Each volume has the following
variables:
-##### `name`
-This specifies the name of the volume.
+- `name`
-##### `type`
-This specifies the type of volume on which the file system will reside.
-Valid values for `type`: `lvm`, `disk` or `raid`.
-The default is determined according to the OS and release (currently `lvm`).
+ This specifies the name of the volume.
-##### `disks`
-This specifies the set of disks to use as backing storage for the file system.
-This is currently only relevant for volumes of type `disk`, where the list
-must contain only a single item.
+- `type`
-##### `size`
-The `size` specifies the size of the file system. The format for this is intended to
-be human-readable, e.g.: "10g", "50 GiB". The size of LVM volumes can be specified as a
-percentage of the pool/VG size, eg: "50%" as of v1.4.2.
+ This specifies the type of volume on which the file system will reside.
+ Valid values for `type`: `lvm`, `disk` or `raid`.
+ The default is determined according to the OS and release (currently `lvm`).
-When using `compression` or `deduplication`, `size` can be set higher than actual available space,
-e.g.: 3 times the size of the volume, based on duplicity and/or compressibility of stored data.
+- `disks`
-__NOTE__: The requested volume size may be reduced as necessary so the volume can
- fit in the available pool space, but only if the required reduction is
- not more than 2% of the requested volume size.
-
-##### `fs_type`
-This indicates the desired file system type to use, e.g.: "xfs", "ext4", "swap".
-The default is determined according to the OS and release
-(currently `xfs` for all the supported systems).
-
-##### `fs_label`
-The `fs_label` is a string to be used for a file system label.
-
-##### `fs_create_options`
-The `fs_create_options` specifies custom arguments to `mkfs` as a string.
-
-##### `mount_point`
-The `mount_point` specifies the directory on which the file system will be mounted.
-
-##### `mount_options`
-The `mount_options` specifies custom mount options as a string, e.g.: 'ro'.
-
-##### `mount_user`
-The `mount_user` specifies desired owner of the mount directory.
-
-##### `mount_group`
-The `mount_group` specifies desired group of the mount directory.
-
-##### `mount_mode`
-The `mount_mode` specifies desired permissions of the mount directory.
-
-##### `raid_level`
-Specifies RAID level. LVM RAID can be created as well.
-"Regular" RAID volume requires type to be `raid`.
-LVM RAID needs that volume has `storage_pools` parent with type `lvm`,
-`raid_disks` need to be specified as well.
-Accepted values are:
-* for LVM RAID volume: `raid0`, `raid1`, `raid4`, `raid5`, `raid6`, `raid10`, `striped`, `mirror`
-* for RAID volume: `linear`, `raid0`, `raid1`, `raid4`, `raid5`, `raid6`, `raid10`
-
-__WARNING__: Changing `raid_level` for a volume is a destructive operation, meaning
- all data on that volume will be lost as part of the process of
- removing old and adding new RAID. RAID reshaping is currently not
- supported.
-
-##### `raid_device_count`
-When type is `raid` specifies number of active RAID devices.
-
-##### `raid_spare_count`
-When type is `raid` specifies number of spare RAID devices.
-
-##### `raid_metadata_version`
-When type is `raid` specifies RAID metadata version as a string, e.g.: '1.0'.
-
-##### `raid_chunk_size`
-When type is `raid` specifies RAID chunk size as a string, e.g.: '512 KiB'.
-Chunk size has to be multiple of 4 KiB.
-
-##### `raid_stripe_size`
-When type is `lvm` specifies LVM RAID stripe size as a string, e.g.: '512 KiB'.
-
-##### `raid_disks`
-Specifies which disks should be used for LVM RAID volume.
-`raid_level` needs to be specified and volume has to have `storage_pools` parent with type `lvm`.
-Accepts sublist of `disks` of parent `storage_pools`.
-In case multiple LVM RAID volumes within the same storage pool, the same disk can be used
-in multiple `raid_disks`.
-
-##### `encryption`
-This specifies whether the volume will be encrypted using LUKS.
-__WARNING__: Toggling encryption for a volume is a destructive operation, meaning
- all data on that volume will be removed as part of the process of
- adding/removing the encryption layer.
-
-##### `encryption_password`
-This string specifies a password or passphrase used to unlock/open the LUKS volume.
-
-##### `encryption_key`
-This string specifies the full path to the key file on the managed nodes used to unlock
-the LUKS volume(s). It is the responsibility of the user of this role to securely copy
-this file to the managed nodes, or otherwise ensure that the file is on the managed
-nodes.
-
-##### `encryption_cipher`
-This string specifies a non-default cipher to be used by LUKS.
-
-##### `encryption_key_size`
-This integer specifies the LUKS key size (in bits).
-
-##### `encryption_luks_version`
-This integer specifies the LUKS version to use.
-
-##### `deduplication`
-This specifies whether the Virtual Data Optimizer (VDO) will be used.
-When set, duplicate data stored on storage volume will be
-deduplicated resulting in more storage capacity.
-Can be used together with `compression` and `vdo_pool_size`.
-Volume has to be part of the LVM `storage_pool`.
-Limit one VDO `storage_volume` per `storage_pool`.
-Underlying volume has to be at least 9 GB (bare minimum is around 5 GiB).
-
-##### `compression`
-This specifies whether the Virtual Data Optimizer (VDO) will be used.
-When set, data stored on storage volume will be
-compressed resulting in more storage capacity.
-Volume has to be part of the LVM `storage_pool`.
-Can be used together with `deduplication` and `vdo_pool_size`.
-Limit one VDO `storage_volume` per `storage_pool`.
-
-##### `vdo_pool_size`
-When Virtual Data Optimizer (VDO) is used, this specifies the actual size the volume
-will take on the device. Virtual size of VDO volume is set by `size` parameter.
-`vdo_pool_size` format is intended to be human-readable,
-e.g.: "30g", "50GiB".
-Default value is equal to the size of the volume.
+ This specifies the set of disks to use as backing storage for the file system.
+ This is currently only relevant for volumes of type `disk`, where the list
+ must contain only a single item.
+
+- `size`
+
+ The `size` specifies the size of the file system. The format for this is intended to
+ be human-readable, e.g.: "10g", "50 GiB". The size of LVM volumes can be specified as a
+ percentage of the pool/VG size, eg: "50%" as of v1.4.2.
+
+ When using `compression` or `deduplication`, `size` can be set higher than actual available space,
+ e.g.: 3 times the size of the volume, based on duplicity and/or compressibility of stored data.
+
+ __NOTE__: The requested volume size may be reduced as necessary so the volume can
+ fit in the available pool space, but only if the required reduction is
+ not more than 2% of the requested volume size.
+
+- `fs_type`
+
+ This indicates the desired file system type to use, e.g.: "xfs", "ext4", "swap".
+ The default is determined according to the OS and release
+ (currently `xfs` for all the supported systems).
+
+- `fs_label`
+
+ The `fs_label` is a string to be used for a file system label.
+
+- `fs_create_options`
+
+ The `fs_create_options` specifies custom arguments to `mkfs` as a string.
+
+- `mount_point`
+
+ The `mount_point` specifies the directory on which the file system will be mounted.
+
+- `mount_options`
+
+ The `mount_options` specifies custom mount options as a string, e.g.: 'ro'.
+
+- `mount_user`
+
+ The `mount_user` specifies desired owner of the mount directory.
+
+- `mount_group`
+
+ The `mount_group` specifies desired group of the mount directory.
+
+- `mount_mode`
+
+ The `mount_mode` specifies desired permissions of the mount directory.
+
+- `raid_level`
+
+ Specifies RAID level. LVM RAID can be created as well.
+ "Regular" RAID volume requires type to be `raid`.
+ LVM RAID needs that volume has `storage_pools` parent with type `lvm`,
+ `raid_disks` need to be specified as well.
+ Accepted values are:
+
+ - for LVM RAID volume: `raid0`, `raid1`, `raid4`, `raid5`, `raid6`, `raid10`, `striped`, `mirror`
+ - for RAID volume: `linear`, `raid0`, `raid1`, `raid4`, `raid5`, `raid6`, `raid10`
+
+ __WARNING__: Changing `raid_level` for a volume is a destructive operation, meaning
+ all data on that volume will be lost as part of the process of
+ removing old and adding new RAID. RAID reshaping is currently not
+ supported.
+
+- `raid_device_count`
+
+ When type is `raid` specifies number of active RAID devices.
+
+- `raid_spare_count`
+
+ When type is `raid` specifies number of spare RAID devices.
+
+- `raid_metadata_version`
+
+ When type is `raid` specifies RAID metadata version as a string, e.g.: '1.0'.
+
+- `raid_chunk_size`
+
+ When type is `raid` specifies RAID chunk size as a string, e.g.: '512 KiB'.
+ Chunk size has to be multiple of 4 KiB.
+
+- `raid_stripe_size`
+
+ When type is `lvm` specifies LVM RAID stripe size as a string, e.g.: '512 KiB'.
+
+- `raid_disks`
+
+ Specifies which disks should be used for LVM RAID volume.
+ `raid_level` needs to be specified and volume has to have `storage_pools` parent with type `lvm`.
+ Accepts sublist of `disks` of parent `storage_pools`.
+ In case multiple LVM RAID volumes within the same storage pool, the same disk can be used
+ in multiple `raid_disks`.
+
+- `encryption`
+
+ This specifies whether the volume will be encrypted using LUKS.
+ __WARNING__: Toggling encryption for a volume is a destructive operation, meaning
+ all data on that volume will be removed as part of the process of
+ adding/removing the encryption layer.
+
+- `encryption_password`
+
+ This string specifies a password or passphrase used to unlock/open the LUKS volume.
+
+- `encryption_key`
+
+ This string specifies the full path to the key file on the managed nodes used to unlock
+ the LUKS volume(s). It is the responsibility of the user of this role to securely copy
+ this file to the managed nodes, or otherwise ensure that the file is on the managed
+ nodes.
+
+- `encryption_cipher`
+
+ This string specifies a non-default cipher to be used by LUKS.
+
+- `encryption_key_size`
+
+ This integer specifies the LUKS key size (in bits).
+
+- `encryption_luks_version`
+
+ This integer specifies the LUKS version to use.
+
+- `deduplication`
+
+ This specifies whether the Virtual Data Optimizer (VDO) will be used.
+ When set, duplicate data stored on storage volume will be
+ deduplicated resulting in more storage capacity.
+ Can be used together with `compression` and `vdo_pool_size`.
+ Volume has to be part of the LVM `storage_pool`.
+ Limit one VDO `storage_volume` per `storage_pool`.
+ Underlying volume has to be at least 9 GB (bare minimum is around 5 GiB).
+
+- `compression`
+
+ This specifies whether the Virtual Data Optimizer (VDO) will be used.
+ When set, data stored on storage volume will be
+ compressed resulting in more storage capacity.
+ Volume has to be part of the LVM `storage_pool`.
+ Can be used together with `deduplication` and `vdo_pool_size`.
+ Limit one VDO `storage_volume` per `storage_pool`.
+
+- `vdo_pool_size`
+
+ When Virtual Data Optimizer (VDO) is used, this specifies the actual size the volume
+ will take on the device. Virtual size of VDO volume is set by `size` parameter.
+ `vdo_pool_size` format is intended to be human-readable,
+ e.g.: "30g", "50GiB".
+ Default value is equal to the size of the volume.
+
+### `cached`
-#### `cached`
This specifies whether the volume should be cached or not.
This is currently supported only for LVM volumes where dm-cache
is used.
-#### `cache_size`
+### `cache_size`
+
Size of the cache. `cache_size` format is intended to be human-readable,
e.g.: "30g", "50GiB".
-#### `cache_mode`
+### `cache_mode`
+
Mode for the cache. Supported values include `writethrough` (default) and `writeback`.
-#### `cache_devices`
+### `cache_devices`
+
List of devices that will be used for the cache. These should be either physical volumes or
drives these physical volumes are allocated on. Generally you want to select fast devices like
SSD or NVMe drives for cache.
-#### `thin`
+### `thin`
+
Whether the volume should be thinly provisioned or not.
This is supported only for LVM volumes.
-#### `thin_pool_name`
+### `thin_pool_name`
+
For `thin` volumes, this can be used to specify the name of the LVM thin pool that will be used
for the volume. If the pool with the provided name already exists, the volume will be added to that
pool. If it doesn't exist a new pool named `thin_pool_name` will be created.
If not specified:
-* if there are no existing thin pools present, a new thin pool will be created with an automatically
- generated name,
-* if there is exactly one existing thin pool, the thin volume will be added to it and
-* if there are multiple thin pools present an exception will be raised.
-#### `thin_pool_size`
+- if there are no existing thin pools present, a new thin pool will be created with an automatically
+ generated name,
+- if there is exactly one existing thin pool, the thin volume will be added to it and
+- if there are multiple thin pools present an exception will be raised.
+
+### `thin_pool_size`
+
Size for the thin pool. `thin_pool_size` format is intended to be human-readable,
e.g.: "30g", "50GiB".
-#### `storage_safe_mode`
+### `storage_safe_mode`
+
When true (the default), an error will occur instead of automatically removing existing devices and/or formatting.
-#### `storage_udevadm_trigger`
+### `storage_udevadm_trigger`
+
When true (the default is false), the role will use udevadm trigger
to cause udev changes to take effect immediately. This may help on some
platforms with "buggy" udev.