Skip to content

storage1.staging: Adapt to be a rancher node

Antoine R. Dumont requested to merge adapt-storage1-as-rancher-node into production
make swh-helm-diff
*** Running octocatalog-diff on host storage1.internal.staging.swh.network
I, [2023-11-22T15:08:15.139548 #3524849]  INFO -- : Catalogs compiled for storage1.internal.staging.swh.network
I, [2023-11-22T15:08:15.488983 #3524849]  INFO -- : Diffs computed for storage1.internal.staging.swh.network
diff origin/production/storage1.internal.staging.swh.network current/storage1.internal.staging.swh.network
*******************************************
+ Exec[create /tmp] =>
   parameters =>
     "command": "mkdir -p /tmp",
     "creates": "/tmp",
     "path": [
       "/bin",
       "/usr/bin",
       "/sbin",
       "/usr/sbin"
     ]
*******************************************
+ Exec[enforce-sysctl-value-vm.max_map_count] =>
   parameters =>
     "command": "/sbin/sysctl -w vm.max_map_count=2097152",
     "unless": "/usr/bin/test \"$(/sbin/sysctl -n vm.max_map_count)\" = 2097152"
*******************************************
+ Exec[sysctl-vm.max_map_count] =>
   parameters =>
     "command": "sysctl -p /etc/sysctl.d/vm.max_map_count.conf",
     "path": [
       "/usr/sbin",
       "/sbin",
       "/usr/bin",
       "/bin"
     ],
     "refreshonly": true
*******************************************
+ Exec[update-sysctl.conf-vm.max_map_count] =>
   parameters =>
     "command": "sed -i -e 's#^vm.max_map_count *=.*#vm.max_map_count = 2097152#'...
     "onlyif": "grep -E '^vm.max_map_count *=' /etc/sysctl.conf",
     "path": [
       "/usr/sbin",
       "/sbin",
       "/usr/bin",
       "/bin"
     ],
     "refreshonly": true
*******************************************
+ File[/etc/rancher/rke2/config.yaml.d/50-snapshotter.yaml] =>
   parameters =>
     "content": "# File managed by puppet - modifications will be lost\nsnapshott...
     "group": "root",
     "mode": "0644",
     "owner": "root"
*******************************************
+ File[/etc/rancher/rke2/config.yaml.d] =>
   parameters =>
     "ensure": "directory",
     "group": "root",
     "mode": "0755",
     "owner": "root"
*******************************************
+ File[/etc/rancher/rke2] =>
   parameters =>
     "ensure": "directory",
     "group": "root",
     "mode": "0755",
     "owner": "root"
*******************************************
+ File[/etc/rancher] =>
   parameters =>
     "ensure": "directory",
     "group": "root",
     "mode": "0755",
     "owner": "root"
*******************************************
+ File[/etc/sysctl.d/99-sysctl.conf] =>
   parameters =>
     "ensure": "link",
     "group": "root",
     "owner": "root",
     "target": "../sysctl.conf"
*******************************************
+ File[/etc/sysctl.d/vm.max_map_count.conf] =>
   parameters =>
     "content": "vm.max_map_count = 2097152\n",
     "group": "root",
     "mode": "0644",
     "notify": [
       "Exec[sysctl-vm.max_map_count]",
       "Exec[update-sysctl.conf-vm.max_map_count]"
     ],
     "owner": "root"
*******************************************
+ File[/etc/sysctl.d] =>
   parameters =>
     "ensure": "directory",
     "group": "root",
     "mode": "0755",
     "owner": "root",
     "purge": false,
     "recurse": false
*******************************************
+ File[/tmp]
*******************************************
+ Mount[/tmp] =>
   parameters =>
     "device": "tmpfs",
     "dump": 0,
     "ensure": "present",
     "fstype": "tmpfs",
     "options": "size=15997071360,nr_inodes=200m,noexec,nosuid,nodev,relatime,rw"...
     "pass": 0
*******************************************
+ Mount[swap] =>
   parameters =>
     "device": "LABEL=\"swap\"",
     "dump": 0,
     "ensure": "present",
     "fstype": "swap",
     "options": "sw",
     "pass": 0
*******************************************
+ Sysctl[vm.max_map_count] =>
   parameters =>
     "enforce": true,
     "suffix": ".conf",
     "value": "2097152"
*******************************************
+ Zfs[data/kubelet] =>
   parameters =>
     "atime": "off",
     "compression": "zstd",
     "ensure": "present",
     "mountpoint": "/var/lib/kubelet"
*******************************************
+ Zfs[data/rancher] =>
   parameters =>
     "atime": "off",
     "compression": "zstd",
     "ensure": "present",
     "mountpoint": "/var/lib/rancher"
*******************************************
+ Zfs[data/volumes] =>
   parameters =>
     "atime": "off",
     "compression": "zstd",
     "ensure": "present",
     "mountpoint": "/srv/kubernetes/volumes"
*******************************************
+ Zpool[data] =>
   parameters =>
     "disk": "vdb",
     "ensure": "present"
*******************************************
*** End octocatalog-diff on storage1.internal.staging.swh.network

Refs. swh/infra/sysadm-environment#4780 (closed)

Refs. swh/infra/sysadm-environment#5164 (closed)

Edited by Antoine R. Dumont

Merge request reports

Loading