Signed-off-by: Friedrich Weber <f.we...@proxmox.com>
---

Notes:
    new in v3

 PVE/CLI/Makefile   |    1 +
 PVE/CLI/pve8to9.pm | 1543 ++++++++++++++++++++++++++++++++++++++++++++
 bin/Makefile       |   12 +-
 bin/pve8to9        |    8 +
 4 files changed, 1563 insertions(+), 1 deletion(-)
 create mode 100644 PVE/CLI/pve8to9.pm
 create mode 100755 bin/pve8to9

diff --git a/PVE/CLI/Makefile b/PVE/CLI/Makefile
index 04b60493..9ff2aeb5 100644
--- a/PVE/CLI/Makefile
+++ b/PVE/CLI/Makefile
@@ -9,6 +9,7 @@ SOURCES = \
        pvenode.pm \
        pvesh.pm \
        pve7to8.pm \
+       pve8to9.pm \
 
 all:
 
diff --git a/PVE/CLI/pve8to9.pm b/PVE/CLI/pve8to9.pm
new file mode 100644
index 00000000..42b7073b
--- /dev/null
+++ b/PVE/CLI/pve8to9.pm
@@ -0,0 +1,1543 @@
+package PVE::CLI::pve8to9;
+
+use strict;
+use warnings;
+
+use Cwd ();
+
+use PVE::API2::APT;
+use PVE::API2::Ceph;
+use PVE::API2::LXC;
+use PVE::API2::Qemu;
+use PVE::API2::Certificates;
+use PVE::API2::Cluster::Ceph;
+
+use PVE::AccessControl;
+use PVE::Ceph::Tools;
+use PVE::Cluster;
+use PVE::Corosync;
+use PVE::INotify;
+use PVE::JSONSchema;
+use PVE::NodeConfig;
+use PVE::RPCEnvironment;
+use PVE::Storage;
+use PVE::Storage::Plugin;
+use PVE::Tools qw(run_command split_list file_get_contents);
+use PVE::QemuConfig;
+use PVE::QemuServer;
+use PVE::VZDump::Common;
+use PVE::LXC;
+use PVE::LXC::Config;
+use PVE::LXC::Setup;
+
+use Term::ANSIColor;
+
+use PVE::CLIHandler;
+
+use base qw(PVE::CLIHandler);
+
+my $nodename = PVE::INotify::nodename();
+
+my $upgraded = 0; # set in check_pve_packages
+
+sub setup_environment {
+    PVE::RPCEnvironment->setup_default_cli_env();
+}
+
+my $new_suite = 'bookworm';
+my $old_suite = 'bullseye';
+my $older_suites = {
+    buster => 1,
+    stretch => 1,
+    jessie => 1,
+};
+
+my ($min_pve_major, $min_pve_minor, $min_pve_pkgrel) = (7, 4, 1);
+
+my $ceph_release2code = {
+    '12' => 'Luminous',
+    '13' => 'Mimic',
+    '14' => 'Nautilus',
+    '15' => 'Octopus',
+    '16' => 'Pacific',
+    '17' => 'Quincy',
+    '18' => 'Reef',
+};
+my $ceph_supported_release = 17; # the version we support for upgrading (i.e., 
available on both)
+my $ceph_supported_code_name = $ceph_release2code->{"$ceph_supported_release"}
+    or die "inconsistent source code, could not map expected ceph version to 
code name!";
+
+my $forced_legacy_cgroup = 0;
+
+my $counters = {
+    pass => 0,
+    skip => 0,
+    notice => 0,
+    warn => 0,
+    fail => 0,
+};
+
+my $level2color = {
+    pass => 'green',
+    notice => 'bold',
+    warn => 'yellow',
+    fail => 'bold red',
+};
+
+my $log_line = sub {
+    my ($level, $line) = @_;
+
+    $counters->{$level}++ if defined($level) && defined($counters->{$level});
+
+    my $color = $level2color->{$level} // '';
+    print color($color) if $color && $color ne '';
+
+    print uc($level), ': ' if defined($level);
+    print "$line\n";
+
+    print color('reset');
+};
+
+sub log_pass { $log_line->('pass', @_); }
+sub log_info { $log_line->('info', @_); }
+sub log_skip { $log_line->('skip', @_); }
+sub log_notice { $log_line->('notice', @_); }
+sub log_warn { $log_line->('warn', @_);  }
+sub log_fail { $log_line->('fail', @_); }
+
+my $print_header_first = 1;
+sub print_header {
+    my ($h) = @_;
+    print "\n" if !$print_header_first;
+    print "= $h =\n\n";
+    $print_header_first = 0;
+}
+
+my $get_systemd_unit_state = sub {
+    my ($unit, $suppress_stderr) = @_;
+
+    my $state;
+    my $filter_output = sub {
+       $state = shift;
+       chomp $state;
+    };
+
+    my %extra = (outfunc => $filter_output, noerr => 1);
+    $extra{errfunc} = sub {  } if $suppress_stderr;
+
+    eval {
+       run_command(['systemctl', 'is-enabled', "$unit"], %extra);
+       return if !defined($state);
+       run_command(['systemctl', 'is-active', "$unit"], %extra);
+    };
+
+    return $state // 'unknown';
+};
+my $log_systemd_unit_state = sub {
+    my ($unit, $no_fail_on_inactive) = @_;
+
+    my $log_method = \&log_warn;
+
+    my $state = $get_systemd_unit_state->($unit);
+    if ($state eq 'active') {
+       $log_method = \&log_pass;
+    } elsif ($state eq 'inactive') {
+       $log_method = $no_fail_on_inactive ? \&log_warn : \&log_fail;
+    } elsif ($state eq 'failed') {
+       $log_method = \&log_fail;
+    }
+
+    $log_method->("systemd unit '$unit' is in state '$state'");
+};
+
+my $versions;
+my $get_pkg = sub {
+    my ($pkg) = @_;
+
+    $versions = eval { PVE::API2::APT->versions({ node => $nodename }) } if 
!defined($versions);
+
+    if (!defined($versions)) {
+       my $msg = "unable to retrieve package version information";
+       $msg .= "- $@" if $@;
+       log_fail("$msg");
+       return undef;
+    }
+
+    my $pkgs = [ grep { $_->{Package} eq $pkg } @$versions ];
+    if (!defined $pkgs || $pkgs == 0) {
+       log_fail("unable to determine installed $pkg version.");
+       return undef;
+    } else {
+       return $pkgs->[0];
+    }
+};
+
+sub check_pve_packages {
+    print_header("CHECKING VERSION INFORMATION FOR PVE PACKAGES");
+
+    print "Checking for package updates..\n";
+    my $updates = eval { PVE::API2::APT->list_updates({ node => $nodename }); 
};
+    if (!defined($updates)) {
+       log_warn("$@") if $@;
+       log_fail("unable to retrieve list of package updates!");
+    } elsif (@$updates > 0) {
+       my $pkgs = join(', ', map { $_->{Package} } @$updates);
+       log_warn("updates for the following packages are available:\n  $pkgs");
+    } else {
+       log_pass("all packages up-to-date");
+    }
+
+    print "\nChecking proxmox-ve package version..\n";
+    if (defined(my $proxmox_ve = $get_pkg->('proxmox-ve'))) {
+       # TODO: update to native version for pve8to9
+       my $min_pve_ver = "$min_pve_major.$min_pve_minor-$min_pve_pkgrel";
+
+       my ($maj, $min, $pkgrel) = $proxmox_ve->{OldVersion} =~ 
m/^(\d+)\.(\d+)[.-](\d+)/;
+
+       if ($maj > $min_pve_major) {
+           log_pass("already upgraded to Proxmox VE " . ($min_pve_major + 1));
+           $upgraded = 1;
+       } elsif ($maj >= $min_pve_major && $min >= $min_pve_minor && $pkgrel >= 
$min_pve_pkgrel) {
+           log_pass("proxmox-ve package has version >= $min_pve_ver");
+       } else {
+           log_fail("proxmox-ve package is too old, please upgrade to >= 
$min_pve_ver!");
+       }
+
+       # FIXME: better differentiate between 6.2 from bullseye or bookworm
+       my $kinstalled = 'proxmox-kernel-6.2';
+       if (!$upgraded) {
+           # we got a few that avoided 5.15 in cluster with mixed CPUs, so 
allow older too
+           $kinstalled = 'pve-kernel-5.15';
+       }
+
+       my $kernel_version_is_expected = sub {
+           my ($version) = @_;
+
+           return $version =~ m/^(?:5\.(?:13|15)|6\.2)/ if !$upgraded;
+
+           if ($version =~ m/^6\.(?:2\.(?:[2-9]\d+|1[6-8]|1\d\d+)|5)[^~]*$/) {
+               return 1;
+           } elsif ($version =~ m/^(\d+).(\d+)[^~]*-pve$/) {
+               return $1 >= 6 && $2 >= 2;
+           }
+           return 0;
+       };
+
+       print "\nChecking running kernel version..\n";
+       my $kernel_ver = $proxmox_ve->{RunningKernel};
+       if (!defined($kernel_ver)) {
+           log_fail("unable to determine running kernel version.");
+       } elsif ($kernel_version_is_expected->($kernel_ver)) {
+           if ($upgraded) {
+               log_pass("running new kernel '$kernel_ver' after upgrade.");
+           } else {
+               log_pass("running kernel '$kernel_ver' is considered suitable 
for upgrade.");
+           }
+       } elsif ($get_pkg->($kinstalled)) {
+           # with 6.2 kernel being available in both we might want to 
fine-tune the check?
+           log_warn("a suitable kernel ($kinstalled) is installed, but an 
unsuitable ($kernel_ver) is booted, missing reboot?!");
+       } else {
+           log_warn("unexpected running and installed kernel '$kernel_ver'.");
+       }
+
+       if ($upgraded && $kernel_version_is_expected->($kernel_ver)) {
+           my $outdated_kernel_meta_pkgs = [];
+           for my $kernel_meta_version ('5.4', '5.11', '5.13', '5.15') {
+               my $pkg = "pve-kernel-${kernel_meta_version}";
+               if ($get_pkg->($pkg)) {
+                   push @$outdated_kernel_meta_pkgs, $pkg;
+               }
+           }
+           if (scalar(@$outdated_kernel_meta_pkgs) > 0) {
+               log_info(
+                   "Found outdated kernel meta-packages, taking up extra space 
on boot partitions.\n"
+                   ."      After a successful upgrade, you can remove them 
using this command:\n"
+                   ."      apt remove " . join(' ', 
$outdated_kernel_meta_pkgs->@*)
+               );
+           }
+       }
+    } else {
+       log_fail("proxmox-ve package not found!");
+    }
+}
+
+
+sub check_storage_health {
+    print_header("CHECKING CONFIGURED STORAGES");
+    my $cfg = PVE::Storage::config();
+
+    my $ctime = time();
+
+    my $info = PVE::Storage::storage_info($cfg);
+
+    for my $storeid (sort keys %$info) {
+       my $d = $info->{$storeid};
+       if ($d->{enabled}) {
+           if ($d->{active}) {
+               log_pass("storage '$storeid' enabled and active.");
+           } else {
+               log_warn("storage '$storeid' enabled but not active!");
+           }
+       } else {
+           log_skip("storage '$storeid' disabled.");
+       }
+    }
+
+    check_storage_content();
+    eval { check_storage_content_dirs() };
+    log_fail("failed to check storage content directories - $@") if $@;
+}
+
+sub check_cluster_corosync {
+    print_header("CHECKING CLUSTER HEALTH/SETTINGS");
+
+    if (!PVE::Corosync::check_conf_exists(1)) {
+       log_skip("standalone node.");
+       return;
+    }
+
+    $log_systemd_unit_state->('pve-cluster.service');
+    $log_systemd_unit_state->('corosync.service');
+
+    if (PVE::Cluster::check_cfs_quorum(1)) {
+       log_pass("Cluster Filesystem is quorate.");
+    } else {
+       log_fail("Cluster Filesystem readonly, lost quorum?!");
+    }
+
+    my $conf = PVE::Cluster::cfs_read_file('corosync.conf');
+    my $conf_nodelist = PVE::Corosync::nodelist($conf);
+    my $node_votes = 0;
+
+    print "\nAnalzying quorum settings and state..\n";
+    if (!defined($conf_nodelist)) {
+       log_fail("unable to retrieve nodelist from corosync.conf");
+    } else {
+       if (grep { $conf_nodelist->{$_}->{quorum_votes} != 1 } keys 
%$conf_nodelist) {
+           log_warn("non-default quorum_votes distribution detected!");
+       }
+       map { $node_votes += $conf_nodelist->{$_}->{quorum_votes} // 0 } keys 
%$conf_nodelist;
+    }
+
+    my ($expected_votes, $total_votes);
+    my $filter_output = sub {
+       my $line = shift;
+       ($expected_votes) = $line =~ /^Expected votes:\s*(\d+)\s*$/
+           if !defined($expected_votes);
+       ($total_votes) = $line =~ /^Total votes:\s*(\d+)\s*$/
+           if !defined($total_votes);
+    };
+    eval {
+       run_command(['corosync-quorumtool', '-s'], outfunc => $filter_output, 
noerr => 1);
+    };
+
+    if (!defined($expected_votes)) {
+       log_fail("unable to get expected number of votes, assuming 0.");
+       $expected_votes = 0;
+    }
+    if (!defined($total_votes)) {
+       log_fail("unable to get expected number of votes, assuming 0.");
+       $total_votes = 0;
+    }
+
+    my $cfs_nodelist = PVE::Cluster::get_clinfo()->{nodelist};
+    my $offline_nodes = grep { $cfs_nodelist->{$_}->{online} != 1 } keys 
%$cfs_nodelist;
+    if ($offline_nodes > 0) {
+       log_fail("$offline_nodes nodes are offline!");
+    }
+
+    my $qdevice_votes = 0;
+    if (my $qdevice_setup = $conf->{main}->{quorum}->{device}) {
+       $qdevice_votes = $qdevice_setup->{votes} // 1;
+    }
+
+    log_info("configured votes - nodes: $node_votes");
+    log_info("configured votes - qdevice: $qdevice_votes");
+    log_info("current expected votes: $expected_votes");
+    log_info("current total votes: $total_votes");
+
+    log_warn("expected votes set to non-standard value '$expected_votes'.")
+       if $expected_votes != $node_votes + $qdevice_votes;
+    log_warn("total votes < expected votes: $total_votes/$expected_votes!")
+       if $total_votes < $expected_votes;
+
+    my $conf_nodelist_count = scalar(keys %$conf_nodelist);
+    my $cfs_nodelist_count = scalar(keys %$cfs_nodelist);
+    log_warn("cluster consists of less than three quorum-providing nodes!")
+       if $conf_nodelist_count < 3 && $conf_nodelist_count + $qdevice_votes < 
3;
+
+    log_fail("corosync.conf ($conf_nodelist_count) and pmxcfs 
($cfs_nodelist_count) don't agree about size of nodelist.")
+       if $conf_nodelist_count != $cfs_nodelist_count;
+
+    print "\nChecking nodelist entries..\n";
+    my $nodelist_pass = 1;
+    for my $cs_node (sort keys %$conf_nodelist) {
+       my $entry = $conf_nodelist->{$cs_node};
+       if (!defined($entry->{name})) {
+           $nodelist_pass = 0;
+           log_fail("$cs_node: no name entry in corosync.conf.");
+       }
+       if (!defined($entry->{nodeid})) {
+           $nodelist_pass = 0;
+           log_fail("$cs_node: no nodeid configured in corosync.conf.");
+       }
+       my $gotLinks = 0;
+       for my $link (0..7) {
+           $gotLinks++ if defined($entry->{"ring${link}_addr"});
+       }
+       if ($gotLinks <= 0) {
+           $nodelist_pass = 0;
+           log_fail("$cs_node: no ringX_addr (0 <= X <= 7) link defined in 
corosync.conf.");
+       }
+
+       my $verify_ring_ip = sub {
+           my $key = shift;
+           if (defined(my $ring = $entry->{$key})) {
+               my ($resolved_ip, undef) = 
PVE::Corosync::resolve_hostname_like_corosync($ring, $conf);
+               if (defined($resolved_ip)) {
+                   if ($resolved_ip ne $ring) {
+                       $nodelist_pass = 0;
+                       log_warn(
+                           "$cs_node: $key '$ring' resolves to 
'$resolved_ip'.\n"
+                           ." Consider replacing it with the currently 
resolved IP address."
+                       );
+                   }
+               } else {
+                   $nodelist_pass = 0;
+                   log_fail(
+                       "$cs_node: unable to resolve $key '$ring' to an IP 
address according to Corosync's"
+                       ." resolve strategy - cluster will potentially fail 
with Corosync 3.x/kronosnet!"
+                   );
+               }
+           }
+       };
+       for my $link (0..7) {
+           $verify_ring_ip->("ring${link}_addr");
+       }
+    }
+    log_pass("nodelist settings OK") if $nodelist_pass;
+
+    print "\nChecking totem settings..\n";
+    my $totem = $conf->{main}->{totem};
+    my $totem_pass = 1;
+
+    my $transport = $totem->{transport};
+    if (defined($transport)) {
+       if ($transport ne 'knet') {
+           $totem_pass = 0;
+           log_fail("Corosync transport explicitly set to '$transport' instead 
of implicit default!");
+       }
+    }
+
+    # TODO: are those values still up-to-date?
+    if ((!defined($totem->{secauth}) || $totem->{secauth} ne 'on') && 
(!defined($totem->{crypto_cipher}) || $totem->{crypto_cipher} eq 'none')) {
+       $totem_pass = 0;
+       log_fail("Corosync authentication/encryption is not explicitly enabled 
(secauth / crypto_cipher / crypto_hash)!");
+    } elsif (defined($totem->{crypto_cipher}) && $totem->{crypto_cipher} eq 
'3des') {
+       $totem_pass = 0;
+       log_fail("Corosync encryption cipher set to '3des', no longer supported 
in Corosync 3.x!"); # FIXME: can be removed?
+    }
+
+    log_pass("totem settings OK") if $totem_pass;
+    print "\n";
+    log_info("run 'pvecm status' to get detailed cluster status..");
+
+    if (defined(my $corosync = $get_pkg->('corosync'))) {
+       if ($corosync->{OldVersion} =~ m/^2\./) {
+           log_fail("\ncorosync 2.x installed, cluster-wide upgrade to 3.x 
needed!");
+       } elsif ($corosync->{OldVersion} !~ m/^3\./) {
+           log_fail("\nunexpected corosync version installed: 
$corosync->{OldVersion}!");
+       }
+    }
+}
+
+sub check_ceph {
+    print_header("CHECKING HYPER-CONVERGED CEPH STATUS");
+
+    if (PVE::Ceph::Tools::check_ceph_inited(1)) {
+       log_info("hyper-converged ceph setup detected!");
+    } else {
+       log_skip("no hyper-converged ceph setup detected!");
+       return;
+    }
+
+    log_info("getting Ceph status/health information..");
+    my $ceph_status = eval { PVE::API2::Ceph->status({ node => $nodename }); };
+    my $noout = eval { PVE::API2::Cluster::Ceph->get_flag({ flag => "noout" 
}); };
+    if ($@) {
+       log_fail("failed to get 'noout' flag status - $@");
+    }
+
+    my $noout_wanted = 1;
+
+    if (!$ceph_status || !$ceph_status->{health}) {
+       log_fail("unable to determine Ceph status!");
+    } else {
+       my $ceph_health = $ceph_status->{health}->{status};
+       if (!$ceph_health) {
+           log_fail("unable to determine Ceph health!");
+       } elsif ($ceph_health eq 'HEALTH_OK') {
+           log_pass("Ceph health reported as 'HEALTH_OK'.");
+       } elsif ($ceph_health eq 'HEALTH_WARN' && $noout && (keys 
%{$ceph_status->{health}->{checks}} == 1)) {
+               log_pass("Ceph health reported as 'HEALTH_WARN' with a single 
failing check and 'noout' flag set.");
+       } else {
+               log_warn(
+                   "Ceph health reported as '$ceph_health'.\n      Use the PVE 
dashboard or 'ceph -s'"
+                   ." to determine the specific issues and try to resolve 
them."
+               );
+       }
+    }
+
+    # TODO: check OSD min-required version, if to low it breaks stuff!
+
+    log_info("checking local Ceph version..");
+    if (my $release = eval { PVE::Ceph::Tools::get_local_version(1) }) {
+       my $code_name = $ceph_release2code->{"$release"} || 'unknown';
+       if ($release == $ceph_supported_release) {
+           log_pass("found expected Ceph $ceph_supported_release 
$ceph_supported_code_name release.")
+       } elsif ($release > $ceph_supported_release) {
+           log_warn(
+               "found newer Ceph release $release $code_name as the expected 
$ceph_supported_release"
+               ." $ceph_supported_code_name, installed third party repos?!"
+           )
+       } else {
+           log_fail(
+               "Hyper-converged Ceph $release $code_name is to old for 
upgrade!\n"
+               ."      Upgrade Ceph first to $ceph_supported_code_name 
following our how-to:\n"
+               ."      <https://pve.proxmox.com/wiki/Category:Ceph_Upgrade>"
+           );
+       }
+    } else {
+       log_fail("unable to determine local Ceph version!");
+    }
+
+    log_info("getting Ceph daemon versions..");
+    my $ceph_versions = eval { PVE::Ceph::Tools::get_cluster_versions(undef, 
1); };
+    if (!$ceph_versions) {
+       log_fail("unable to determine Ceph daemon versions!");
+    } else {
+       my $services = [
+           { 'key' => 'mon', 'name' => 'monitor' },
+           { 'key' => 'mgr', 'name' => 'manager' },
+           { 'key' => 'mds', 'name' => 'MDS' },
+           { 'key' => 'osd', 'name' => 'OSD' },
+       ];
+
+       my $ceph_versions_simple = {};
+       my $ceph_versions_commits = {};
+       for my $type (keys %$ceph_versions) {
+           for my $full_version (keys $ceph_versions->{$type}->%*) {
+               if ($full_version =~ m/^(.*) \((.*)\).*\(.*\)$/) {
+                   # String is in the form of
+                   # ceph version 17.2.6 
(810db68029296377607028a6c6da1ec06f5a2b27) quincy (stable)
+                   # only check the first part, e.g. 'ceph version 17.2.6', 
the commit hash can
+                   # be different
+                   $ceph_versions_simple->{$type}->{$1} = 1;
+                   $ceph_versions_commits->{$type}->{$2} = 1;
+               }
+           }
+       }
+
+       for my $service (@$services) {
+           my ($name, $key) = $service->@{'name', 'key'};
+           if (my $service_versions = $ceph_versions_simple->{$key}) {
+               if (keys %$service_versions == 0) {
+                   log_skip("no running instances detected for daemon type 
$name.");
+               } elsif (keys %$service_versions == 1) {
+                   log_pass("single running version detected for daemon type 
$name.");
+               } else {
+                   log_warn("multiple running versions detected for daemon 
type $name!");
+               }
+           } else {
+               log_skip("unable to determine versions of running Ceph $name 
instances.");
+           }
+           my $service_commits = $ceph_versions_commits->{$key};
+           log_info("different builds of same version detected for an $name. 
Are you in the middle of the upgrade?")
+               if $service_commits && keys %$service_commits > 1;
+       }
+
+       my $overall_versions = $ceph_versions->{overall};
+       if (!$overall_versions) {
+           log_warn("unable to determine overall Ceph daemon versions!");
+       } elsif (keys %$overall_versions == 1) {
+           log_pass("single running overall version detected for all Ceph 
daemon types.");
+           $noout_wanted = !$upgraded; # off post-upgrade, on pre-upgrade
+       } elsif (keys $ceph_versions_simple->{overall}->%* != 1) {
+           log_warn("overall version mismatch detected, check 'ceph versions' 
output for details!");
+       }
+    }
+
+    if ($noout) {
+       if ($noout_wanted) {
+           log_pass("'noout' flag set to prevent rebalancing during 
cluster-wide upgrades.");
+       } else {
+           log_warn("'noout' flag set, Ceph cluster upgrade seems finished.");
+       }
+    } elsif ($noout_wanted) {
+       log_warn("'noout' flag not set - recommended to prevent rebalancing 
during upgrades.");
+    }
+
+    log_info("checking Ceph config..");
+    my $conf = PVE::Cluster::cfs_read_file('ceph.conf');
+    if (%$conf) {
+       my $global = $conf->{global};
+
+       my $global_monhost = $global->{mon_host} // $global->{"mon host"} // 
$global->{"mon-host"};
+       if (!defined($global_monhost)) {
+           log_warn(
+               "No 'mon_host' entry found in ceph config.\n  It's recommended 
to add mon_host with"
+               ." all monitor addresses (without ports) to the global section."
+           );
+       }
+
+       my $ipv6 = $global->{ms_bind_ipv6} // $global->{"ms bind ipv6"} // 
$global->{"ms-bind-ipv6"};
+       if ($ipv6) {
+           my $ipv4 = $global->{ms_bind_ipv4} // $global->{"ms bind ipv4"} // 
$global->{"ms-bind-ipv4"};
+           if ($ipv6 eq 'true' && (!defined($ipv4) || $ipv4 ne 'false')) {
+               log_warn(
+                   "'ms_bind_ipv6' is enabled but 'ms_bind_ipv4' is not 
disabled.\n  Make sure to"
+                   ." disable 'ms_bind_ipv4' for ipv6 only clusters, or add an 
ipv4 network to public/cluster network."
+               );
+           }
+       }
+
+       if (defined($global->{keyring})) {
+           log_warn(
+               "[global] config section contains 'keyring' option, which will 
prevent services from"
+               ." starting with Nautilus.\n Move 'keyring' option to [client] 
section instead."
+           );
+       }
+
+    } else {
+       log_warn("Empty ceph config found");
+    }
+
+    my $local_ceph_ver = PVE::Ceph::Tools::get_local_version(1);
+    if (defined($local_ceph_ver)) {
+       if ($local_ceph_ver <= 14) {
+           log_fail("local Ceph version too low, at least Octopus required..");
+       }
+    } else {
+       log_fail("unable to determine local Ceph version.");
+    }
+}
+
+sub check_backup_retention_settings {
+    log_info("Checking backup retention settings..");
+
+    my $pass = 1;
+
+    my $maxfiles_msg = "parameter 'maxfiles' is deprecated with PVE 7.x and 
will be removed in a " .
+       "future version, use 'prune-backups' instead.";
+
+    eval {
+       my $confdesc = PVE::VZDump::Common::get_confdesc();
+       # vzdump.conf by itself doesn't need to honor any 'requires'
+       delete $confdesc->{$_}->{requires} for keys $confdesc->%*;
+
+       my $fn = "/etc/vzdump.conf";
+       my $raw = PVE::Tools::file_get_contents($fn);
+
+       my $conf_schema = { type => 'object', properties => $confdesc, };
+       my $param = PVE::JSONSchema::parse_config($conf_schema, $fn, $raw);
+
+       if (defined($param->{maxfiles})) {
+           $pass = 0;
+           log_warn("$fn - $maxfiles_msg");
+       }
+    };
+    if (my $err = $@) {
+       $pass = 0;
+       log_warn("unable to parse node's VZDump configuration - $err");
+    }
+
+    my $storage_cfg = PVE::Storage::config();
+
+    for my $storeid (keys $storage_cfg->{ids}->%*) {
+       my $scfg = $storage_cfg->{ids}->{$storeid};
+
+       if (defined($scfg->{maxfiles})) {
+           $pass = 0;
+           log_warn("storage '$storeid' - $maxfiles_msg");
+       }
+    }
+
+    eval {
+       my $vzdump_cron = PVE::Cluster::cfs_read_file('vzdump.cron');
+
+       # only warn once, there might be many jobs...
+       if (scalar(grep { defined($_->{maxfiles}) } $vzdump_cron->{jobs}->@*)) {
+           $pass = 0;
+           log_warn("/etc/pve/vzdump.cron - $maxfiles_msg");
+       }
+    };
+    if (my $err = $@) {
+       $pass = 0;
+       log_warn("unable to parse node's VZDump configuration - $err");
+    }
+
+    log_pass("no backup retention problems found.") if $pass;
+}
+
+sub check_cifs_credential_location {
+    log_info("checking CIFS credential location..");
+
+    my $regex = qr/^(.*)\.cred$/;
+
+    my $found;
+
+    PVE::Tools::dir_glob_foreach('/etc/pve/priv/', $regex, sub {
+       my ($filename) = @_;
+
+       my ($basename) = $filename =~ $regex;
+
+       log_warn(
+           "CIFS credentials '/etc/pve/priv/$filename' will be moved to"
+           ." '/etc/pve/priv/storage/$basename.pw' during the update"
+       );
+
+       $found = 1;
+    });
+
+    log_pass("no CIFS credentials at outdated location found.") if !$found;
+}
+
+sub check_custom_pool_roles {
+    log_info("Checking permission system changes..");
+
+    if (! -f "/etc/pve/user.cfg") {
+       log_skip("user.cfg does not exist");
+       return;
+    }
+
+    my $raw = eval { PVE::Tools::file_get_contents('/etc/pve/user.cfg'); };
+    if ($@) {
+       log_fail("Failed to read '/etc/pve/user.cfg' - $@");
+       return;
+    }
+
+    my $roles = {};
+    while ($raw =~ /^\s*(.+?)\s*$/gm) {
+       my $line = $1;
+       my @data;
+
+       for my $d (split (/:/, $line)) {
+           $d =~ s/^\s+//;
+           $d =~ s/\s+$//;
+           push @data, $d
+       }
+
+       my $et = shift @data;
+       if ($et eq 'role') {
+           my ($role, $privlist) = @data;
+           if (!PVE::AccessControl::verify_rolename($role, 1)) {
+               warn "user config - ignore role '$role' - invalid characters in 
role name\n";
+               next;
+           }
+
+           $roles->{$role} = {} if !$roles->{$role};
+           for my $priv (split_list($privlist)) {
+               $roles->{$role}->{$priv} = 1;
+           }
+       } elsif ($et eq 'acl') {
+           my ($propagate, $pathtxt, $uglist, $rolelist) = @data;
+           for my $role (split_list($rolelist)) {
+               if ($role eq 'PVESysAdmin' || $role eq 'PVEAdmin') {
+                   log_warn(
+                       "found ACL entry on '$pathtxt' for '$uglist' with role 
'$role' - this role"
+                       ." will no longer have 'Permissions.Modify' after the 
upgrade!"
+                   );
+               }
+           }
+       }
+    }
+
+    log_info("Checking custom role IDs for clashes with new 'PVE' 
namespace..");
+    my ($custom_roles, $pve_namespace_clashes) = (0, 0);
+    for my $role (sort keys %{$roles}) {
+       next if PVE::AccessControl::role_is_special($role);
+       $custom_roles++;
+
+       if ($role =~ /^PVE/i) {
+           log_warn("custom role '$role' clashes with 'PVE' namespace for 
built-in roles");
+           $pve_namespace_clashes++;
+       }
+    }
+    if ($pve_namespace_clashes > 0) {
+       log_fail("$pve_namespace_clashes custom role(s) will clash with 'PVE' 
namespace for built-in roles enforced in Proxmox VE 8");
+    } elsif ($custom_roles > 0) {
+       log_pass("none of the $custom_roles custom roles will clash with newly 
enforced 'PVE' namespace")
+    } else {
+       log_pass("no custom roles defined, so no clash with 'PVE' role ID 
namespace enforced in Proxmox VE 8")
+    }
+}
+
+my sub check_max_length {
+    my ($raw, $max_length, $warning) = @_;
+    log_warn($warning) if defined($raw) && length($raw) > $max_length; 
+}
+
+sub check_node_and_guest_configurations {
+    log_info("Checking node and guest description/note length..");
+
+    my @affected_nodes = grep {
+       my $desc = PVE::NodeConfig::load_config($_)->{desc};
+       defined($desc) && length($desc) > 64 * 1024
+    } PVE::Cluster::get_nodelist();
+
+    if (scalar(@affected_nodes) > 0) {
+       log_warn("Node config description of the following nodes too long for 
new limit of 64 KiB:\n    "
+           . join(', ', @affected_nodes));
+    } else {
+       log_pass("All node config descriptions fit in the new limit of 64 KiB");
+    }
+
+    my $affected_guests_long_desc = [];
+    my $affected_cts_cgroup_keys = [];
+
+    my $cts = PVE::LXC::config_list();
+    for my $vmid (sort { $a <=> $b } keys %$cts) {
+       my $conf = PVE::LXC::Config->load_config($vmid);
+
+       my $desc = $conf->{description};
+       push @$affected_guests_long_desc, "CT $vmid" if defined($desc) && 
length($desc) > 8 * 1024;
+
+       my $lxc_raw_conf = $conf->{lxc};
+       push @$affected_cts_cgroup_keys, "CT $vmid"  if (grep (@$_[0] =~ 
/^lxc\.cgroup\./, @$lxc_raw_conf));
+    }
+    my $vms = PVE::QemuServer::config_list();
+    for my $vmid (sort { $a <=> $b } keys %$vms) {
+       my $desc = PVE::QemuConfig->load_config($vmid)->{description};
+       push @$affected_guests_long_desc, "VM $vmid" if defined($desc) && 
length($desc) > 8 * 1024;
+    }
+    if (scalar($affected_guests_long_desc->@*) > 0) {
+       log_warn("Guest config description of the following virtual-guests too 
long for new limit of 64 KiB:\n"
+           ."    " . join(", ", $affected_guests_long_desc->@*));
+    } else {
+       log_pass("All guest config descriptions fit in the new limit of 8 KiB");
+    }
+
+    log_info("Checking container configs for deprecated lxc.cgroup entries");
+
+    if (scalar($affected_cts_cgroup_keys->@*) > 0) {
+       if ($forced_legacy_cgroup) {
+           log_pass("Found legacy 'lxc.cgroup' keys, but system explicitly 
configured for legacy hybrid cgroup hierarchy.");
+       }  else {
+           log_warn("The following CTs have 'lxc.cgroup' keys configured, 
which will be ignored in the new default unified cgroupv2:\n"
+               ."    " . join(", ", $affected_cts_cgroup_keys->@*) ."\n"
+               ."    Often it can be enough to change to the new 'lxc.cgroup2' 
prefix after the upgrade to Proxmox VE 7.x");
+       }
+    } else {
+       log_pass("No legacy 'lxc.cgroup' keys found.");
+    }
+}
+
+sub check_storage_content {
+    log_info("Checking storage content type configuration..");
+
+    my $found;
+    my $pass = 1;
+
+    my $storage_cfg = PVE::Storage::config();
+
+    for my $storeid (sort keys $storage_cfg->{ids}->%*) {
+       my $scfg = $storage_cfg->{ids}->{$storeid};
+
+       next if $scfg->{shared};
+       next if !PVE::Storage::storage_check_enabled($storage_cfg, $storeid, 
undef, 1);
+
+       my $valid_content = 
PVE::Storage::Plugin::valid_content_types($scfg->{type});
+
+       if (scalar(keys $scfg->{content}->%*) == 0 && !$valid_content->{none}) {
+           $pass = 0;
+           log_fail("storage '$storeid' does not support configured content 
type 'none'");
+           delete $scfg->{content}->{none}; # scan for guest images below
+       }
+
+       next if $scfg->{content}->{images};
+       next if $scfg->{content}->{rootdir};
+
+       # Skip 'iscsi(direct)' (and foreign plugins with potentially similar 
behavior) with 'none',
+       # because that means "use LUNs directly" and vdisk_list() in PVE 6.x 
still lists those.
+       # It's enough to *not* skip 'dir', because it is the only other storage 
that supports 'none'
+       # and 'images' or 'rootdir', hence being potentially misconfigured.
+       next if $scfg->{type} ne 'dir' && $scfg->{content}->{none};
+
+       eval { PVE::Storage::activate_storage($storage_cfg, $storeid) };
+       if (my $err = $@) {
+           log_warn("activating '$storeid' failed - $err");
+           next;
+       }
+
+       my $res = eval { PVE::Storage::vdisk_list($storage_cfg, $storeid); };
+       if (my $err = $@) {
+           log_warn("listing images on '$storeid' failed - $err");
+           next;
+       }
+       my @volids = map { $_->{volid} } $res->{$storeid}->@*;
+
+       my $number = scalar(@volids);
+       if ($number > 0) {
+           log_info(
+               "storage '$storeid' - neither content type 'images' nor 
'rootdir' configured, but"
+               ."found $number guest volume(s)"
+           );
+       }
+    }
+
+    my $check_volid = sub {
+       my ($volid, $vmid, $vmtype, $reference) = @_;
+
+       my $guesttext = $vmtype eq 'qemu' ? 'VM' : 'CT';
+       my $prefix = "$guesttext $vmid - volume '$volid' ($reference)";
+
+       my ($storeid) = PVE::Storage::parse_volume_id($volid, 1);
+       return if !defined($storeid);
+
+       my $scfg = $storage_cfg->{ids}->{$storeid};
+       if (!$scfg) {
+           $pass = 0;
+           log_warn("$prefix - storage does not exist!");
+           return;
+       }
+
+       # cannot use parse_volname for containers, as it can return 'images'
+       # but containers cannot have ISO images attached, so assume 'rootdir'
+       my $vtype = 'rootdir';
+       if ($vmtype eq 'qemu') {
+           ($vtype) = eval { PVE::Storage::parse_volname($storage_cfg, 
$volid); };
+           return if $@;
+       }
+
+       if (!$scfg->{content}->{$vtype}) {
+           $found = 1;
+           $pass = 0;
+           log_warn("$prefix - storage does not have content type '$vtype' 
configured.");
+       }
+    };
+
+    my $cts = PVE::LXC::config_list();
+    for my $vmid (sort { $a <=> $b } keys %$cts) {
+       my $conf = PVE::LXC::Config->load_config($vmid);
+
+       my $volhash = {};
+
+       my $check = sub {
+           my ($ms, $mountpoint, $reference) = @_;
+
+           my $volid = $mountpoint->{volume};
+           return if !$volid || $mountpoint->{type} ne 'volume';
+
+           return if $volhash->{$volid}; # volume might be referenced multiple 
times
+
+           $volhash->{$volid} = 1;
+
+           $check_volid->($volid, $vmid, 'lxc', $reference);
+       };
+
+       my $opts = { include_unused => 1 };
+       PVE::LXC::Config->foreach_volume_full($conf, $opts, $check, 'in 
config');
+       for my $snapname (keys $conf->{snapshots}->%*) {
+           my $snap = $conf->{snapshots}->{$snapname};
+           PVE::LXC::Config->foreach_volume_full($snap, $opts, $check, "in 
snapshot '$snapname'");
+       }
+    }
+
+    my $vms = PVE::QemuServer::config_list();
+    for my $vmid (sort { $a <=> $b } keys %$vms) {
+       my $conf = PVE::QemuConfig->load_config($vmid);
+
+       my $volhash = {};
+
+       my $check = sub {
+           my ($key, $drive, $reference) = @_;
+
+           my $volid = $drive->{file};
+           return if $volid =~ m|^/|;
+           return if $volhash->{$volid}; # volume might be referenced multiple 
times
+
+           $volhash->{$volid} = 1;
+           $check_volid->($volid, $vmid, 'qemu', $reference);
+       };
+
+       my $opts = {
+           extra_keys => ['vmstate'],
+           include_unused => 1,
+       };
+       # startup from a suspended state works even without 'images' content 
type on the
+       # state storage, so do not check 'vmstate' for $conf
+       PVE::QemuConfig->foreach_volume_full($conf, { include_unused => 1 }, 
$check, 'in config');
+       for my $snapname (keys $conf->{snapshots}->%*) {
+           my $snap = $conf->{snapshots}->{$snapname};
+           PVE::QemuConfig->foreach_volume_full($snap, $opts, $check, "in 
snapshot '$snapname'");
+       }
+    }
+
+    if ($found) {
+       log_warn("Proxmox VE enforces stricter content type checks since 7.0. 
The guests above " .
+           "might not work until the storage configuration is fixed.");
+    }
+
+    if ($pass) {
+       log_pass("no storage content problems found");
+    }
+}
+
+sub check_storage_content_dirs {
+    my $storage_cfg = PVE::Storage::config();
+
+    # check that content dirs are pairwise inequal
+    my $any_problematic = 0;
+    for my $storeid (sort keys $storage_cfg->{ids}->%*) {
+       my $scfg = $storage_cfg->{ids}->{$storeid};
+
+       next if !PVE::Storage::storage_check_enabled($storage_cfg, $storeid, 
undef, 1);
+       next if !$scfg->{path} || !$scfg->{content};
+
+       eval { PVE::Storage::activate_storage($storage_cfg, $storeid) };
+       if (my $err = $@) {
+           log_warn("activating '$storeid' failed - $err");
+           next;
+       }
+
+       my $resolved_subdirs = {};
+       my $plugin = PVE::Storage::Plugin->lookup($scfg->{type});
+       for my $vtype (keys $scfg->{content}->%*) {
+           my $abs_subdir = Cwd::abs_path($plugin->get_subdir($scfg, $vtype));
+           next if !defined($abs_subdir);
+           push $resolved_subdirs->{$abs_subdir}->@*, $vtype;
+       }
+       for my $subdir (keys $resolved_subdirs->%*) {
+           if (scalar($resolved_subdirs->{$subdir}->@*) > 1) {
+               my $types = join(", ", $resolved_subdirs->{$subdir}->@*);
+               log_warn("storage '$storeid' uses directory $subdir for 
multiple content types ($types).");
+               $any_problematic = 1;
+            }
+       }
+    }
+    if ($any_problematic) {
+       log_fail("re-using directory for multiple content types (see above) is 
no longer supported in Proxmox VE 8!")
+    } else {
+       log_pass("no storage re-uses a directory for multiple content types.")
+    }
+}
+
+sub check_containers_cgroup_compat {
+    if ($forced_legacy_cgroup) {
+       log_warn("System explicitly configured for legacy hybrid cgroup 
hierarchy.\n"
+           ."     NOTE: support for the hybrid cgroup hierarchy will be 
removed in future Proxmox VE 9 (~ 2025)."
+       );
+    }
+
+    my $supports_cgroupv2 = sub {
+       my ($conf, $rootdir, $ctid) = @_;
+
+       my $get_systemd_version = sub {
+           my ($self) = @_;
+
+           my @dirs = (
+               '/lib/systemd',
+               '/usr/lib/systemd',
+               '/usr/lib/x86_64-linux-gnu/systemd',
+               '/usr/lib64/systemd'
+           );
+           my $libsd;
+           for my $dir (@dirs) {
+               $libsd = PVE::Tools::dir_glob_regex($dir, 
"libsystemd-shared-.+\.so");
+               last if defined($libsd);
+           }
+           if (defined($libsd) && $libsd =~ 
/libsystemd-shared-(\d+)(\.\d-\d)?(\.fc\d\d)?\.so/) {
+               return $1;
+           }
+
+           return undef;
+       };
+
+       my  $unified_cgroupv2_support = sub {
+           my ($self) = @_;
+
+           # https://www.freedesktop.org/software/systemd/man/systemd.html
+           # systemd is installed as symlink to /sbin/init
+           my $systemd = CORE::readlink('/sbin/init');
+
+           # assume non-systemd init will run with unified cgroupv2
+           if (!defined($systemd) || $systemd !~ m@/systemd$@) {
+               return 1;
+           }
+
+           # systemd version 232 (e.g. debian stretch) supports the unified 
hierarchy
+           my $sdver = $get_systemd_version->();
+           if (!defined($sdver) || $sdver < 232) {
+               return 0;
+           }
+
+           return 1;
+       };
+
+       my $ostype = $conf->{ostype};
+       if (!defined($ostype)) {
+           log_warn("Found CT ($ctid) without 'ostype' set!");
+       } elsif ($ostype eq 'devuan' || $ostype eq 'alpine') {
+           return 1; # no systemd, no cgroup problems
+       }
+
+       my $lxc_setup = PVE::LXC::Setup->new($conf, $rootdir);
+       return $lxc_setup->protected_call($unified_cgroupv2_support);
+    };
+
+    my $log_problem = sub {
+       my ($ctid) = @_;
+       my $extra = $forced_legacy_cgroup ? '' : " or set 
systemd.unified_cgroup_hierarchy=0 in the Proxmox VE hosts' kernel cmdline";
+       log_warn(
+           "Found at least one CT ($ctid) which does not support running in a 
unified cgroup v2 layout\n"
+           ."    Consider upgrading the Containers distro${extra}! Skipping 
further CT compat checks."
+       );
+    };
+
+    my $cts = eval { PVE::API2::LXC->vmlist({ node => $nodename }) };
+    if ($@) {
+       log_warn("Failed to retrieve information about this node's CTs - $@");
+       return;
+    }
+
+    if (!defined($cts) || !scalar(@$cts)) {
+       log_skip("No containers on node detected.");
+       return;
+    }
+
+    my @running_cts = sort { $a <=> $b } grep { $_->{status} eq 'running' } 
@$cts;
+    my @offline_cts = sort { $a <=> $b } grep { $_->{status} ne 'running' } 
@$cts;
+
+    for my $ct (@running_cts) {
+       my $ctid = $ct->{vmid};
+       my $pid = eval { PVE::LXC::find_lxc_pid($ctid) };
+       if (my $err = $@) {
+           log_warn("Failed to get PID for running CT $ctid - $err");
+           next;
+       }
+       my $rootdir = "/proc/$pid/root";
+       my $conf = PVE::LXC::Config->load_config($ctid);
+
+       my $ret = eval { $supports_cgroupv2->($conf, $rootdir, $ctid) };
+       if (my $err = $@) {
+           log_warn("Failed to get cgroup support status for CT $ctid - $err");
+           next;
+       }
+       if (!$ret) {
+           $log_problem->($ctid);
+           return;
+       }
+    }
+
+    my $storage_cfg = PVE::Storage::config();
+    for my $ct (@offline_cts) {
+       my $ctid = $ct->{vmid};
+       my ($conf, $rootdir, $ret);
+       eval {
+           $conf = PVE::LXC::Config->load_config($ctid);
+           $rootdir = PVE::LXC::mount_all($ctid, $storage_cfg, $conf);
+           $ret = $supports_cgroupv2->($conf, $rootdir, $ctid);
+       };
+       if (my $err = $@) {
+           log_warn("Failed to load config and mount CT $ctid - $err");
+           eval { PVE::LXC::umount_all($ctid, $storage_cfg, $conf) };
+           next;
+       }
+       if (!$ret) {
+           $log_problem->($ctid);
+           eval { PVE::LXC::umount_all($ctid, $storage_cfg, $conf) };
+           last;
+       }
+
+       eval { PVE::LXC::umount_all($ctid, $storage_cfg, $conf) };
+    }
+};
+
+sub check_lxcfs_fuse_version {
+    log_info("Checking if LXCFS is running with FUSE3 library, if already 
upgraded..");
+    if (!$upgraded) {
+       log_skip("not yet upgraded, no need to check the FUSE library version 
LXCFS uses");
+       return;
+    }
+
+    my $lxcfs_pid = eval { file_get_contents('/run/lxcfs.pid') };
+    if (my $err = $@) {
+       log_fail("failed to get LXCFS pid - $err");
+       return;
+    }
+    chomp $lxcfs_pid;
+
+    my $lxcfs_maps = eval { file_get_contents("/proc/${lxcfs_pid}/maps") };
+    if (my $err = $@) {
+       log_fail("failed to get LXCFS maps - $err");
+       return;
+    }
+
+    if ($lxcfs_maps =~ /\/libfuse.so.2/s) {
+       log_warn("systems seems to be upgraded but LXCFS is still running with 
FUSE 2 library, not yet rebooted?")
+    } elsif ($lxcfs_maps =~ /\/libfuse3.so.3/s) {
+       log_pass("systems seems to be upgraded and LXCFS is running with FUSE 3 
library")
+    }
+    return;
+}
+
+sub check_apt_repos {
+    log_info("Checking if the suite for the Debian security repository is 
correct..");
+
+    my $found = 0;
+
+    my $dir = '/etc/apt/sources.list.d';
+    my $in_dir = 0;
+
+    # TODO: check that (original) debian and Proxmox VE mirrors are present.
+
+    my ($found_suite, $found_suite_where);
+    my ($mismatches, $strange_suites);
+
+    my $check_file = sub {
+       my ($file) = @_;
+
+       $file = "${dir}/${file}" if $in_dir;
+
+       my $raw = eval { PVE::Tools::file_get_contents($file) };
+       return if !defined($raw);
+       my @lines = split(/\n/, $raw);
+
+       my $number = 0;
+       for my $line (@lines) {
+           $number++;
+
+           next if length($line) == 0; # split would result in undef then...
+
+           ($line) = split(/#/, $line);
+
+           next if $line !~ m/^deb[[:space:]]/; # is case sensitive
+
+           my $suite;
+           if ($line =~ m|deb\s+\w+://\S+\s+(\S*)|i) {
+               $suite = $1;
+           } else {
+               next;
+           }
+           my $where = "in ${file}:${number}";
+
+           $suite =~ 
s/-(?:(?:proposed-)?updates|backports|debug|security)(?:-debug)?$//;
+           if ($suite ne $old_suite && $suite ne $new_suite && 
!$older_suites->{$suite}) {
+               push $strange_suites->@*, { suite => $suite, where => $where };
+               next;
+           }
+
+           if (!defined($found_suite)) {
+               $found_suite = $suite;
+               $found_suite_where = $where;
+           } elsif ($suite ne $found_suite) {
+               if (!defined($mismatches)) {
+                   $mismatches = [];
+                   push $mismatches->@*,
+                       { suite => $found_suite, where => $found_suite_where},
+                       { suite => $suite, where => $where};
+               } else {
+                   push $mismatches->@*, { suite => $suite, where => $where};
+               }
+           }
+       }
+    };
+
+    $check_file->("/etc/apt/sources.list");
+
+    $in_dir = 1;
+
+    PVE::Tools::dir_glob_foreach($dir, '^.*\.list$', $check_file);
+
+    if ($strange_suites) {
+       my @strange_list = map { "found suite $_->{suite} at $_->{where}" } 
$strange_suites->@*;
+       log_notice(
+           "found unusual suites that are neither old '$old_suite' nor new 
'$new_suite':"
+           ."\n    " . join("\n    ", @strange_list)
+           ."\n  Please ensure these repositories are shipping compatible 
packages for the upgrade!"
+       );
+    }
+    if (defined($mismatches)) {
+       my @mismatch_list = map { "found suite $_->{suite} at $_->{where}" } 
$mismatches->@*;
+
+       log_fail(
+           "Found mixed old and new package repository suites, fix before 
upgrading! Mismatches:"
+           ."\n    " . join("\n    ", @mismatch_list)
+           ."\n  Configure the same base-suite for all Proxmox and Debian 
provided repos and ask"
+           ." original vendor for any third-party repos."
+           ."\n  E.g., for the upgrade to Proxmox VE ".($min_pve_major + 1)." 
use the '$new_suite' suite."
+       );
+    } elsif (defined($strange_suites)) {
+       log_notice("found no suite mismatches, but found at least one strange 
suite");
+    } else {
+       log_pass("found no suite mismatch");
+    }
+}
+
+sub check_nvidia_vgpu_service {
+    log_info("Checking for existence of NVIDIA vGPU Manager..");
+
+    my $msg = "NVIDIA vGPU Service found, possibly not compatible with newer 
kernel versions, check"
+        ." with their documentation and 
https://pve.proxmox.com/wiki/Upgrade_from_7_to_8#Known_upgrade_issues.";;
+
+    my $state = $get_systemd_unit_state->("nvidia-vgpu-mgr.service", 1);
+    if ($state && $state eq 'active') {
+       log_warn("Running $msg");
+    } elsif ($state && $state ne 'unknown') {
+       log_warn($msg);
+    } else {
+       log_pass("No NVIDIA vGPU Service found.");
+    }
+}
+
+sub check_time_sync {
+    my $unit_active = sub { return $get_systemd_unit_state->($_[0], 1) eq 
'active' ? $_[0] : undef };
+
+    log_info("Checking for supported & active NTP service..");
+    if ($unit_active->('systemd-timesyncd.service')) {
+       log_warn(
+           "systemd-timesyncd is not the best choice for time-keeping on 
servers, due to only applying"
+           ." updates on boot.\n  While not necessary for the upgrade it's 
recommended to use one of:\n"
+           ."    * chrony (Default in new Proxmox VE installations)\n    * 
ntpsec\n    * openntpd\n"
+       );
+    } elsif ($unit_active->('ntp.service')) {
+       log_info("Debian deprecated and removed the ntp package for Bookworm, 
but the system"
+           ." will automatically migrate to the 'ntpsec' replacement package 
on upgrade.");
+    } elsif (my $active_ntp = ($unit_active->('chrony.service') || 
$unit_active->('openntpd.service') || $unit_active->('ntpsec.service'))) {
+       log_pass("Detected active time synchronisation unit '$active_ntp'");
+    } else {
+       log_warn(
+           "No (active) time synchronisation daemon (NTP) detected, but 
synchronized systems are important,"
+           ." especially for cluster and/or ceph!"
+       );
+    }
+}
+
+sub check_bootloader {
+    log_info("Checking bootloader configuration...");
+
+    if (! -d '/sys/firmware/efi') {
+       log_skip("System booted in legacy-mode - no need for additional 
packages");
+       return;
+    }
+
+    if ( -f "/etc/kernel/proxmox-boot-uuids") {
+       if (!$upgraded) {
+           log_skip("not yet upgraded, no need to check the presence of 
systemd-boot");
+           return;
+       }
+       if ( -f "/usr/share/doc/systemd-boot/changelog.Debian.gz") {
+           log_pass("bootloader packages installed correctly");
+           return;
+       }
+       log_warn(
+           "proxmox-boot-tool is used for bootloader configuration in uefi 
mode"
+           . " but the separate systemd-boot package is not installed,"
+           . " initializing new ESPs will not work until the package is 
installed"
+       );
+       return;
+    } elsif ( ! -f "/usr/share/doc/grub-efi-amd64/changelog.Debian.gz" ) {
+       log_warn(
+           "System booted in uefi mode but grub-efi-amd64 meta-package not 
installed,"
+           . " new grub versions will not be installed to /boot/efi!"
+           . " Install grub-efi-amd64."
+       );
+       return;
+    } else {
+       log_pass("bootloader packages installed correctly");
+    }
+}
+
+sub check_dkms_modules {
+    log_info("Check for dkms modules...");
+
+    my $count;
+    my $set_count = sub {
+       $count = scalar @_;
+    };
+
+    my $sig_pipe = $SIG{PIPE};
+    $SIG{PIPE} = "DEFAULT";
+    my $exit_code = eval {
+       run_command(['dkms', 'status', '-k', '`uname -r`'], outfunc => 
$set_count, noerr => 1)
+    };
+    $SIG{PIPE} = $sig_pipe;
+
+    if ($exit_code != 0) {
+       log_skip("could not get dkms status");
+    } elsif (!$count) {
+       log_pass("no dkms modules found");
+    } else {
+       log_warn("dkms modules found, this might cause issues during upgrade.");
+    }
+}
+
+sub check_misc {
+    print_header("MISCELLANEOUS CHECKS");
+    my $ssh_config = eval { PVE::Tools::file_get_contents('/root/.ssh/config') 
};
+    if (defined($ssh_config)) {
+       log_fail("Unsupported SSH Cipher configured for root in 
/root/.ssh/config: $1")
+           if $ssh_config =~ /^Ciphers .*(blowfish|arcfour|3des).*$/m;
+    } else {
+       log_skip("No SSH config file found.");
+    }
+
+    log_info("Checking common daemon services..");
+    $log_systemd_unit_state->('pveproxy.service');
+    $log_systemd_unit_state->('pvedaemon.service');
+    $log_systemd_unit_state->('pvescheduler.service');
+    $log_systemd_unit_state->('pvestatd.service');
+
+    check_time_sync();
+
+    my $root_free = PVE::Tools::df('/', 10);
+    log_warn("Less than 5 GB free space on root file system.")
+       if defined($root_free) && $root_free->{avail} < 5 * 1000*1000*1000;
+
+    log_info("Checking for running guests..");
+    my $running_guests = 0;
+
+    my $vms = eval { PVE::API2::Qemu->vmlist({ node => $nodename }) };
+    log_warn("Failed to retrieve information about this node's VMs - $@") if 
$@;
+    $running_guests += grep { $_->{status} eq 'running' } @$vms if 
defined($vms);
+
+    my $cts = eval { PVE::API2::LXC->vmlist({ node => $nodename }) };
+    log_warn("Failed to retrieve information about this node's CTs - $@") if 
$@;
+    $running_guests += grep { $_->{status} eq 'running' } @$cts if 
defined($cts);
+
+    if ($running_guests > 0) {
+       log_warn("$running_guests running guest(s) detected - consider 
migrating or stopping them.")
+    } else {
+       log_pass("no running guest detected.")
+    }
+
+    log_info("Checking if the local node's hostname '$nodename' is 
resolvable..");
+    my $local_ip = eval { PVE::Network::get_ip_from_hostname($nodename) };
+    if ($@) {
+       log_warn("Failed to resolve hostname '$nodename' to IP - $@");
+    } else {
+       log_info("Checking if resolved IP is configured on local node..");
+       my $cidr = Net::IP::ip_is_ipv6($local_ip) ? "$local_ip/128" : 
"$local_ip/32";
+       my $configured_ips = PVE::Network::get_local_ip_from_cidr($cidr);
+       my $ip_count = scalar(@$configured_ips);
+
+       if ($ip_count <= 0) {
+           log_fail("Resolved node IP '$local_ip' not configured or active for 
'$nodename'");
+       } elsif ($ip_count > 1) {
+           log_warn("Resolved node IP '$local_ip' active on multiple 
($ip_count) interfaces!");
+       } else {
+           log_pass("Resolved node IP '$local_ip' configured and active on 
single interface.");
+       }
+    }
+
+    log_info("Check node certificate's RSA key size");
+    my $certs = PVE::API2::Certificates->info({ node => $nodename });
+    my $certs_check = {
+       'rsaEncryption' => {
+           minsize => 2048,
+           name => 'RSA',
+       },
+       'id-ecPublicKey' => {
+           minsize => 224,
+           name => 'ECC',
+       },
+    };
+
+    my $certs_check_failed = 0;
+    for my $cert (@$certs) {
+       my ($type, $size, $fn) = $cert->@{qw(public-key-type public-key-bits 
filename)};
+
+       if (!defined($type) || !defined($size)) {
+           log_warn("'$fn': cannot check certificate, failed to get it's type 
or size!");
+       }
+
+       my $check = $certs_check->{$type};
+       if (!defined($check)) {
+           log_warn("'$fn': certificate's public key type '$type' unknown!");
+           next;
+       }
+
+       if ($size < $check->{minsize}) {
+           log_fail("'$fn', certificate's $check->{name} public key size is 
less than 2048 bit");
+           $certs_check_failed = 1;
+       } else {
+           log_pass("Certificate '$fn' passed Debian Busters (and newer) 
security level for TLS connections ($size >= 2048)");
+       }
+    }
+
+    check_backup_retention_settings();
+    check_cifs_credential_location();
+    check_custom_pool_roles();
+    check_lxcfs_fuse_version();
+    check_node_and_guest_configurations();
+    check_apt_repos();
+    check_nvidia_vgpu_service();
+    check_bootloader();
+    check_dkms_modules();
+}
+
+my sub colored_if {
+    my ($str, $color, $condition) = @_;
+    return "". ($condition ? colored($str, $color) : $str);
+}
+
+__PACKAGE__->register_method ({
+    name => 'checklist',
+    path => 'checklist',
+    method => 'GET',
+    description => 'Check (pre-/post-)upgrade conditions.',
+    parameters => {
+       additionalProperties => 0,
+       properties => {
+           full => {
+               description => 'perform additional, expensive checks.',
+               type => 'boolean',
+               optional => 1,
+               default => 0,
+           },
+       },
+    },
+    returns => { type => 'null' },
+    code => sub {
+       my ($param) = @_;
+
+       my $kernel_cli = PVE::Tools::file_get_contents('/proc/cmdline');
+       if ($kernel_cli =~ /systemd.unified_cgroup_hierarchy=0/){
+           $forced_legacy_cgroup = 1;
+       }
+
+       check_pve_packages();
+       check_cluster_corosync();
+       check_ceph();
+       check_storage_health();
+       check_misc();
+
+       if ($param->{full}) {
+           check_containers_cgroup_compat();
+       } else {
+           log_skip("NOTE: Expensive checks, like CT cgroupv2 compat, not 
performed without '--full' parameter");
+       }
+
+       print_header("SUMMARY");
+
+       my $total = 0;
+       $total += $_ for values %$counters;
+
+       print "TOTAL:    $total\n";
+       print colored("PASSED:   $counters->{pass}\n", 'green');
+       print "SKIPPED:  $counters->{skip}\n";
+       print colored_if("WARNINGS: $counters->{warn}\n", 'yellow', 
$counters->{warn} > 0);
+       print colored_if("FAILURES: $counters->{fail}\n", 'bold red', 
$counters->{fail} > 0);
+
+       if ($counters->{warn} > 0 || $counters->{fail} > 0) {
+           my $color = $counters->{fail} > 0 ? 'bold red' : 'yellow';
+           print colored("\nATTENTION: Please check the output for detailed 
information!\n", $color);
+           print colored("Try to solve the problems one at a time and then run 
this checklist tool again.\n", $color) if $counters->{fail} > 0;
+       }
+
+       return undef;
+    }});
+
+our $cmddef = [ __PACKAGE__, 'checklist', [], {}];
+
+1;
diff --git a/bin/Makefile b/bin/Makefile
index 82d5c633..cfe81a22 100644
--- a/bin/Makefile
+++ b/bin/Makefile
@@ -13,7 +13,7 @@ CLITOOLS = \
        pvenode \
        pvesh \
        pve7to8 \
-
+       pve8to9 \
 
 SCRIPTS =                      \
        $(SERVICES)             \
@@ -73,6 +73,16 @@ pve7to8.1:
        printf ".SH SYNOPSIS\npve7to8 [--full]\n" >> $@.tmp
        mv $@.tmp $@
 
+
+pve8to9.1:
+       printf ".TH PVE8TO9 1\n.SH NAME\npve8to9 \- Proxmox VE upgrade checker 
script for 8.4+ to current 9.x\n" > $@.tmp
+       printf ".SH DESCRIPTION\nThis tool will help you to detect common 
pitfalls and misconfguration\
+        before, and during the upgrade of a Proxmox VE system\n" >> $@.tmp
+       printf "Any failure must be addressed before the upgrade, and any 
warning must be addressed, \
+        or at least carefully evaluated, if a false-positive is suspected\n" 
>> $@.tmp
+       printf ".SH SYNOPSIS\npve8to9 [--full]\n" >> $@.tmp
+       mv $@.tmp $@
+
 pveversion.1.pod: pveversion
 pveupgrade.1.pod: pveupgrade
 pvereport.1.pod: pvereport
diff --git a/bin/pve8to9 b/bin/pve8to9
new file mode 100755
index 00000000..29f3c7ad
--- /dev/null
+++ b/bin/pve8to9
@@ -0,0 +1,8 @@
+#!/usr/bin/perl
+
+use strict;
+use warnings;
+
+use PVE::CLI::pve8to9;
+
+PVE::CLI::pve8to9->run_cli_handler();
-- 
2.39.5



_______________________________________________
pve-devel mailing list
pve-devel@lists.proxmox.com
https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-devel

Reply via email to