-----============= acceptance-small: conf-sanity ============----- Fri Apr 19 08:49:56 EDT 2024 excepting tests: 102 106 115 32newtarball 110 skipping tests SLOW=no: 45 69 106 111 114 Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:-f) Stopping client oleg233-client.virtnet /mnt/lustre opts:-f Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:-f) Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server Stopping /mnt/lustre-ost2 (opts:-f) on oleg233-server oleg233-server: oleg233-server.virtnet: executing set_hostid Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions libkmod: kmod_module_get_holders: could not open '/sys/module/pcc_cpufreq/holders': No such file or directory loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions Formatting mgs, mds, osts Format mds1: /dev/mapper/mds1_flakey Format mds2: /dev/mapper/mds2_flakey Format ost1: /dev/mapper/ost1_flakey Format ost2: /dev/mapper/ost2_flakey start mds service on oleg233-server Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/mds1_flakey Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/mds2_flakey Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/ost1_flakey Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid oleg233-server: oleg233-server.virtnet: executing wait_import_state FULL os[cp].lustre-OST0000-osc-MDT0000.ost_server_uuid 50 oleg233-server: os[cp].lustre-OST0000-osc-MDT0000.ost_server_uuid in FULL state after 0 sec oleg233-server: oleg233-server.virtnet: executing wait_import_state FULL os[cp].lustre-OST0000-osc-MDT0001.ost_server_uuid 50 oleg233-server: os[cp].lustre-OST0000-osc-MDT0001.ost_server_uuid in FULL state after 0 sec stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 76a: set permanent params with lctl across mounts ========================================================== 08:51:19 (1713531079) start mds service on oleg233-server Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre Change MGS params max_dirty_mb: 467 new_max_dirty_mb: 457 Waiting 90s for '457' Updated after 2s: want '457' got '457' 457 Check the value is stored after remount Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:) Stopping client oleg233-client.virtnet /mnt/lustre opts: Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:) Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server Checking servers environments Checking clients oleg233-client.virtnet environments Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions Setup mgs, mdt, osts Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 Starting ost2: -o localrecov /dev/mapper/ost2_flakey /mnt/lustre-ost2 seq.cli-lustre-OST0001-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/ost2_flakey Started lustre-OST0001 mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre Starting client oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre Started clients oleg233-client.virtnet: 192.168.202.133@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Using TIMEOUT=20 osc.lustre-OST0000-osc-ffff8800aaf3a800.idle_timeout=debug osc.lustre-OST0001-osc-ffff8800aaf3a800.idle_timeout=debug disable quota as required Change OST params client_cache_count: 128 new_client_cache_count: 256 Waiting 90s for '256' Updated after 3s: want '256' got '256' 256 Check the value is stored after remount Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:) Stopping client oleg233-client.virtnet /mnt/lustre opts: Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:) Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server Stopping /mnt/lustre-ost2 (opts:-f) on oleg233-server Checking servers environments Checking clients oleg233-client.virtnet environments Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions Setup mgs, mdt, osts Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 Starting ost2: -o localrecov /dev/mapper/ost2_flakey /mnt/lustre-ost2 seq.cli-lustre-OST0001-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0001 mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre Starting client oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre Started clients oleg233-client.virtnet: 192.168.202.133@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Using TIMEOUT=20 osc.lustre-OST0000-osc-ffff8800aa6ab000.idle_timeout=debug osc.lustre-OST0001-osc-ffff8800aa6ab000.idle_timeout=debug disable quota as required 256 Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:) Stopping client oleg233-client.virtnet /mnt/lustre opts: Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:) Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server Stopping /mnt/lustre-ost2 (opts:-f) on oleg233-server PASS 76a (165s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 76b: verify params log setup correctly ========================================================== 08:54:05 (1713531245) Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:) Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:) Checking servers environments Checking clients oleg233-client.virtnet environments Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions Setup mgs, mdt, osts Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 Starting ost2: -o localrecov /dev/mapper/ost2_flakey /mnt/lustre-ost2 seq.cli-lustre-OST0001-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0001 mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre Starting client oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre Started clients oleg233-client.virtnet: 192.168.202.133@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Using TIMEOUT=20 osc.lustre-OST0000-osc-ffff8800aa6ac000.idle_timeout=debug osc.lustre-OST0001-osc-ffff8800aa6ac000.idle_timeout=debug disable quota as required mgs.MGS.live.params= fsname: params flags: 0x20 gen: 2 Secure RPC Config Rules: imperative_recovery_state: state: startup nonir_clients: 0 nidtbl_version: 2 notify_duration_total: 0.000000000 notify_duation_max: 0.000000000 notify_count: 0 Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:) Stopping client oleg233-client.virtnet /mnt/lustre opts: Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:) Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server Stopping /mnt/lustre-ost2 (opts:-f) on oleg233-server PASS 76b (64s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 76c: verify changelog_mask is applied with lctl set_param -P ========================================================== 08:55:11 (1713531311) Checking servers environments Checking clients oleg233-client.virtnet environments Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions libkmod: kmod_module_get_holders: could not open '/sys/module/acpi_cpufreq/holders': No such file or directory libkmod: kmod_module_get_holders: could not open '/sys/module/pcc_cpufreq/holders': No such file or directory libkmod: kmod_module_get_holders: could not open '/sys/module/acpi_cpufreq/holders': No such file or directory loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions Setup mgs, mdt, osts Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 Starting ost2: -o localrecov /dev/mapper/ost2_flakey /mnt/lustre-ost2 seq.cli-lustre-OST0001-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0001 mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre Starting client oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre Started clients oleg233-client.virtnet: 192.168.202.133@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Using TIMEOUT=20 osc.lustre-OST0000-osc-ffff8800aaf3d800.idle_timeout=debug osc.lustre-OST0001-osc-ffff8800aaf3d800.idle_timeout=debug disable quota as required Change changelog_mask pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Check the value is stored after mds remount stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server start mds service on oleg233-server Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 20 sec oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:) Stopping client oleg233-client.virtnet /mnt/lustre opts: Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:) Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server Stopping /mnt/lustre-ost2 (opts:-f) on oleg233-server PASS 76c (105s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 76d: verify llite.*.xattr_cache can be set by 'lctl set_param -P' correctly ========================================================== 08:56:58 (1713531418) Checking servers environments Checking clients oleg233-client.virtnet environments Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions Setup mgs, mdt, osts Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 Starting ost2: -o localrecov /dev/mapper/ost2_flakey /mnt/lustre-ost2 seq.cli-lustre-OST0001-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0001 mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre Starting client oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre Started clients oleg233-client.virtnet: 192.168.202.133@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Using TIMEOUT=20 osc.lustre-OST0000-osc-ffff8800aaf3f800.idle_timeout=debug osc.lustre-OST0001-osc-ffff8800aaf3f800.idle_timeout=debug disable quota as required lctl set_param -P llite.*.xattr_cache=0 Waiting 90s for '0' Updated after 2s: want '0' got '0' Check llite.*.xattr_cache on client /mnt/lustre umount lustre on /mnt/lustre..... Stopping client oleg233-client.virtnet /mnt/lustre (opts:) mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre Check llite.*.xattr_cache on the new client /mnt/lustre2 mount lustre on /mnt/lustre2..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre2 umount lustre on /mnt/lustre2..... Stopping client oleg233-client.virtnet /mnt/lustre2 (opts:) Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:) Stopping client oleg233-client.virtnet /mnt/lustre opts: Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:) Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server Stopping /mnt/lustre-ost2 (opts:-f) on oleg233-server PASS 76d (63s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 77: comma-separated MGS NIDs and failover node NIDs ========================================================== 08:58:03 (1713531483) SKIP: conf-sanity test_77 mixed loopback and real device not working SKIP 77 (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 78: run resize2fs on MDT and OST filesystems ========================================================== 08:58:05 (1713531485) Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:-f) Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:-f) oleg233-server: oleg233-server.virtnet: executing set_hostid Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions oleg233-server: libkmod: kmod_module_get_holders: could not open '/sys/module/acpi_cpufreq/holders': No such file or directory Formatting mgs, mds, osts Format mds1: /dev/mapper/mds1_flakey Format ost1: /dev/mapper/ost1_flakey start mds service on oleg233-server Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/mds1_flakey Started lustre-MDT0000 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=131072 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/ost1_flakey Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre create test files UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 83240 1616 73832 3% /mnt/lustre[MDT:0] lustre-OST0000_UUID 124712 1388 110724 2% /mnt/lustre[OST:0] filesystem_summary: 124712 1388 110724 2% /mnt/lustre UUID Inodes IUsed IFree IUse% Mounted on lustre-MDT0000_UUID 72000 272 71728 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 45008 302 44706 1% /mnt/lustre[OST:0] filesystem_summary: 44978 272 44706 1% /mnt/lustre 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0373451 s, 28.1 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0374352 s, 28.0 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0370034 s, 28.3 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0351808 s, 29.8 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.033809 s, 31.0 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0441019 s, 23.8 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0357527 s, 29.3 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0653054 s, 16.1 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0356214 s, 29.4 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0383431 s, 27.3 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0349678 s, 30.0 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0407339 s, 25.7 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0358014 s, 29.3 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.040631 s, 25.8 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0380275 s, 27.6 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0429838 s, 24.4 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0390317 s, 26.9 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0444224 s, 23.6 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0404591 s, 25.9 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0443059 s, 23.7 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0362933 s, 28.9 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0376954 s, 27.8 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0396739 s, 26.4 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0441925 s, 23.7 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0402266 s, 26.1 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0399354 s, 26.3 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0503861 s, 20.8 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0378517 s, 27.7 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.038326 s, 27.4 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.039369 s, 26.6 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0437003 s, 24.0 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0484582 s, 21.6 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.038391 s, 27.3 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0399566 s, 26.2 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0390842 s, 26.8 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0379486 s, 27.6 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0409103 s, 25.6 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0392315 s, 26.7 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0435291 s, 24.1 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0331121 s, 31.7 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0401719 s, 26.1 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0383461 s, 27.3 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0419507 s, 25.0 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.045278 s, 23.2 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0360208 s, 29.1 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.036913 s, 28.4 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0374547 s, 28.0 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0643752 s, 16.3 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0395993 s, 26.5 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.038851 s, 27.0 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0386508 s, 27.1 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0379794 s, 27.6 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0430592 s, 24.4 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0405699 s, 25.8 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0370198 s, 28.3 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0410766 s, 25.5 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0349944 s, 30.0 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0391575 s, 26.8 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0386174 s, 27.2 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0381555 s, 27.5 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.04039 s, 26.0 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.038957 s, 26.9 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0418475 s, 25.1 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0352013 s, 29.8 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0386373 s, 27.1 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0369472 s, 28.4 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0383508 s, 27.3 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0385168 s, 27.2 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.06107 s, 17.2 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0388124 s, 27.0 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0347878 s, 30.1 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0384527 s, 27.3 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0356682 s, 29.4 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0350581 s, 29.9 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0345207 s, 30.4 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.037941 s, 27.6 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0360822 s, 29.1 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0370689 s, 28.3 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0382817 s, 27.4 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.040976 s, 25.6 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0379133 s, 27.7 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0588232 s, 17.8 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0340121 s, 30.8 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0358625 s, 29.2 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0405999 s, 25.8 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.39239 s, 2.7 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0434377 s, 24.1 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0372408 s, 28.2 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0553178 s, 19.0 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0372933 s, 28.1 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0400848 s, 26.2 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0377898 s, 27.7 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0379036 s, 27.7 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0344037 s, 30.5 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.059199 s, 17.7 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0352301 s, 29.8 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0356366 s, 29.4 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0506226 s, 20.7 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0384809 s, 27.2 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0887963 s, 11.8 MB/s umount lustre on /mnt/lustre..... Stopping client oleg233-client.virtnet /mnt/lustre (opts:) stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server unloading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing unload_modules_local modules unloaded. e2fsck -d -v -t -t -f -y /dev/mapper/mds1_flakey -m8 oleg233-server: e2fsck 1.46.2.wc5 (26-Mar-2022) oleg233-server: Use max possible thread num: 1 instead Pass 1: Checking inodes, blocks, and sizes [Thread 0] Scan group range [0, 3) [Thread 0] jumping to group 0 [Thread 0] e2fsck_pass1_run:2564: increase inode 81 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 82 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 83 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 84 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 85 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 86 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 87 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 88 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 89 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 90 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 91 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 92 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 93 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 94 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 95 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 96 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 97 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 98 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 99 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 100 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 101 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 102 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 103 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 104 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 105 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 106 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 107 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 108 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 109 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 110 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 111 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 112 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 113 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 114 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 115 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 116 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 117 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 118 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 119 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 120 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 121 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 122 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 123 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 124 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 125 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 126 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 127 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 128 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 129 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 130 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 131 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 132 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 133 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 134 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 135 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 136 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 137 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 138 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 139 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 140 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 141 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 142 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 143 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 144 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 145 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 146 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 147 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 148 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 149 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 150 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 151 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 152 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 153 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 154 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 155 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 156 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 157 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 158 badness 0 to 2 for 10084 [Thread 0] group 1 finished [Thread 0] e2fsck_pass1_run:2564: increase inode 24033 badness 0 to 2 for 10084 [Thread 0] group 2 finished [Thread 0] e2fsck_pass1_run:2564: increase inode 48044 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 48045 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 48046 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 48047 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 48048 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 48050 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 48051 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 48052 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 48053 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 48054 badness 0 to 2 for 10084 [Thread 0] group 3 finished [Thread 0] Pass 1: Memory used: 272k/0k (141k/132k), time: 0.00/ 0.00/ 0.00 [Thread 0] Pass 1: I/O read: 1MB, write: 0MB, rate: 283.61MB/s [Thread 0] Scanned group range [0, 3), inodes 373 Pass 2: Checking directory structure Pass 2: Memory used: 272k/0k (95k/178k), time: 0.00/ 0.00/ 0.00 Pass 2: I/O read: 1MB, write: 0MB, rate: 297.71MB/s Pass 3: Checking directory connectivity Peak memory: Memory used: 272k/0k (95k/178k), time: 0.01/ 0.01/ 0.00 Pass 3A: Memory used: 272k/0k (95k/178k), time: 0.00/ 0.00/ 0.00 Pass 3A: I/O read: 0MB, write: 0MB, rate: 0.00MB/s Pass 3: Memory used: 272k/0k (93k/180k), time: 0.00/ 0.00/ 0.00 Pass 3: I/O read: 1MB, write: 0MB, rate: 5988.02MB/s Pass 4: Checking reference counts Pass 4: Memory used: 272k/0k (67k/206k), time: 0.00/ 0.00/ 0.00 Pass 4: I/O read: 0MB, write: 0MB, rate: 0.00MB/s Pass 5: Checking group summary information Pass 5: Memory used: 272k/0k (67k/206k), time: 0.00/ 0.00/ 0.00 Pass 5: I/O read: 1MB, write: 0MB, rate: 282.01MB/s 372 inodes used (0.52%, out of 72000) 4 non-contiguous files (1.1%) 0 non-contiguous directories (0.0%) # of inodes with ind/dind/tind blocks: 0/0/0 22546 blocks used (50.10%, out of 45000) 0 bad blocks 1 large file 244 regular files 118 directories 0 character device files 0 block device files 0 fifos 0 links 0 symbolic links (0 fast symbolic links) 0 sockets ------------ 362 files Memory used: 272k/0k (66k/207k), time: 0.02/ 0.01/ 0.01 I/O read: 1MB, write: 1MB, rate: 41.16MB/s e2fsck -d -v -t -t -f -y /dev/mapper/ost1_flakey -m8 oleg233-server: e2fsck 1.46.2.wc5 (26-Mar-2022) oleg233-server: Use max possible thread num: 1 instead Pass 1: Checking inodes, blocks, and sizes [Thread 0] Scan group range [0, 2) [Thread 0] jumping to group 0 [Thread 0] e2fsck_pass1_run:2564: increase inode 81 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 82 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 89 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 90 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 91 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 92 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 93 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 94 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 95 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 96 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 97 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 98 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 99 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 100 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 101 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 102 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 103 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 104 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 105 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 106 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 107 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 140 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 142 badness 0 to 2 for 10084 [Thread 0] group 1 finished [Thread 0] group 2 finished [Thread 0] Pass 1: Memory used: 264k/0k (132k/133k), time: 0.01/ 0.00/ 0.01 [Thread 0] Pass 1: I/O read: 5MB, write: 0MB, rate: 553.96MB/s [Thread 0] Scanned group range [0, 2), inodes 398 Pass 2: Checking directory structure Pass 2: Memory used: 264k/0k (87k/178k), time: 0.00/ 0.00/ 0.00 Pass 2: I/O read: 1MB, write: 0MB, rate: 338.18MB/s Pass 3: Checking directory connectivity Peak memory: Memory used: 264k/0k (92k/173k), time: 0.02/ 0.00/ 0.01 Pass 3A: Memory used: 264k/0k (92k/173k), time: 0.00/ 0.00/ 0.00 Pass 3A: I/O read: 0MB, write: 0MB, rate: 0.00MB/s Pass 3: Memory used: 264k/0k (84k/181k), time: 0.00/ 0.00/ 0.00 Pass 3: I/O read: 1MB, write: 0MB, rate: 3663.00MB/s Pass 4: Checking reference counts Pass 4: Memory used: 264k/0k (65k/200k), time: 0.00/ 0.00/ 0.00 Pass 4: I/O read: 0MB, write: 0MB, rate: 0.00MB/s Pass 5: Checking group summary information Pass 5: Memory used: 264k/0k (65k/200k), time: 0.00/ 0.00/ 0.00 Pass 5: I/O read: 1MB, write: 0MB, rate: 274.12MB/s 398 inodes used (0.88%, out of 45008) 2 non-contiguous files (0.5%) 0 non-contiguous directories (0.0%) # of inodes with ind/dind/tind blocks: 0/0/0 Extent depth histogram: 392 37721 blocks used (83.82%, out of 45000) 0 bad blocks 1 large file 216 regular files 172 directories 0 character device files 0 block device files 0 fifos 0 links 0 symbolic links (0 fast symbolic links) 0 sockets ------------ 388 files Memory used: 264k/0k (64k/201k), time: 0.03/ 0.01/ 0.01 I/O read: 2MB, write: 1MB, rate: 77.78MB/s oleg233-server: resize2fs 1.46.2.wc5 (26-Mar-2022) Resizing the filesystem on /dev/mapper/mds1_flakey to 640000 (4k) blocks. The filesystem on /dev/mapper/mds1_flakey is now 640000 (4k) blocks long. oleg233-server: resize2fs 1.46.2.wc5 (26-Mar-2022) Resizing the filesystem on /dev/mapper/ost1_flakey to 1048576 (4k) blocks. The filesystem on /dev/mapper/ost1_flakey is now 1048576 (4k) blocks long. e2fsck -d -v -t -t -f -y /dev/mapper/mds1_flakey -m8 oleg233-server: e2fsck 1.46.2.wc5 (26-Mar-2022) oleg233-server: Use max possible thread num: 2 instead Pass 1: Checking inodes, blocks, and sizes [Thread 0] Scan group range [0, 16) [Thread 1] Scan group range [16, 33) [Thread 1] jumping to group 16 [Thread 0] jumping to group 0 [Thread 1] group 17 finished [Thread 1] group 18 finished [Thread 1] group 19 finished [Thread 1] group 20 finished [Thread 1] group 21 finished [Thread 1] group 22 finished [Thread 1] group 23 finished [Thread 1] group 24 finished [Thread 1] group 25 finished [Thread 1] group 26 finished [Thread 1] group 27 finished [Thread 1] group 28 finished [Thread 1] group 29 finished [Thread 1] group 30 finished [Thread 1] group 31 finished [Thread 1] group 32 finished [Thread 1] group 33 finished [Thread 1] Pass 1: Memory used: 632k/0k (378k/255k), time: 0.00/ 0.00/ 0.00 [Thread 1] Pass 1: I/O read: 1MB, write: 0MB, rate: 1239.16MB/s [Thread 1] Scanned group range [16, 33), inodes 0 [Thread 0] e2fsck_pass1_run:2564: increase inode 81 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 82 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 83 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 84 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 85 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 86 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 87 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 88 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 89 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 90 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 91 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 92 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 93 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 94 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 95 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 96 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 97 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 98 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 99 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 100 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 101 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 102 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 103 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 104 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 105 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 106 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 107 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 108 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 109 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 110 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 111 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 112 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 113 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 114 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 115 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 116 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 117 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 118 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 119 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 120 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 121 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 122 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 123 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 124 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 125 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 126 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 127 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 128 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 129 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 130 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 131 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 132 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 133 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 134 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 135 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 136 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 137 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 138 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 139 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 140 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 141 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 142 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 143 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 144 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 145 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 146 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 147 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 148 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 149 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 150 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 151 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 152 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 153 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 154 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 155 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 156 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 157 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 158 badness 0 to 2 for 10084 [Thread 0] group 1 finished [Thread 0] e2fsck_pass1_run:2564: increase inode 24033 badness 0 to 2 for 10084 [Thread 0] group 2 finished [Thread 0] e2fsck_pass1_run:2564: increase inode 48044 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 48045 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 48046 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 48047 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 48048 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 48050 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 48051 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 48052 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 48053 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 48054 badness 0 to 2 for 10084 [Thread 0] group 3 finished [Thread 0] group 4 finished [Thread 0] group 5 finished [Thread 0] group 6 finished [Thread 0] group 7 finished [Thread 0] group 8 finished [Thread 0] group 9 finished [Thread 0] group 10 finished [Thread 0] group 11 finished [Thread 0] group 12 finished [Thread 0] group 13 finished [Thread 0] group 14 finished [Thread 0] group 15 finished [Thread 0] group 16 finished [Thread 0] Pass 1: Memory used: 688k/0k (355k/334k), time: 0.00/ 0.00/ 0.00 [Thread 0] Pass 1: I/O read: 1MB, write: 0MB, rate: 262.26MB/s [Thread 0] Scanned group range [0, 16), inodes 373 Pass 2: Checking directory structure Pass 2: Memory used: 632k/0k (200k/433k), time: 0.00/ 0.00/ 0.00 Pass 2: I/O read: 1MB, write: 0MB, rate: 322.79MB/s Pass 3: Checking directory connectivity Peak memory: Memory used: 632k/0k (200k/433k), time: 0.04/ 0.04/ 0.01 Pass 3A: Memory used: 632k/0k (200k/433k), time: 0.00/ 0.00/ 0.00 Pass 3A: I/O read: 0MB, write: 0MB, rate: 0.00MB/s Pass 3: Memory used: 632k/0k (198k/435k), time: 0.00/ 0.00/ 0.00 Pass 3: I/O read: 1MB, write: 0MB, rate: 9523.81MB/s Pass 4: Checking reference counts Pass 4: Memory used: 632k/0k (72k/561k), time: 0.02/ 0.02/ 0.00 Pass 4: I/O read: 0MB, write: 0MB, rate: 0.00MB/s Pass 5: Checking group summary information Pass 5: Memory used: 632k/0k (70k/563k), time: 0.00/ 0.00/ 0.00 Pass 5: I/O read: 1MB, write: 0MB, rate: 257.20MB/s 372 inodes used (0.05%, out of 792000) 4 non-contiguous files (1.1%) 0 non-contiguous directories (0.0%) # of inodes with ind/dind/tind blocks: 0/0/0 202726 blocks used (31.68%, out of 640000) 0 bad blocks 1 large file 244 regular files 118 directories 0 character device files 0 block device files 0 fifos 0 links 0 symbolic links (0 fast symbolic links) 0 sockets ------------ 362 files Memory used: 632k/0k (69k/564k), time: 0.07/ 0.06/ 0.01 I/O read: 1MB, write: 1MB, rate: 13.95MB/s e2fsck -d -v -t -t -f -y /dev/mapper/ost1_flakey -m8 oleg233-server: e2fsck 1.46.2.wc5 (26-Mar-2022) oleg233-server: Use max possible thread num: 1 instead Pass 1: Checking inodes, blocks, and sizes [Thread 0] Scan group range [0, 32) [Thread 0] jumping to group 0 [Thread 0] e2fsck_pass1_run:2564: increase inode 81 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 82 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 89 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 90 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 91 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 92 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 93 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 94 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 95 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 96 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 97 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 98 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 99 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 100 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 101 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 102 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 103 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 104 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 105 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 106 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 107 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 140 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 142 badness 0 to 2 for 10084 [Thread 0] group 1 finished [Thread 0] group 2 finished [Thread 0] group 3 finished [Thread 0] group 4 finished [Thread 0] group 5 finished [Thread 0] group 6 finished [Thread 0] group 7 finished [Thread 0] group 8 finished [Thread 0] group 9 finished [Thread 0] group 10 finished [Thread 0] group 11 finished [Thread 0] group 12 finished [Thread 0] group 13 finished [Thread 0] group 14 finished [Thread 0] group 15 finished [Thread 0] group 16 finished [Thread 0] group 17 finished [Thread 0] group 18 finished [Thread 0] group 19 finished [Thread 0] group 20 finished [Thread 0] group 21 finished [Thread 0] group 22 finished [Thread 0] group 23 finished [Thread 0] group 24 finished [Thread 0] group 25 finished [Thread 0] group 26 finished [Thread 0] group 27 finished [Thread 0] group 28 finished [Thread 0] group 29 finished [Thread 0] group 30 finished [Thread 0] group 31 finished [Thread 0] group 32 finished [Thread 0] Pass 1: Memory used: 468k/0k (344k/125k), time: 0.01/ 0.00/ 0.01 [Thread 0] Pass 1: I/O read: 5MB, write: 0MB, rate: 651.30MB/s [Thread 0] Scanned group range [0, 32), inodes 398 Pass 2: Checking directory structure Pass 2: Memory used: 680k/0k (298k/383k), time: 0.00/ 0.00/ 0.00 Pass 2: I/O read: 1MB, write: 0MB, rate: 353.61MB/s Pass 3: Checking directory connectivity Peak memory: Memory used: 680k/0k (299k/382k), time: 0.03/ 0.02/ 0.01 Pass 3A: Memory used: 680k/0k (299k/382k), time: 0.00/ 0.00/ 0.00 Pass 3A: I/O read: 0MB, write: 0MB, rate: 0.00MB/s Pass 3: Memory used: 680k/0k (296k/385k), time: 0.00/ 0.00/ 0.00 Pass 3: I/O read: 1MB, write: 0MB, rate: 8474.58MB/s Pass 4: Checking reference counts Pass 4: Memory used: 564k/0k (66k/499k), time: 0.02/ 0.02/ 0.00 Pass 4: I/O read: 0MB, write: 0MB, rate: 0.00MB/s Pass 5: Checking group summary information Pass 5: Memory used: 564k/0k (65k/500k), time: 0.00/ 0.00/ 0.00 Pass 5: I/O read: 1MB, write: 0MB, rate: 231.80MB/s 398 inodes used (0.06%, out of 720128) 2 non-contiguous files (0.5%) 0 non-contiguous directories (0.0%) # of inodes with ind/dind/tind blocks: 0/0/0 Extent depth histogram: 392 128315 blocks used (12.24%, out of 1048576) 0 bad blocks 1 large file 216 regular files 172 directories 0 character device files 0 block device files 0 fifos 0 links 0 symbolic links (0 fast symbolic links) 0 sockets ------------ 388 files Memory used: 564k/0k (64k/501k), time: 0.05/ 0.04/ 0.01 I/O read: 2MB, write: 1MB, rate: 40.44MB/s start mds service on oleg233-server Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions libkmod: kmod_module_get_holders: could not open '/sys/module/acpi_cpufreq/holders': No such file or directory ../libcfs/libcfs/libcfs options: 'cpu_npartitions=2' ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' quota/lquota options: 'hash_lqs_cur_bits=3' loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions oleg233-server: libkmod: kmod_module_get_holders: could not open '/sys/module/pcc_cpufreq/holders': No such file or directory oleg233-server: ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' oleg233-server: quota/lquota options: 'hash_lqs_cur_bits=3' Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=131072 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre check files after expanding the MDT and OST filesystems /mnt/lustre/d78.conf-sanity/f78.conf-sanity-1 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-1 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-2 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-2 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-3 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-3 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-4 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-4 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-5 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-5 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-6 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-6 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-7 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-7 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-8 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-8 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-9 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-9 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-10 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-10 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-11 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-11 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-12 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-12 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-13 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-13 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-14 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-14 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-15 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-15 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-16 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-16 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-17 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-17 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-18 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-18 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-19 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-19 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-20 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-20 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-21 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-21 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-22 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-22 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-23 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-23 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-24 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-24 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-25 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-25 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-26 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-26 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-27 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-27 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-28 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-28 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-29 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-29 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-30 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-30 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-31 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-31 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-32 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-32 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-33 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-33 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-34 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-34 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-35 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-35 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-36 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-36 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-37 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-37 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-38 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-38 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-39 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-39 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-40 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-40 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-41 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-41 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-42 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-42 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-43 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-43 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-44 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-44 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-45 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-45 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-46 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-46 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-47 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-47 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-48 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-48 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-49 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-49 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-50 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-50 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-51 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-51 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-52 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-52 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-53 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-53 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-54 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-54 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-55 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-55 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-56 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-56 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-57 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-57 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-58 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-58 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-59 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-59 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-60 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-60 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-61 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-61 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-62 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-62 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-63 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-63 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-64 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-64 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-65 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-65 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-66 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-66 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-67 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-67 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-68 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-68 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-69 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-69 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-70 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-70 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-71 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-71 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-72 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-72 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-73 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-73 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-74 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-74 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-75 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-75 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-76 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-76 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-77 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-77 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-78 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-78 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-79 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-79 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-80 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-80 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-81 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-81 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-82 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-82 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-83 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-83 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-84 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-84 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-85 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-85 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-86 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-86 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-87 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-87 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-88 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-88 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-89 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-89 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-90 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-90 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-91 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-91 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-92 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-92 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-93 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-93 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-94 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-94 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-95 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-95 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-96 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-96 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-97 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-97 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-98 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-98 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-99 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-99 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-100 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-100 has size 1048576 OK create more files after expanding the MDT and OST filesystems 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0431989 s, 24.3 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.046952 s, 22.3 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0442959 s, 23.7 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0459162 s, 22.8 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0409742 s, 25.6 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0327845 s, 32.0 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0353035 s, 29.7 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0533975 s, 19.6 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0377368 s, 27.8 MB/s 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0341508 s, 30.7 MB/s umount lustre on /mnt/lustre..... Stopping client oleg233-client.virtnet /mnt/lustre (opts:) stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server unloading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing unload_modules_local modules unloaded. e2fsck -d -v -t -t -f -y /dev/mapper/mds1_flakey -m8 oleg233-server: e2fsck 1.46.2.wc5 (26-Mar-2022) oleg233-server: Use max possible thread num: 2 instead Pass 1: Checking inodes, blocks, and sizes [Thread 0] Scan group range [0, 16) [Thread 1] Scan group range [16, 33) [Thread 0] jumping to group 0 [Thread 1] jumping to group 16 [Thread 1] group 17 finished [Thread 1] group 18 finished [Thread 1] group 19 finished [Thread 1] group 20 finished [Thread 1] group 21 finished [Thread 1] group 22 finished [Thread 1] group 23 finished [Thread 1] group 24 finished [Thread 1] group 25 finished [Thread 1] group 26 finished [Thread 1] group 27 finished [Thread 1] group 28 finished [Thread 1] group 29 finished [Thread 1] group 30 finished [Thread 1] group 31 finished [Thread 1] group 32 finished [Thread 1] group 33 finished [Thread 1] Pass 1: Memory used: 632k/0k (378k/255k), time: 0.00/ 0.00/ 0.00 [Thread 1] Pass 1: I/O read: 1MB, write: 0MB, rate: 1569.86MB/s [Thread 1] Scanned group range [16, 33), inodes 0 [Thread 0] e2fsck_pass1_run:2564: increase inode 81 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 82 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 83 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 84 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 85 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 86 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 87 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 88 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 89 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 90 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 91 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 92 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 93 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 94 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 95 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 96 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 97 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 98 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 99 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 100 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 101 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 102 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 103 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 104 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 105 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 106 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 107 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 108 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 109 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 110 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 111 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 112 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 113 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 114 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 115 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 116 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 117 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 118 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 119 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 120 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 121 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 122 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 123 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 124 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 125 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 126 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 127 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 128 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 129 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 130 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 131 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 132 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 133 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 134 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 135 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 136 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 137 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 138 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 139 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 140 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 141 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 142 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 143 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 144 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 145 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 146 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 147 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 148 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 149 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 150 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 151 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 152 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 153 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 154 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 155 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 156 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 157 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 158 badness 0 to 2 for 10084 [Thread 0] group 1 finished [Thread 0] e2fsck_pass1_run:2564: increase inode 24033 badness 0 to 2 for 10084 [Thread 0] group 2 finished [Thread 0] e2fsck_pass1_run:2564: increase inode 48044 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 48045 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 48046 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 48047 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 48048 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 48050 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 48051 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 48052 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 48053 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 48054 badness 0 to 2 for 10084 [Thread 0] group 3 finished [Thread 0] group 4 finished [Thread 0] group 5 finished [Thread 0] group 6 finished [Thread 0] group 7 finished [Thread 0] group 8 finished [Thread 0] group 9 finished [Thread 0] group 10 finished [Thread 0] group 11 finished [Thread 0] group 12 finished [Thread 0] group 13 finished [Thread 0] group 14 finished [Thread 0] group 15 finished [Thread 0] group 16 finished [Thread 0] Pass 1: Memory used: 688k/0k (355k/334k), time: 0.01/ 0.00/ 0.00 [Thread 0] Pass 1: I/O read: 1MB, write: 0MB, rate: 197.51MB/s [Thread 0] Scanned group range [0, 16), inodes 383 Pass 2: Checking directory structure Pass 2: Memory used: 632k/0k (200k/433k), time: 0.00/ 0.00/ 0.00 Pass 2: I/O read: 1MB, write: 0MB, rate: 314.07MB/s Pass 3: Checking directory connectivity Peak memory: Memory used: 632k/0k (200k/433k), time: 0.04/ 0.03/ 0.01 Pass 3A: Memory used: 632k/0k (200k/433k), time: 0.00/ 0.00/ 0.00 Pass 3A: I/O read: 0MB, write: 0MB, rate: 0.00MB/s Pass 3: Memory used: 632k/0k (198k/435k), time: 0.00/ 0.00/ 0.00 Pass 3: I/O read: 1MB, write: 0MB, rate: 5347.59MB/s Pass 4: Checking reference counts Pass 4: Memory used: 632k/0k (72k/561k), time: 0.02/ 0.02/ 0.00 Pass 4: I/O read: 0MB, write: 0MB, rate: 0.00MB/s Pass 5: Checking group summary information Pass 5: Memory used: 632k/0k (70k/563k), time: 0.00/ 0.00/ 0.00 Pass 5: I/O read: 1MB, write: 0MB, rate: 214.59MB/s 382 inodes used (0.05%, out of 792000) 4 non-contiguous files (1.0%) 0 non-contiguous directories (0.0%) # of inodes with ind/dind/tind blocks: 0/0/0 202726 blocks used (31.68%, out of 640000) 0 bad blocks 1 large file 254 regular files 118 directories 0 character device files 0 block device files 0 fifos 0 links 0 symbolic links (0 fast symbolic links) 0 sockets ------------ 372 files Memory used: 632k/0k (69k/564k), time: 0.07/ 0.06/ 0.01 I/O read: 1MB, write: 1MB, rate: 14.68MB/s e2fsck -d -v -t -t -f -y /dev/mapper/ost1_flakey -m8 oleg233-server: e2fsck 1.46.2.wc5 (26-Mar-2022) oleg233-server: Use max possible thread num: 1 instead Pass 1: Checking inodes, blocks, and sizes [Thread 0] Scan group range [0, 32) [Thread 0] jumping to group 0 [Thread 0] e2fsck_pass1_run:2564: increase inode 81 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 82 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 89 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 90 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 91 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 92 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 93 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 94 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 95 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 96 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 97 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 98 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 99 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 100 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 101 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 102 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 103 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 104 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 105 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 106 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 107 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 140 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 142 badness 0 to 2 for 10084 [Thread 0] group 1 finished [Thread 0] group 2 finished [Thread 0] group 3 finished [Thread 0] group 4 finished [Thread 0] group 5 finished [Thread 0] group 6 finished [Thread 0] group 7 finished [Thread 0] group 8 finished [Thread 0] group 9 finished [Thread 0] group 10 finished [Thread 0] group 11 finished [Thread 0] group 12 finished [Thread 0] group 13 finished [Thread 0] group 14 finished [Thread 0] group 15 finished [Thread 0] group 16 finished [Thread 0] group 17 finished [Thread 0] group 18 finished [Thread 0] group 19 finished [Thread 0] group 20 finished [Thread 0] group 21 finished [Thread 0] group 22 finished [Thread 0] group 23 finished [Thread 0] group 24 finished [Thread 0] group 25 finished [Thread 0] group 26 finished [Thread 0] group 27 finished [Thread 0] group 28 finished [Thread 0] group 29 finished [Thread 0] group 30 finished [Thread 0] group 31 finished [Thread 0] group 32 finished [Thread 0] Pass 1: Memory used: 472k/0k (345k/128k), time: 0.01/ 0.00/ 0.00 [Thread 0] Pass 1: I/O read: 5MB, write: 0MB, rate: 631.79MB/s [Thread 0] Scanned group range [0, 32), inodes 402 Pass 2: Checking directory structure Pass 2: Memory used: 684k/0k (299k/386k), time: 0.00/ 0.00/ 0.00 Pass 2: I/O read: 1MB, write: 0MB, rate: 413.91MB/s Pass 3: Checking directory connectivity Peak memory: Memory used: 684k/0k (299k/386k), time: 0.03/ 0.02/ 0.01 Pass 3A: Memory used: 684k/0k (299k/386k), time: 0.00/ 0.00/ 0.00 Pass 3A: I/O read: 0MB, write: 0MB, rate: 0.00MB/s Pass 3: Memory used: 684k/0k (296k/389k), time: 0.00/ 0.00/ 0.00 Pass 3: I/O read: 1MB, write: 0MB, rate: 6849.31MB/s Pass 4: Checking reference counts Pass 4: Memory used: 568k/0k (67k/502k), time: 0.02/ 0.02/ 0.00 Pass 4: I/O read: 0MB, write: 0MB, rate: 0.00MB/s Pass 5: Checking group summary information Pass 5: Memory used: 568k/0k (65k/504k), time: 0.01/ 0.01/ 0.00 Pass 5: I/O read: 1MB, write: 0MB, rate: 181.62MB/s 402 inodes used (0.06%, out of 720128) 4 non-contiguous files (1.0%) 0 non-contiguous directories (0.0%) # of inodes with ind/dind/tind blocks: 0/0/0 Extent depth histogram: 396 130875 blocks used (12.48%, out of 1048576) 0 bad blocks 1 large file 220 regular files 172 directories 0 character device files 0 block device files 0 fifos 0 links 0 symbolic links (0 fast symbolic links) 0 sockets ------------ 392 files Memory used: 568k/0k (64k/505k), time: 0.05/ 0.04/ 0.01 I/O read: 2MB, write: 1MB, rate: 41.08MB/s oleg233-server: resize2fs 1.46.2.wc5 (26-Mar-2022) Resizing the filesystem on /dev/mapper/mds1_flakey to 377837 (4k) blocks. The filesystem on /dev/mapper/mds1_flakey is now 377837 (4k) blocks long. oleg233-server: resize2fs 1.46.2.wc5 (26-Mar-2022) Resizing the filesystem on /dev/mapper/ost1_flakey to 591846 (4k) blocks. The filesystem on /dev/mapper/ost1_flakey is now 589824 (4k) blocks long. e2fsck -d -v -t -t -f -y /dev/mapper/mds1_flakey -m8 oleg233-server: e2fsck 1.46.2.wc5 (26-Mar-2022) oleg233-server: Use max possible thread num: 1 instead Pass 1: Checking inodes, blocks, and sizes [Thread 0] Scan group range [0, 20) [Thread 0] jumping to group 0 [Thread 0] e2fsck_pass1_run:2564: increase inode 81 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 82 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 83 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 84 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 85 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 86 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 87 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 88 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 89 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 90 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 91 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 92 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 93 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 94 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 95 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 96 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 97 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 98 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 99 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 100 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 101 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 102 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 103 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 104 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 105 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 106 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 107 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 108 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 109 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 110 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 111 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 112 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 113 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 114 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 115 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 116 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 117 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 118 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 119 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 120 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 121 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 122 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 123 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 124 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 125 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 126 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 127 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 128 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 129 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 130 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 131 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 132 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 133 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 134 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 135 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 136 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 137 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 138 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 139 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 140 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 141 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 142 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 143 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 144 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 145 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 146 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 147 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 148 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 149 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 150 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 151 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 152 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 153 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 154 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 155 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 156 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 157 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 158 badness 0 to 2 for 10084 [Thread 0] group 1 finished [Thread 0] e2fsck_pass1_run:2564: increase inode 24033 badness 0 to 2 for 10084 [Thread 0] group 2 finished [Thread 0] e2fsck_pass1_run:2564: increase inode 48044 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 48045 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 48046 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 48047 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 48048 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 48050 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 48051 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 48052 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 48053 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 48054 badness 0 to 2 for 10084 [Thread 0] group 3 finished [Thread 0] group 4 finished [Thread 0] group 5 finished [Thread 0] group 6 finished [Thread 0] group 7 finished [Thread 0] group 8 finished [Thread 0] group 9 finished [Thread 0] group 10 finished [Thread 0] group 11 finished [Thread 0] group 12 finished [Thread 0] group 13 finished [Thread 0] group 14 finished [Thread 0] group 15 finished [Thread 0] group 16 finished [Thread 0] group 17 finished [Thread 0] group 18 finished [Thread 0] group 19 finished [Thread 0] group 20 finished [Thread 0] Pass 1: Memory used: 400k/0k (270k/131k), time: 0.00/ 0.00/ 0.00 [Thread 0] Pass 1: I/O read: 1MB, write: 0MB, rate: 309.02MB/s [Thread 0] Scanned group range [0, 20), inodes 383 Pass 2: Checking directory structure Pass 2: Memory used: 576k/0k (224k/353k), time: 0.00/ 0.00/ 0.00 Pass 2: I/O read: 1MB, write: 0MB, rate: 296.56MB/s Pass 3: Checking directory connectivity Peak memory: Memory used: 576k/0k (224k/353k), time: 0.03/ 0.02/ 0.00 Pass 3A: Memory used: 576k/0k (224k/353k), time: 0.00/ 0.00/ 0.00 Pass 3A: I/O read: 0MB, write: 0MB, rate: 0.00MB/s Pass 3: Memory used: 576k/0k (222k/355k), time: 0.00/ 0.00/ 0.00 Pass 3: I/O read: 1MB, write: 0MB, rate: 10989.01MB/s Pass 4: Checking reference counts Pass 4: Memory used: 500k/0k (68k/433k), time: 0.01/ 0.01/ 0.00 Pass 4: I/O read: 0MB, write: 0MB, rate: 0.00MB/s Pass 5: Checking group summary information Pass 5: Memory used: 500k/0k (67k/434k), time: 0.00/ 0.00/ 0.00 Pass 5: I/O read: 1MB, write: 0MB, rate: 344.00MB/s 382 inodes used (0.08%, out of 480000) 4 non-contiguous files (1.0%) 0 non-contiguous directories (0.0%) # of inodes with ind/dind/tind blocks: 0/0/0 124660 blocks used (32.99%, out of 377837) 0 bad blocks 1 large file 254 regular files 118 directories 0 character device files 0 block device files 0 fifos 0 links 0 symbolic links (0 fast symbolic links) 0 sockets ------------ 372 files Memory used: 500k/0k (66k/435k), time: 0.04/ 0.04/ 0.00 I/O read: 1MB, write: 1MB, rate: 23.28MB/s e2fsck -d -v -t -t -f -y /dev/mapper/ost1_flakey -m8 oleg233-server: e2fsck 1.46.2.wc5 (26-Mar-2022) oleg233-server: Use max possible thread num: 1 instead Pass 1: Checking inodes, blocks, and sizes [Thread 0] Scan group range [0, 18) [Thread 0] jumping to group 0 [Thread 0] e2fsck_pass1_run:2564: increase inode 81 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 82 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 89 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 90 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 91 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 92 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 93 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 94 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 95 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 96 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 97 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 98 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 99 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 100 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 101 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 102 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 103 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 104 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 105 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 106 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 107 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 140 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 142 badness 0 to 2 for 10084 [Thread 0] group 1 finished [Thread 0] group 2 finished [Thread 0] group 3 finished [Thread 0] group 4 finished [Thread 0] group 5 finished [Thread 0] group 6 finished [Thread 0] group 7 finished [Thread 0] group 8 finished [Thread 0] group 9 finished [Thread 0] group 10 finished [Thread 0] group 11 finished [Thread 0] group 12 finished [Thread 0] group 13 finished [Thread 0] group 14 finished [Thread 0] group 15 finished [Thread 0] group 16 finished [Thread 0] group 17 finished [Thread 0] group 18 finished [Thread 0] Pass 1: Memory used: 372k/0k (246k/127k), time: 0.01/ 0.00/ 0.00 [Thread 0] Pass 1: I/O read: 5MB, write: 0MB, rate: 670.51MB/s [Thread 0] Scanned group range [0, 18), inodes 402 Pass 2: Checking directory structure Pass 2: Memory used: 532k/0k (200k/333k), time: 0.00/ 0.00/ 0.00 Pass 2: I/O read: 1MB, write: 0MB, rate: 362.06MB/s Pass 3: Checking directory connectivity Peak memory: Memory used: 532k/0k (200k/333k), time: 0.02/ 0.01/ 0.01 Pass 3A: Memory used: 532k/0k (200k/333k), time: 0.00/ 0.00/ 0.00 Pass 3A: I/O read: 0MB, write: 0MB, rate: 0.00MB/s Pass 3: Memory used: 532k/0k (197k/336k), time: 0.00/ 0.00/ 0.00 Pass 3: I/O read: 1MB, write: 0MB, rate: 5747.13MB/s Pass 4: Checking reference counts Pass 4: Memory used: 468k/0k (66k/403k), time: 0.01/ 0.01/ 0.00 Pass 4: I/O read: 0MB, write: 0MB, rate: 0.00MB/s Pass 5: Checking group summary information Pass 5: Memory used: 468k/0k (65k/404k), time: 0.00/ 0.00/ 0.00 Pass 5: I/O read: 1MB, write: 0MB, rate: 270.93MB/s 402 inodes used (0.10%, out of 405072) 4 non-contiguous files (1.0%) 0 non-contiguous directories (0.0%) # of inodes with ind/dind/tind blocks: 0/0/0 Extent depth histogram: 396 89417 blocks used (15.16%, out of 589824) 0 bad blocks 1 large file 220 regular files 172 directories 0 character device files 0 block device files 0 fifos 0 links 0 symbolic links (0 fast symbolic links) 0 sockets ------------ 392 files Memory used: 468k/0k (64k/405k), time: 0.04/ 0.03/ 0.01 I/O read: 2MB, write: 1MB, rate: 52.79MB/s start mds service on oleg233-server Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions ../libcfs/libcfs/libcfs options: 'cpu_npartitions=2' ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' quota/lquota options: 'hash_lqs_cur_bits=3' loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions oleg233-server: libkmod: kmod_module_get_holders: could not open '/sys/module/acpi_cpufreq/holders': No such file or directory oleg233-server: ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' oleg233-server: quota/lquota options: 'hash_lqs_cur_bits=3' Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=131072 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre check files after shrinking the MDT and OST filesystems /mnt/lustre/d78.conf-sanity/f78.conf-sanity-1 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-1 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-2 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-2 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-3 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-3 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-4 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-4 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-5 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-5 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-6 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-6 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-7 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-7 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-8 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-8 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-9 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-9 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-10 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-10 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-11 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-11 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-12 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-12 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-13 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-13 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-14 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-14 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-15 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-15 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-16 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-16 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-17 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-17 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-18 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-18 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-19 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-19 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-20 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-20 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-21 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-21 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-22 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-22 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-23 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-23 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-24 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-24 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-25 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-25 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-26 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-26 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-27 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-27 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-28 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-28 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-29 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-29 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-30 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-30 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-31 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-31 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-32 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-32 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-33 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-33 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-34 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-34 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-35 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-35 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-36 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-36 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-37 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-37 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-38 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-38 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-39 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-39 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-40 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-40 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-41 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-41 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-42 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-42 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-43 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-43 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-44 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-44 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-45 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-45 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-46 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-46 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-47 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-47 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-48 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-48 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-49 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-49 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-50 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-50 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-51 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-51 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-52 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-52 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-53 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-53 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-54 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-54 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-55 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-55 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-56 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-56 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-57 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-57 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-58 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-58 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-59 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-59 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-60 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-60 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-61 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-61 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-62 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-62 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-63 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-63 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-64 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-64 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-65 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-65 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-66 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-66 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-67 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-67 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-68 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-68 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-69 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-69 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-70 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-70 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-71 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-71 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-72 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-72 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-73 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-73 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-74 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-74 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-75 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-75 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-76 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-76 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-77 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-77 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-78 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-78 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-79 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-79 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-80 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-80 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-81 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-81 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-82 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-82 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-83 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-83 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-84 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-84 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-85 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-85 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-86 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-86 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-87 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-87 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-88 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-88 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-89 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-89 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-90 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-90 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-91 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-91 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-92 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-92 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-93 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-93 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-94 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-94 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-95 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-95 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-96 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-96 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-97 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-97 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-98 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-98 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-99 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-99 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-100 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-100 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-101 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-101 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-102 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-102 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-103 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-103 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-104 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-104 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-105 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-105 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-106 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-106 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-107 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-107 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-108 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-108 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-109 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-109 has size 1048576 OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-110 has type file OK /mnt/lustre/d78.conf-sanity/f78.conf-sanity-110 has size 1048576 OK umount lustre on /mnt/lustre..... Stopping client oleg233-client.virtnet /mnt/lustre (opts:) stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server unloading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing unload_modules_local modules unloaded. Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:-f) Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:-f) pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 oleg233-server: oleg233-server.virtnet: executing set_hostid Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions ../libcfs/libcfs/libcfs options: 'cpu_npartitions=2' ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' quota/lquota options: 'hash_lqs_cur_bits=3' loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions oleg233-server: ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' oleg233-server: quota/lquota options: 'hash_lqs_cur_bits=3' Formatting mgs, mds, osts Format mds1: /dev/mapper/mds1_flakey Format mds2: /dev/mapper/mds2_flakey Format ost1: /dev/mapper/ost1_flakey Format ost2: /dev/mapper/ost2_flakey start mds service on oleg233-server Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/mds1_flakey Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/mds2_flakey Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/ost1_flakey Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid oleg233-server: oleg233-server.virtnet: executing wait_import_state FULL os[cp].lustre-OST0000-osc-MDT0000.ost_server_uuid 50 oleg233-server: os[cp].lustre-OST0000-osc-MDT0000.ost_server_uuid in FULL state after 3 sec oleg233-server: oleg233-server.virtnet: executing wait_import_state FULL os[cp].lustre-OST0000-osc-MDT0001.ost_server_uuid 50 oleg233-server: os[cp].lustre-OST0000-osc-MDT0001.ost_server_uuid in FULL state after 0 sec stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server PASS 78 (175s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 79: format MDT/OST without mgs option (should return errors) ========================================================== 09:01:01 (1713531661) Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions oleg233-server: oleg233-server: mkfs.lustre FATAL: Must specify --mgs or --mgsnode oleg233-server: mkfs.lustre: exiting with 22 (Invalid argument) pdsh@oleg233-client: oleg233-server: ssh exited with exit code 22 oleg233-server: oleg233-server: mkfs.lustre FATAL: Must specify --mgs or --mgsnode oleg233-server: mkfs.lustre: exiting with 22 (Invalid argument) pdsh@oleg233-client: oleg233-server: ssh exited with exit code 22 oleg233-server: oleg233-server: mkfs.lustre FATAL: Must specify --mgsnode oleg233-server: mkfs.lustre: exiting with 22 (Invalid argument) pdsh@oleg233-client: oleg233-server: ssh exited with exit code 22 Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:-f) Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:-f) oleg233-server: oleg233-server.virtnet: executing set_hostid Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions Formatting mgs, mds, osts Format mds1: /dev/mapper/mds1_flakey Format mds2: /dev/mapper/mds2_flakey Format ost1: /dev/mapper/ost1_flakey Format ost2: /dev/mapper/ost2_flakey start mds service on oleg233-server Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/mds1_flakey Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/mds2_flakey Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/ost1_flakey Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid oleg233-server: oleg233-server.virtnet: executing wait_import_state FULL os[cp].lustre-OST0000-osc-MDT0000.ost_server_uuid 50 oleg233-server: os[cp].lustre-OST0000-osc-MDT0000.ost_server_uuid in FULL state after 3 sec oleg233-server: oleg233-server.virtnet: executing wait_import_state FULL os[cp].lustre-OST0000-osc-MDT0001.ost_server_uuid 50 oleg233-server: os[cp].lustre-OST0000-osc-MDT0001.ost_server_uuid in FULL state after 0 sec stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server PASS 79 (61s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 80: mgc import reconnect race ======== 09:02:04 (1713531724) start mds service on oleg233-server Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid fail_val=10 fail_loc=0x906 fail_val=10 fail_loc=0x906 start ost2 service on oleg233-server Starting ost2: -o localrecov /dev/mapper/ost2_flakey /mnt/lustre-ost2 seq.cli-lustre-OST0001-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/ost2_flakey Started lustre-OST0001 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0001-osc-[-0-9a-f]*.ost_server_uuid fail_loc=0 stop ost2 service on oleg233-server Stopping /mnt/lustre-ost2 (opts:-f) on oleg233-server stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server PASS 80 (73s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 81: sparse OST indexing ============== 09:03:19 (1713531799) SKIP: conf-sanity test_81 needs >= 3 OSTs SKIP 81 (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 82a: specify OSTs for file (succeed) or directory (succeed) ========================================================== 09:03:22 (1713531802) SKIP: conf-sanity test_82a needs >= 3 OSTs SKIP 82a (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 82b: specify OSTs for file with --pool and --ost-list options ========================================================== 09:03:25 (1713531805) SKIP: conf-sanity test_82b needs >= 4 OSTs SKIP 82b (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 83: ENOSPACE on OST doesn't cause message VFS: Busy inodes after unmount ... ========================================================== 09:03:28 (1713531808) mount the OST /dev/mapper/ost1_flakey as a ldiskfs filesystem mnt_opts -o loop run llverfs in partial mode on the OST ldiskfs /mnt/lustre-ost1 oleg233-server: oleg233-server.virtnet: executing run_llverfs /mnt/lustre-ost1 -vpl no oleg233-server: oleg233-server: llverfs: write /mnt/lustre-ost1/llverfs_dir00142/file000@0+1048576 short: 368640 written oleg233-server: Timestamp: 1713531813 oleg233-server: dirs: 147, fs blocks: 37602 oleg233-server: write_done: /mnt/lustre-ost1/llverfs_dir00142/file000, current: 305.685 MB/s, overall: 305.685 MB/s, ETA: 0:00:00 oleg233-server: oleg233-server: read_done: /mnt/lustre-ost1/llverfs_dir00141/file000, current: 3507.91 MB/s, overall: 3507.91 MB/s, ETA: 0:00:00 oleg233-server: unmount the OST /dev/mapper/ost1_flakey Stopping /mnt/lustre-ost1 (opts:) on oleg233-server checking for existing Lustre data: found Read previous values: Target: lustre-MDT0000 Index: 0 Lustre FS: lustre Mount type: ldiskfs Flags: 0x5 (MDT MGS ) Persistent mount opts: user_xattr,errors=remount-ro Parameters: sys.timeout=20 mdt.identity_upcall=/home/green/git/lustre-release/lustre/utils/l_getidentity Permanent disk data: Target: lustre=MDT0000 Index: 0 Lustre FS: lustre Mount type: ldiskfs Flags: 0x105 (MDT MGS writeconf ) Persistent mount opts: user_xattr,errors=remount-ro Parameters: sys.timeout=20 mdt.identity_upcall=/home/green/git/lustre-release/lustre/utils/l_getidentity checking for existing Lustre data: found Read previous values: Target: lustre-MDT0001 Index: 1 Lustre FS: lustre Mount type: ldiskfs Flags: 0x1 (MDT ) Persistent mount opts: user_xattr,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 mdt.identity_upcall=/home/green/git/lustre-release/lustre/utils/l_getidentity Permanent disk data: Target: lustre=MDT0001 Index: 1 Lustre FS: lustre Mount type: ldiskfs Flags: 0x101 (MDT writeconf ) Persistent mount opts: user_xattr,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 mdt.identity_upcall=/home/green/git/lustre-release/lustre/utils/l_getidentity checking for existing Lustre data: found Read previous values: Target: lustre-OST0000 Index: 0 Lustre FS: lustre Mount type: ldiskfs Flags: 0x62 (OST first_time update ) Persistent mount opts: ,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 Permanent disk data: Target: lustre=OST0000 Index: 0 Lustre FS: lustre Mount type: ldiskfs Flags: 0x162 (OST first_time update writeconf ) Persistent mount opts: ,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 checking for existing Lustre data: found Read previous values: Target: lustre-OST0001 Index: 1 Lustre FS: lustre Mount type: ldiskfs Flags: 0x2 (OST ) Persistent mount opts: ,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 Permanent disk data: Target: lustre=OST0001 Index: 1 Lustre FS: lustre Mount type: ldiskfs Flags: 0x102 (OST writeconf ) Persistent mount opts: ,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 oleg233-server: mount.lustre: mount /dev/mapper/ost1_flakey at /mnt/lustre-ost1 failed: No space left on device pdsh@oleg233-client: oleg233-server: ssh exited with exit code 28 oleg233-server: error: set_param: param_path 'seq/cli-lustre': No such file or directory oleg233-server: error: set_param: setting 'seq/cli-lustre'='OST0000-super.width=65536': No such file or directory pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 Start of /dev/mapper/ost1_flakey on ost1 failed 28 string err Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:-f) Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:-f) oleg233-server: oleg233-server.virtnet: executing set_hostid Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions Formatting mgs, mds, osts Format mds1: /dev/mapper/mds1_flakey Format mds2: /dev/mapper/mds2_flakey Format ost1: /dev/mapper/ost1_flakey Format ost2: /dev/mapper/ost2_flakey start mds service on oleg233-server Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/mds1_flakey Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/mds2_flakey Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/ost1_flakey Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid oleg233-server: oleg233-server.virtnet: executing wait_import_state FULL os[cp].lustre-OST0000-osc-MDT0000.ost_server_uuid 50 oleg233-server: os[cp].lustre-OST0000-osc-MDT0000.ost_server_uuid in FULL state after 0 sec oleg233-server: oleg233-server.virtnet: executing wait_import_state FULL os[cp].lustre-OST0000-osc-MDT0001.ost_server_uuid 50 oleg233-server: os[cp].lustre-OST0000-osc-MDT0001.ost_server_uuid in FULL state after 0 sec stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server PASS 83 (70s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 84: check recovery_hard_time ========= 09:04:40 (1713531880) start mds service on oleg233-server start mds service on oleg233-server Starting mds1: -o localrecov -o recovery_time_hard=60,recovery_time_soft=60 /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov -o recovery_time_hard=60,recovery_time_soft=60 /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid oleg233-server: oleg233-server.virtnet: executing wait_import_state FULL os[cp].lustre-OST0000-osc-MDT0000.ost_server_uuid 50 oleg233-server: os[cp].lustre-OST0000-osc-MDT0000.ost_server_uuid in FULL state after 0 sec oleg233-server: oleg233-server.virtnet: executing wait_import_state FULL os[cp].lustre-OST0000-osc-MDT0001.ost_server_uuid 50 oleg233-server: os[cp].lustre-OST0000-osc-MDT0001.ost_server_uuid in FULL state after 0 sec start ost2 service on oleg233-server Starting ost2: -o localrecov /dev/mapper/ost2_flakey /mnt/lustre-ost2 seq.cli-lustre-OST0001-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/ost2_flakey Started lustre-OST0001 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0001-osc-[-0-9a-f]*.ost_server_uuid oleg233-server: oleg233-server.virtnet: executing wait_import_state FULL os[cp].lustre-OST0001-osc-MDT0000.ost_server_uuid 50 oleg233-server: os[cp].lustre-OST0001-osc-MDT0000.ost_server_uuid in FULL state after 2 sec oleg233-server: oleg233-server.virtnet: executing wait_import_state FULL os[cp].lustre-OST0001-osc-MDT0001.ost_server_uuid 50 oleg233-server: os[cp].lustre-OST0001-osc-MDT0001.ost_server_uuid in FULL state after 0 sec recovery_time=60, timeout=20, wrap_up=5 mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre mount lustre on /mnt/lustre2..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre2 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 95248 1668 84924 2% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 95248 1532 85060 2% /mnt/lustre[MDT:1] lustre-OST0000_UUID 142216 1524 126692 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 142216 1524 126692 2% /mnt/lustre[OST:1] filesystem_summary: 284432 3048 253384 2% /mnt/lustre total: 1000 open/close in 5.47 seconds: 182.93 ops/second fail_loc=0x20000709 fail_val=5 Failing mds1 on oleg233-server Stopping /mnt/lustre-mds1 (opts:) on oleg233-server 09:05:28 (1713531928) shut down Failover mds1 to oleg233-server e2fsck -d -v -t -t -f -n /dev/mapper/mds1_flakey -m8 oleg233-server: e2fsck 1.46.2.wc5 (26-Mar-2022) oleg233-server: Use max possible thread num: 1 instead Warning: skipping journal recovery because doing a read-only filesystem check. Pass 1: Checking inodes, blocks, and sizes [Thread 0] Scan group range [0, 3) [Thread 0] jumping to group 0 [Thread 0] e2fsck_pass1_run:2564: increase inode 81 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 82 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 83 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 84 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 85 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 86 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 87 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 88 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 89 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 90 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 91 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 92 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 93 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 94 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 95 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 96 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 97 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 98 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 99 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 100 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 101 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 102 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 103 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 104 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 105 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 106 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 107 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 108 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 109 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 110 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 111 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 112 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 113 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 114 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 115 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 116 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 117 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 118 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 119 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 120 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 121 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 122 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 123 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 124 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 125 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 126 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 127 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 128 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 129 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 130 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 131 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 132 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 133 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 134 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 135 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 136 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 137 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 138 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 139 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 140 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 141 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 142 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 143 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 144 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 145 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 146 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 147 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 148 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 149 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 150 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 151 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 152 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 153 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 154 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 155 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 156 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 157 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 158 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 159 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 160 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 161 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 162 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 163 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 164 badness 0 to 2 for 10084 [Thread 0] group 1 finished [Thread 0] e2fsck_pass1_run:2564: increase inode 26697 badness 0 to 2 for 10084 [Thread 0] group 2 finished [Thread 0] e2fsck_pass1_run:2564: increase inode 53372 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 53373 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 53374 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 53375 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 53376 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 53377 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 53378 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 53379 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 53380 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 53381 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 53382 badness 0 to 2 for 10084 [Thread 0] group 3 finished [Thread 0] Pass 1: Memory used: 264k/0k (140k/125k), time: 0.00/ 0.00/ 0.00 [Thread 0] Pass 1: I/O read: 1MB, write: 0MB, rate: 201.41MB/s [Thread 0] Scanned group range [0, 3), inodes 277 Pass 2: Checking directory structure Pass 2: Memory used: 264k/0k (97k/168k), time: 0.01/ 0.00/ 0.00 Pass 2: I/O read: 1MB, write: 0MB, rate: 183.32MB/s Pass 3: Checking directory connectivity Peak memory: Memory used: 264k/0k (97k/168k), time: 0.02/ 0.01/ 0.01 Pass 3: Memory used: 264k/0k (96k/169k), time: 0.00/ 0.00/ 0.00 Pass 3: I/O read: 0MB, write: 0MB, rate: 0.00MB/s Pass 4: Checking reference counts Pass 4: Memory used: 264k/0k (67k/198k), time: 0.00/ 0.00/ 0.00 Pass 4: I/O read: 0MB, write: 0MB, rate: 0.00MB/s Pass 5: Checking group summary information Free blocks count wrong (25455, counted=25443). Fix? no Free inodes count wrong (79719, counted=79715). Fix? no Pass 5: Memory used: 264k/0k (67k/198k), time: 0.00/ 0.00/ 0.00 Pass 5: I/O read: 1MB, write: 0MB, rate: 222.92MB/s 273 inodes used (0.34%, out of 79992) 5 non-contiguous files (1.8%) 0 non-contiguous directories (0.0%) # of inodes with ind/dind/tind blocks: 0/0/0 24545 blocks used (49.09%, out of 50000) 0 bad blocks 1 large file 150 regular files 117 directories 0 character device files 0 block device files 0 fifos 0 links 0 symbolic links (0 fast symbolic links) 0 sockets ------------ 267 files Memory used: 264k/0k (66k/199k), time: 0.03/ 0.01/ 0.01 I/O read: 1MB, write: 0MB, rate: 35.69MB/s mount facets: mds1 Starting mds1: -o localrecov -o recovery_time_hard=60,recovery_time_soft=60 /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 09:05:43 (1713531943) targets are mounted 09:05:43 (1713531943) facet_failover done oleg233-client: error: invalid path '/mnt/lustre': Input/output error pdsh@oleg233-client: oleg233-client: ssh exited with exit code 5 recovery status status: COMPLETE recovery_start: 1713531947 recovery_duration: 60 completed_clients: 2/3 replayed_requests: 158 last_transno: 8589934750 VBR: DISABLED IR: DISABLED fail_loc=0 umount lustre on /mnt/lustre..... Stopping client oleg233-client.virtnet /mnt/lustre (opts:) umount lustre on /mnt/lustre2..... Stopping client oleg233-client.virtnet /mnt/lustre2 (opts:) stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server stop ost2 service on oleg233-server Stopping /mnt/lustre-ost2 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server PASS 84 (153s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 85: osd_ost init: fail ea_fid_set ==== 09:07:16 (1713532036) fail_loc=0x197 start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server PASS 85 (191s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 86: Replacing mkfs.lustre -G option == 09:10:29 (1713532229) oleg233-server: debugfs 1.46.2.wc5 (26-Mar-2022) oleg233-server: /dev/mapper/ost1_flakey: catastrophic mode - not reading inode or group bitmaps params: --mgsnode=oleg233-server@tcp --fsname=lustre --ost --index=0 --param=sys.timeout=20 --backfstype=ldiskfs --device-size=200000 --mkfsoptions=\"-G 1024 -b 4096 -O flex_bg -E lazy_itable_init\" --reformat /dev/mapper/ost1_flakey Failing mds1 on oleg233-server 09:10:31 (1713532231) shut down Failover mds1 to oleg233-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 09:10:46 (1713532246) targets are mounted 09:10:46 (1713532246) facet_failover done pdsh@oleg233-client: oleg233-client: ssh exited with exit code 95 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid pdsh@oleg233-client: oleg233-client: ssh exited with exit code 95 Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:) Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:) Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server Permanent disk data: Target: lustre:OST0000 Index: 0 Lustre FS: lustre Mount type: ldiskfs Flags: 0x62 (OST first_time update ) Persistent mount opts: ,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 device size = 4096MB formatting backing filesystem ldiskfs on /dev/mapper/ost1_flakey target name lustre:OST0000 kilobytes 200000 options -G 1024 -b 4096 -I 512 -q -O flex_bg,uninit_bg,extents,dir_nlink,quota,project,huge_file,large_dir,^fast_commit -E lazy_itable_init,resize="4290772992",lazy_journal_init,packed_meta_blocks -F mkfs_cmd = mke2fs -j -b 4096 -L lustre:OST0000 -G 1024 -b 4096 -I 512 -q -O flex_bg,uninit_bg,extents,dir_nlink,quota,project,huge_file,large_dir,^fast_commit -E lazy_itable_init,resize="4290772992",lazy_journal_init,packed_meta_blocks -F /dev/mapper/ost1_flakey 200000k Writing CONFIGS/mountdata oleg233-server: debugfs 1.46.2.wc5 (26-Mar-2022) oleg233-server: /dev/mapper/ost1_flakey: catastrophic mode - not reading inode or group bitmaps Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:-f) Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:-f) oleg233-server: oleg233-server.virtnet: executing set_hostid Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions oleg233-server: libkmod: kmod_module_get_holders: could not open '/sys/module/pcc_cpufreq/holders': No such file or directory Formatting mgs, mds, osts Format mds1: /dev/mapper/mds1_flakey Format mds2: /dev/mapper/mds2_flakey Format ost1: /dev/mapper/ost1_flakey Format ost2: /dev/mapper/ost2_flakey start mds service on oleg233-server Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/mds1_flakey Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/mds2_flakey Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/ost1_flakey Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid oleg233-server: oleg233-server.virtnet: executing wait_import_state FULL os[cp].lustre-OST0000-osc-MDT0000.ost_server_uuid 50 oleg233-server: os[cp].lustre-OST0000-osc-MDT0000.ost_server_uuid in FULL state after 0 sec oleg233-server: oleg233-server.virtnet: executing wait_import_state FULL os[cp].lustre-OST0000-osc-MDT0001.ost_server_uuid 50 oleg233-server: os[cp].lustre-OST0000-osc-MDT0001.ost_server_uuid in FULL state after 0 sec stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server PASS 86 (79s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 87: check if MDT inode can hold EAs with N stripes properly ========================================================== 09:11:50 (1713532310) Estimate: at most 353-byte space left in inode. unloading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing unload_modules_local modules unloaded. Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:-f) Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:-f) pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 oleg233-server: oleg233-server.virtnet: executing set_hostid Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions ../libcfs/libcfs/libcfs options: 'cpu_npartitions=2' ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' quota/lquota options: 'hash_lqs_cur_bits=3' loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions oleg233-server: libkmod: kmod_module_get_holders: could not open '/sys/module/acpi_cpufreq/holders': No such file or directory oleg233-server: libkmod: kmod_module_get_holders: could not open '/sys/module/pcc_cpufreq/holders': No such file or directory oleg233-server: libkmod: kmod_module_get_holders: could not open '/sys/module/acpi_cpufreq/holders': No such file or directory oleg233-server: ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' oleg233-server: quota/lquota options: 'hash_lqs_cur_bits=3' Formatting mgs, mds, osts Format mds1: /dev/mapper/mds1_flakey Format mds2: /dev/mapper/mds2_flakey Format ost1: /dev/mapper/ost1_flakey Format ost2: /dev/mapper/ost2_flakey Permanent disk data: Target: lustre:MDT0000 Index: 0 Lustre FS: lustre Mount type: ldiskfs Flags: 0x65 (MDT MGS first_time update ) Persistent mount opts: user_xattr,errors=remount-ro Parameters: sys.timeout=20 mdt.identity_upcall=/home/green/git/lustre-release/lustre/utils/l_getidentity device size = 2500MB formatting backing filesystem ldiskfs on /dev/mapper/mds1_flakey target name lustre:MDT0000 kilobytes 200000 options -b 4096 -I 1024 -i 2560 -q -O uninit_bg,^extents,dirdata,dir_nlink,quota,project,huge_file,ea_inode,large_dir,^fast_commit,flex_bg -E lazy_itable_init,lazy_journal_init,packed_meta_blocks -F mkfs_cmd = mke2fs -j -b 4096 -L lustre:MDT0000 -b 4096 -I 1024 -i 2560 -q -O uninit_bg,^extents,dirdata,dir_nlink,quota,project,huge_file,ea_inode,large_dir,^fast_commit,flex_bg -E lazy_itable_init,lazy_journal_init,packed_meta_blocks -F /dev/mapper/mds1_flakey 200000k Writing CONFIGS/mountdata oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 setup single mount lustre success 4 -rw-r--r-- 1 root root 67108865 Apr 19 09:12 /mnt/lustre-mds1/ROOT/f87.conf-sanity Verified: at most 353-byte space left in inode. Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server Stopping /mnt/lustre-ost2 (opts:-f) on oleg233-server PASS 87 (62s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 88: check the default mount options can be overridden ========================================================== 09:12:54 (1713532374) Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions libkmod: kmod_module_get_holders: could not open '/sys/module/acpi_cpufreq/holders': No such file or directory loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions Permanent disk data: Target: lustre:MDT0000 Index: 0 Lustre FS: lustre Mount type: ldiskfs Flags: 0x65 (MDT MGS first_time update ) Persistent mount opts: user_xattr,errors=remount-ro Parameters: sys.timeout=20 mdt.identity_upcall=/home/green/git/lustre-release/lustre/utils/l_getidentity device size = 2500MB formatting backing filesystem ldiskfs on /dev/mapper/mds1_flakey target name lustre:MDT0000 kilobytes 200000 options -b 4096 -I 1024 -i 2560 -q -O uninit_bg,^extents,dirdata,dir_nlink,quota,project,huge_file,ea_inode,large_dir,^fast_commit,flex_bg -E lazy_itable_init,lazy_journal_init,packed_meta_blocks -F mkfs_cmd = mke2fs -j -b 4096 -L lustre:MDT0000 -b 4096 -I 1024 -i 2560 -q -O uninit_bg,^extents,dirdata,dir_nlink,quota,project,huge_file,ea_inode,large_dir,^fast_commit,flex_bg -E lazy_itable_init,lazy_journal_init,packed_meta_blocks -F /dev/mapper/mds1_flakey 200000k Writing CONFIGS/mountdata Persistent mount opts: user_xattr,errors=remount-ro Persistent mount opts: user_xattr,errors=remount-ro Permanent disk data: Target: lustre:MDT0000 Index: 0 Lustre FS: lustre Mount type: ldiskfs Flags: 0x65 (MDT MGS first_time update ) Persistent mount opts: user_xattr,errors=panic Parameters: sys.timeout=20 mdt.identity_upcall=/home/green/git/lustre-release/lustre/utils/l_getidentity device size = 2500MB formatting backing filesystem ldiskfs on /dev/mapper/mds1_flakey target name lustre:MDT0000 kilobytes 200000 options -b 4096 -I 1024 -i 2560 -q -O uninit_bg,^extents,dirdata,dir_nlink,quota,project,huge_file,ea_inode,large_dir,^fast_commit,flex_bg -E lazy_itable_init,lazy_journal_init,packed_meta_blocks -F mkfs_cmd = mke2fs -j -b 4096 -L lustre:MDT0000 -b 4096 -I 1024 -i 2560 -q -O uninit_bg,^extents,dirdata,dir_nlink,quota,project,huge_file,ea_inode,large_dir,^fast_commit,flex_bg -E lazy_itable_init,lazy_journal_init,packed_meta_blocks -F /dev/mapper/mds1_flakey 200000k Writing CONFIGS/mountdata Persistent mount opts: user_xattr,errors=panic Persistent mount opts: user_xattr,errors=panic PASS 88 (11s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 89: check tunefs --param and --erase-param{s} options ========================================================== 09:13:07 (1713532387) Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:) Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:) tunefs --param failover.node=192.0.2.254@tcp0 tunefs --param failover.node=192.0.2.255@tcp0 tunefs --erase-param failover.node tunefs --erase-params tunefs --param failover.node=192.0.2.254@tcp0 --erase-params Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:-f) Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:-f) oleg233-server: oleg233-server.virtnet: executing set_hostid Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions Formatting mgs, mds, osts Format mds1: /dev/mapper/mds1_flakey Format mds2: /dev/mapper/mds2_flakey Format ost1: /dev/mapper/ost1_flakey Format ost2: /dev/mapper/ost2_flakey start mds service on oleg233-server Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/mds1_flakey Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/mds2_flakey Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL failover.node=192.0.2.254@tcp0,mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL failover.node=192.0.2.254@tcp0,mdc.lustre-MDT0001-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/ost1_flakey Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) failover.node=192.0.2.254@tcp0,osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid oleg233-server: oleg233-server.virtnet: executing wait_import_state FULL os[cp].lustre-OST0000-osc-MDT0000.ost_server_uuid 50 oleg233-server: os[cp].lustre-OST0000-osc-MDT0000.ost_server_uuid in FULL state after 0 sec oleg233-server: oleg233-server.virtnet: executing wait_import_state FULL os[cp].lustre-OST0000-osc-MDT0001.ost_server_uuid 50 oleg233-server: os[cp].lustre-OST0000-osc-MDT0001.ost_server_uuid in FULL state after 0 sec stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server PASS 89 (64s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 90a: check max_mod_rpcs_in_flight is enforced ========================================================== 09:14:13 (1713532453) start mds service on oleg233-server Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre max_mod_rcps_in_flight is 7 creating 8 files ... fail_loc=0x159 launch 6 chmod in parallel ... fail_loc=0 launch 1 additional chmod in parallel ... /mnt/lustre/d90a.conf-sanity/file-7 has perms 0600 OK fail_loc=0x159 launch 7 chmod in parallel ... fail_loc=0 launch 1 additional chmod in parallel ... /mnt/lustre/d90a.conf-sanity/file-8 has perms 0644 OK umount lustre on /mnt/lustre..... Stopping client oleg233-client.virtnet /mnt/lustre (opts:) stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server unloading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing unload_modules_local modules unloaded. pdsh@oleg233-client: oleg233-client: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-client: ssh exited with exit code 2 PASS 90a (88s) error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='0': No such file or directory oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='0': No such file or directory error: get_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory oleg233-server: error: get_param: param_path 'debug_raw_pointers': No such file or directory pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory == conf-sanity test 90b: check max_mod_rpcs_in_flight is enforced after update ========================================================== 09:15:43 (1713532543) start mds service on oleg233-server Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' quota/lquota options: 'hash_lqs_cur_bits=3' loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions oleg233-server: libkmod: kmod_module_get_holders: could not open '/sys/module/acpi_cpufreq/holders': No such file or directory oleg233-server: ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' oleg233-server: quota/lquota options: 'hash_lqs_cur_bits=3' Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre mdc.lustre-MDT0000-mdc-ffff8800b3f03800.max_mod_rpcs_in_flight=1 max_mod_rpcs_in_flight set to 1 creating 2 files ... fail_loc=0x159 launch 0 chmod in parallel ... fail_loc=0 launch 1 additional chmod in parallel ... /mnt/lustre/d90b.conf-sanity1/file-1 has perms 0600 OK fail_loc=0x159 launch 1 chmod in parallel ... fail_loc=0 launch 1 additional chmod in parallel ... /mnt/lustre/d90b.conf-sanity1/file-2 has perms 0644 OK mdc.lustre-MDT0001-mdc-ffff8800b3f03800.max_mod_rpcs_in_flight=5 max_mod_rpcs_in_flight set to 5 creating 6 files ... fail_loc=0x159 launch 4 chmod in parallel ... fail_loc=0 launch 1 additional chmod in parallel ... /mnt/lustre/d90b.conf-sanity2/file-5 has perms 0600 OK fail_loc=0x159 launch 5 chmod in parallel ... fail_loc=0 launch 1 additional chmod in parallel ... /mnt/lustre/d90b.conf-sanity2/file-6 has perms 0644 OK mdt_max_mod_rpcs_in_flight is 8 umount lustre on /mnt/lustre..... Stopping client oleg233-client.virtnet /mnt/lustre (opts:) error: get_param: param_path 'mdt/*/max_mod_rpcs_in_flight': No such file or directory the deprecated max_mod_rpcs_per_client parameter was involved mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre mdc.lustre-MDT0000-mdc-ffff88012bf61800.max_rpcs_in_flight=17 mdc.lustre-MDT0000-mdc-ffff88012bf61800.max_mod_rpcs_in_flight=16 max_mod_rpcs_in_flight set to 16 creating 17 files ... fail_loc=0x159 launch 15 chmod in parallel ... fail_loc=0 launch 1 additional chmod in parallel ... /mnt/lustre/d90b.conf-sanity3/file-16 has perms 0600 OK fail_loc=0x159 launch 16 chmod in parallel ... fail_loc=0 launch 1 additional chmod in parallel ... /mnt/lustre/d90b.conf-sanity3/file-17 has perms 0644 OK error: get_param: param_path 'mdt/*/max_mod_rpcs_in_flight': No such file or directory the deprecated max_mod_rpcs_per_client parameter was involved umount lustre on /mnt/lustre..... Stopping client oleg233-client.virtnet /mnt/lustre (opts:) stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server unloading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing unload_modules_local modules unloaded. pdsh@oleg233-client: oleg233-client: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-client: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 PASS 90b (141s) error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='0': No such file or directory oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='0': No such file or directory error: get_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory oleg233-server: error: get_param: param_path 'debug_raw_pointers': No such file or directory pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory == conf-sanity test 90c: check max_mod_rpcs_in_flight update limits ========================================================== 09:18:06 (1713532686) start mds service on oleg233-server Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions libkmod: kmod_module_get_holders: could not open '/sys/module/acpi_cpufreq/holders': No such file or directory ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' quota/lquota options: 'hash_lqs_cur_bits=3' loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions oleg233-server: ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' oleg233-server: quota/lquota options: 'hash_lqs_cur_bits=3' Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre max_rpcs_in_flight is 8 MDC max_mod_rpcs_in_flight is 7 mdt_max_mod_rpcs_in_flight is 8 error: get_param: param_path 'mdt/*/max_mod_rpcs_in_flight': No such file or directory the deprecated max_mod_rpcs_per_client parameter was involved mdc.lustre-MDT0000-mdc-ffff88012a6f8800.max_mod_rpcs_in_flight=8 umount lustre on /mnt/lustre..... Stopping client oleg233-client.virtnet /mnt/lustre (opts:) error: get_param: param_path 'mdt/*/max_mod_rpcs_in_flight': No such file or directory the deprecated max_mod_rpcs_per_client parameter was involved mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre mdc.lustre-MDT0000-mdc-ffff8800af2ce000.max_rpcs_in_flight=10 error: set_param: setting /sys/fs/lustre/mdc/lustre-MDT0000-mdc-ffff8800af2ce000/max_mod_rpcs_in_flight=9: Numerical result out of range error: set_param: setting 'mdc/lustre-MDT0000-mdc-*/max_mod_rpcs_in_flight'='9': Numerical result out of range Stopping client oleg233-client.virtnet /mnt/lustre (opts:) Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre error: get_param: param_path 'mdt/*/max_mod_rpcs_in_flight': No such file or directory the deprecated max_mod_rpcs_per_client parameter was involved umount lustre on /mnt/lustre..... Stopping client oleg233-client.virtnet /mnt/lustre (opts:) stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server unloading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing unload_modules_local modules unloaded. pdsh@oleg233-client: oleg233-client: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-client: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 PASS 90c (70s) error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='0': No such file or directory oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='0': No such file or directory error: get_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory oleg233-server: error: get_param: param_path 'debug_raw_pointers': No such file or directory pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory == conf-sanity test 90d: check one close RPC is allowed above max_mod_rpcs_in_flight ========================================================== 09:19:18 (1713532758) start mds service on oleg233-server Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions libkmod: kmod_module_get_holders: could not open '/sys/module/pcc_cpufreq/holders': No such file or directory libkmod: kmod_module_get_holders: could not open '/sys/module/acpi_cpufreq/holders': No such file or directory ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' quota/lquota options: 'hash_lqs_cur_bits=3' loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions oleg233-server: ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' oleg233-server: quota/lquota options: 'hash_lqs_cur_bits=3' Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre max_mod_rcps_in_flight is 7 creating 7 files ... multiop /mnt/lustre/d90d.conf-sanity/file-close vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.7349 fail_loc=0x159 launch 7 chmod in parallel ... fail_loc=0 launch 1 additional close in parallel ... umount lustre on /mnt/lustre..... Stopping client oleg233-client.virtnet /mnt/lustre (opts:) stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server unloading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing unload_modules_local modules unloaded. pdsh@oleg233-client: oleg233-client: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-client: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 PASS 90d (83s) error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='0': No such file or directory oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='0': No such file or directory error: get_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory oleg233-server: error: get_param: param_path 'debug_raw_pointers': No such file or directory pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory == conf-sanity test 91: evict-by-nid support ============= 09:20:43 (1713532843) start mds service on oleg233-server Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' quota/lquota options: 'hash_lqs_cur_bits=3' loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions oleg233-server: ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' oleg233-server: quota/lquota options: 'hash_lqs_cur_bits=3' Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre setup single mount lustre success list nids on mdt: mdt.lustre-MDT0000.exports.0@lo mdt.lustre-MDT0000.exports.192.168.202.33@tcp mdt.lustre-MDT0000.exports.clear mdt.lustre-MDT0001.exports.0@lo mdt.lustre-MDT0001.exports.192.168.202.33@tcp mdt.lustre-MDT0001.exports.clear uuid from 192\.168\.202\.33@tcp: mdt.lustre-MDT0000.exports.192.168.202.33@tcp.uuid=148ac5db-dc5c-4e4c-8e1d-aa7b78d357c4 mdt.lustre-MDT0001.exports.192.168.202.33@tcp.uuid=148ac5db-dc5c-4e4c-8e1d-aa7b78d357c4 manual umount lustre on /mnt/lustre.... evict 192\.168\.202\.33@tcp oleg233-server: error: read_param: '/proc/fs/lustre/mdt/lustre-MDT0000/exports/192.168.202.33@tcp/uuid': No such device pdsh@oleg233-client: oleg233-server: ssh exited with exit code 19 oleg233-server: error: read_param: '/proc/fs/lustre/obdfilter/lustre-OST0000/exports/192.168.202.33@tcp/uuid': No such device pdsh@oleg233-client: oleg233-server: ssh exited with exit code 19 oleg233-server: error: read_param: '/proc/fs/lustre/mdt/lustre-MDT0000/exports/192.168.202.33@tcp/uuid': No such device pdsh@oleg233-client: oleg233-server: ssh exited with exit code 19 oleg233-server: error: read_param: '/proc/fs/lustre/obdfilter/lustre-OST0000/exports/192.168.202.33@tcp/uuid': No such device pdsh@oleg233-client: oleg233-server: ssh exited with exit code 19 umount lustre on /mnt/lustre..... stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server unloading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing unload_modules_local modules unloaded. pdsh@oleg233-client: oleg233-client: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-client: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 PASS 91 (89s) error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='0': No such file or directory oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='0': No such file or directory error: get_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory oleg233-server: error: get_param: param_path 'debug_raw_pointers': No such file or directory pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory == conf-sanity test 92: ldev returns MGS NID correctly in command substitution ========================================================== 09:22:15 (1713532935) Host is oleg233-client.virtnet ----- /tmp/ldev.conf ----- oleg233-server - lustre-MGS0000 /dev/mapper/mds1_flakey oleg233-server - lustre-OST0000 /dev/mapper/ost1_flakey oleg233-server - lustre-OST0001 /dev/mapper/ost2_flakey oleg233-server - lustre-MDT0000 /dev/mapper/mds1_flakey oleg233-server - lustre-MDT0001 /dev/mapper/mds2_flakey --- END /tmp/ldev.conf --- ----- /tmp/nids ----- oleg233-server oleg233-server@tcp --- END /tmp/nids --- -- START OF LDEV OUTPUT -- lustre-OST0000: oleg233-server@tcp lustre-OST0001: oleg233-server@tcp lustre-MGS0000: oleg233-server@tcp lustre-MDT0000: oleg233-server@tcp lustre-MDT0001: oleg233-server@tcp --- END OF LDEV OUTPUT --- -- START OF LDEV OUTPUT -- lustre-OST0000: oleg233-server@tcp lustre-MGS0000: oleg233-server@tcp lustre-OST0001: oleg233-server@tcp lustre-MDT0000: oleg233-server@tcp lustre-MDT0001: oleg233-server@tcp --- END OF LDEV OUTPUT --- -- START OF LDEV OUTPUT -- lustre-OST0000: oleg233-server@tcp lustre-OST0001: oleg233-server@tcp lustre-MGS0000: oleg233-server@tcp lustre-MDT0001: oleg233-server@tcp lustre-MDT0000: oleg233-server@tcp --- END OF LDEV OUTPUT --- -- START OF LDEV OUTPUT -- lustre-OST0000: oleg233-server@tcp lustre-OST0001: oleg233-server@tcp lustre-MDT0000: oleg233-server@tcp lustre-MGS0000: oleg233-server@tcp lustre-MDT0001: oleg233-server@tcp --- END OF LDEV OUTPUT --- -- START OF LDEV OUTPUT -- lustre-OST0001: oleg233-server@tcp lustre-OST0000: oleg233-server@tcp lustre-MDT0000: oleg233-server@tcp lustre-MGS0000: oleg233-server@tcp lustre-MDT0001: oleg233-server@tcp --- END OF LDEV OUTPUT --- pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 PASS 92 (4s) error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='0': No such file or directory oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='0': No such file or directory error: get_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory oleg233-server: error: get_param: param_path 'debug_raw_pointers': No such file or directory pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory == conf-sanity test 93: register mulitple MDT at the same time ========================================================== 09:22:20 (1713532940) SKIP: conf-sanity test_93 needs >= 3 MDTs SKIP 93 (1s) error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='0': No such file or directory oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='0': No such file or directory error: get_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory oleg233-server: error: get_param: param_path 'debug_raw_pointers': No such file or directory pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory == conf-sanity test 94: ldev outputs correct labels for file system name query ========================================================== 09:22:23 (1713532943) ----- /tmp/ldev.conf ----- oleg233-server - lustre-MGS0000 /dev/mapper/mds1_flakey oleg233-server - lustre-OST0000 /dev/mapper/ost1_flakey oleg233-server - lustre-OST0001 /dev/mapper/ost2_flakey oleg233-server - lustre-MDT0000 /dev/mapper/mds1_flakey oleg233-server - lustre-MDT0001 /dev/mapper/mds2_flakey --- END /tmp/ldev.conf --- ----- /tmp/nids ----- oleg233-server oleg233-server@tcp --- END /tmp/nids --- -- START OF LDEV OUTPUT -- lustre-MDT0000 lustre-MDT0001 lustre-MGS0000 lustre-OST0000 lustre-OST0001 --- END OF LDEV OUTPUT --- -- START OF EXPECTED OUTPUT -- lustre-MDT0000 lustre-MDT0001 lustre-MGS0000 lustre-OST0000 lustre-OST0001 --- END OF EXPECTED OUTPUT --- pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 PASS 94 (3s) error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='0': No such file or directory oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='0': No such file or directory error: get_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory oleg233-server: error: get_param: param_path 'debug_raw_pointers': No such file or directory pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory == conf-sanity test 95: ldev should only allow one label filter ========================================================== 09:22:29 (1713532949) ----- /tmp/ldev.conf ----- oleg233-server - lustre-MGS0000 /dev/mapper/mds1_flakey oleg233-server - lustre-OST0000 /dev/mapper/ost1_flakey oleg233-server - lustre-OST0001 /dev/mapper/ost2_flakey oleg233-server - lustre-MDT0000 /dev/mapper/mds1_flakey oleg233-server - lustre-MDT0001 /dev/mapper/mds2_flakey --- END /tmp/ldev.conf --- ----- /tmp/nids ----- oleg233-server oleg233-server@tcp --- END /tmp/nids --- pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 PASS 95 (3s) error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='0': No such file or directory oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='0': No such file or directory error: get_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory oleg233-server: error: get_param: param_path 'debug_raw_pointers': No such file or directory pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory == conf-sanity test 96: ldev returns hostname and backend fs correctly in command sub ========================================================== 09:22:34 (1713532954) ----- /tmp/ldev.conf ----- oleg233-server - lustre-MGS0000 /dev/mapper/mds1_flakey oleg233-server - lustre-OST0000 /dev/mapper/ost1_flakey oleg233-server - lustre-OST0001 /dev/mapper/ost2_flakey oleg233-server - lustre-MDT0000 /dev/mapper/mds1_flakey oleg233-server - lustre-MDT0001 /dev/mapper/mds2_flakey --- END /tmp/ldev.conf --- ----- /tmp/nids ----- oleg233-server oleg233-server@tcp --- END /tmp/nids --- -- START OF LDEV OUTPUT -- oleg233-server-ldiskfs oleg233-server-ldiskfs oleg233-server-ldiskfs oleg233-server-ldiskfs oleg233-server-ldiskfs --- END OF LDEV OUTPUT --- -- START OF EXPECTED OUTPUT -- oleg233-server-ldiskfs oleg233-server-ldiskfs oleg233-server-ldiskfs oleg233-server-ldiskfs oleg233-server-ldiskfs --- END OF EXPECTED OUTPUT --- pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 PASS 96 (2s) error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='0': No such file or directory oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='0': No such file or directory error: get_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory oleg233-server: error: get_param: param_path 'debug_raw_pointers': No such file or directory pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory == conf-sanity test 97: ldev returns correct ouput when querying based on role ========================================================== 09:22:38 (1713532958) ----- /tmp/ldev.conf ----- oleg233-server - lustre-MGS0000 /dev/mapper/mds1_flakey oleg233-server - lustre-OST0000 /dev/mapper/ost1_flakey oleg233-server - lustre-OST0001 /dev/mapper/ost2_flakey oleg233-server - lustre-MDT0000 /dev/mapper/mds1_flakey oleg233-server - lustre-MDT0001 /dev/mapper/mds2_flakey --- END /tmp/ldev.conf --- ----- /tmp/nids ----- oleg233-server oleg233-server@tcp --- END /tmp/nids --- MDT role -- START OF LDEV OUTPUT -- lustre-MDT0000 lustre-MDT0001 --- END OF LDEV OUTPUT --- -- START OF EXPECTED OUTPUT -- lustre-MDT0000 lustre-MDT0001 --- END OF EXPECTED OUTPUT --- OST role -- START OF LDEV OUTPUT -- lustre-OST0000 lustre-OST0001 --- END OF LDEV OUTPUT --- -- START OF EXPECTED OUTPUT -- lustre-OST0000 lustre-OST0001 --- END OF EXPECTED OUTPUT --- MGS role -- START OF LDEV OUTPUT -- lustre-MGS0000 --- END OF LDEV OUTPUT --- -- START OF EXPECTED OUTPUT -- lustre-MGS0000 --- END OF EXPECTED OUTPUT --- pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 PASS 97 (3s) error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='0': No such file or directory oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='0': No such file or directory error: get_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory oleg233-server: error: get_param: param_path 'debug_raw_pointers': No such file or directory pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory == conf-sanity test 98: Buffer-overflow check while parsing mount_opts ========================================================== 09:22:43 (1713532963) start mds service on oleg233-server Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' quota/lquota options: 'hash_lqs_cur_bits=3' loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions oleg233-server: ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' oleg233-server: quota/lquota options: 'hash_lqs_cur_bits=3' Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre setup single mount lustre success error: mount options too long umount lustre on /mnt/lustre..... Stopping client oleg233-client.virtnet /mnt/lustre (opts:) stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server unloading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing unload_modules_local modules unloaded. pdsh@oleg233-client: oleg233-client: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-client: ssh exited with exit code 2 PASS 98 (50s) error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='0': No such file or directory oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='0': No such file or directory error: get_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory oleg233-server: error: get_param: param_path 'debug_raw_pointers': No such file or directory pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory == conf-sanity test 99: Adding meta_bg option ============ 09:23:35 (1713533015) oleg233-server: debugfs 1.46.2.wc5 (26-Mar-2022) oleg233-server: /dev/mapper/ost1_flakey: catastrophic mode - not reading inode or group bitmaps params: --mgsnode=oleg233-server@tcp --fsname=lustre --ost --index=0 --param=sys.timeout=20 --backfstype=ldiskfs --device-size=200000 --mkfsoptions=\"-O ^resize_inode,meta_bg -b 4096 -E lazy_itable_init\" --reformat /dev/mapper/ost1_flakey Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' quota/lquota options: 'hash_lqs_cur_bits=3' loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions oleg233-server: ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' oleg233-server: quota/lquota options: 'hash_lqs_cur_bits=3' Permanent disk data: Target: lustre:OST0000 Index: 0 Lustre FS: lustre Mount type: ldiskfs Flags: 0x62 (OST first_time update ) Persistent mount opts: ,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 device size = 4096MB formatting backing filesystem ldiskfs on /dev/mapper/ost1_flakey target name lustre:OST0000 kilobytes 200000 options -b 4096 -I 512 -q -O ^resize_inode,meta_bg,uninit_bg,extents,dir_nlink,quota,project,huge_file,large_dir,^fast_commit,flex_bg -G 256 -E lazy_itable_init,lazy_journal_init,packed_meta_blocks -F mkfs_cmd = mke2fs -j -b 4096 -L lustre:OST0000 -b 4096 -I 512 -q -O ^resize_inode,meta_bg,uninit_bg,extents,dir_nlink,quota,project,huge_file,large_dir,^fast_commit,flex_bg -G 256 -E lazy_itable_init,lazy_journal_init,packed_meta_blocks -F /dev/mapper/ost1_flakey 200000k Writing CONFIGS/mountdata oleg233-server: debugfs 1.46.2.wc5 (26-Mar-2022) oleg233-server: /dev/mapper/ost1_flakey: catastrophic mode - not reading inode or group bitmaps Filesystem features: has_journal ext_attr dir_index filetype meta_bg extent flex_bg large_dir sparse_super large_file huge_file uninit_bg dir_nlink quota project PASS 99 (12s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 100: check lshowmount lists MGS, MDT, OST and 0@lo ========================================================== 09:23:48 (1713533028) Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:-f) Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:-f) oleg233-server: oleg233-server.virtnet: executing set_hostid Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions Formatting mgs, mds, osts Format mds1: /dev/mapper/mds1_flakey Format mds2: /dev/mapper/mds2_flakey Format ost1: /dev/mapper/ost1_flakey Format ost2: /dev/mapper/ost2_flakey start mds service on oleg233-server Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/mds1_flakey Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/mds2_flakey Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/ost1_flakey Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre setup single mount lustre success lustre-MDT0000: lustre-OST0000: umount lustre on /mnt/lustre..... Stopping client oleg233-client.virtnet /mnt/lustre (opts:) stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server unloading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing unload_modules_local modules unloaded. pdsh@oleg233-client: oleg233-client: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-client: ssh exited with exit code 2 PASS 100 (64s) error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='0': No such file or directory oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='0': No such file or directory error: get_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory oleg233-server: error: get_param: param_path 'debug_raw_pointers': No such file or directory pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory == conf-sanity test 101a: Race MDT->OST reconnection with create ========================================================== 09:24:54 (1713533094) start mds service on oleg233-server Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' quota/lquota options: 'hash_lqs_cur_bits=3' loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions oleg233-server: libkmod: kmod_module_get_holders: could not open '/sys/module/acpi_cpufreq/holders': No such file or directory oleg233-server: ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' oleg233-server: quota/lquota options: 'hash_lqs_cur_bits=3' Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre seq.cli-lustre-OST0000-super.width=0x1ffffff - open/close 525 (time 1713533133.95 total 10.01 last 52.43) - open/close 935 (time 1713533144.53 total 20.59 last 38.77) - open/close 1447 (time 1713533155.49 total 31.55 last 46.72) - open/close 1971 (time 1713533165.49 total 41.55 last 52.39) - open/close 2387 (time 1713533176.25 total 52.31 last 38.66) - open/close 2885 (time 1713533186.29 total 62.35 last 49.61) - open/close 3363 (time 1713533196.29 total 72.35 last 47.79) - open/close 3852 (time 1713533206.29 total 82.35 last 48.89) - open/close 4317 (time 1713533216.29 total 92.35 last 46.49) - open/close 4792 (time 1713533226.29 total 102.35 last 47.50) - open/close 5136 (time 1713533236.83 total 112.89 last 32.64) - open/close 5384 (time 1713533247.05 total 123.12 last 24.26) - open/close 5822 (time 1713533257.06 total 133.12 last 43.78) - open/close 6250 (time 1713533267.06 total 143.12 last 42.79) - open/close 6657 (time 1713533277.78 total 153.85 last 37.96) - open/close 7072 (time 1713533288.75 total 164.81 last 37.85) - open/close 7366 (time 1713533299.24 total 175.30 last 28.03) - open/close 7752 (time 1713533310.06 total 186.12 last 35.67) - open/close 8128 (time 1713533320.91 total 196.97 last 34.66) - open/close 8514 (time 1713533331.91 total 207.97 last 35.09) - open/close 8869 (time 1713533342.79 total 218.85 last 32.63) - open/close 9214 (time 1713533353.68 total 229.75 last 31.66) - open/close 9388 (time 1713533363.69 total 239.75 last 17.39) - open/close 9598 (time 1713533373.72 total 249.78 last 20.93) - open/close 9883 (time 1713533384.32 total 260.39 last 26.88) - open/close 10000 (time 1713533392.07 total 268.13 last 15.10) - open/close 10308 (time 1713533402.91 total 278.97 last 28.42) - open/close 10578 (time 1713533413.48 total 289.54 last 25.54) - open/close 10848 (time 1713533424.06 total 300.12 last 25.53) - open/close 11201 (time 1713533434.26 total 310.32 last 34.61) - open/close 12492 (time 1713533444.26 total 320.32 last 129.07) - open/close 14082 (time 1713533454.26 total 330.32 last 158.97) - open/close 15573 (time 1713533464.27 total 340.33 last 149.03) - open/close 17001 (time 1713533474.27 total 350.33 last 142.72) - open/close 18406 (time 1713533484.27 total 360.33 last 140.48) - open/close 19693 (time 1713533494.28 total 370.34 last 128.63) - open/close 20000 (time 1713533496.57 total 372.63 last 134.04) - open/close 21371 (time 1713533506.57 total 382.64 last 137.04) - open/close 22557 (time 1713533516.58 total 392.64 last 118.56) - open/close 23753 (time 1713533526.58 total 402.65 last 119.51) - open/close 24954 (time 1713533536.59 total 412.65 last 120.08) - open/close 26241 (time 1713533546.59 total 422.66 last 128.60) - open/close 27422 (time 1713533556.60 total 432.66 last 118.06) - open/close 28503 (time 1713533566.60 total 442.67 last 108.03) - open/close 29563 (time 1713533576.61 total 452.67 last 105.92) - open/close 30000 (time 1713533580.26 total 456.32 last 119.79) - open/close 31024 (time 1713533590.26 total 466.32 last 102.37) - open/close 32036 (time 1713533600.27 total 476.33 last 101.13) - open/close 33089 (time 1713533610.27 total 486.34 last 105.27) - open/close 34025 (time 1713533620.28 total 496.34 last 93.57) - open/close 35250 (time 1713533630.28 total 506.34 last 122.48) - open/close 36242 (time 1713533640.28 total 516.34 last 99.16) - open/close 37138 (time 1713533650.28 total 526.34 last 89.59) - open/close 37932 (time 1713533660.29 total 536.35 last 79.32) - open/close 38715 (time 1713533670.29 total 546.36 last 78.28) - open/close 39480 (time 1713533680.30 total 556.36 last 76.48) - open/close 40000 (time 1713533687.45 total 563.51 last 72.70) - open/close 40694 (time 1713533697.46 total 573.52 last 69.31) - open/close 41450 (time 1713533707.47 total 583.53 last 75.52) - open/close 42199 (time 1713533717.48 total 593.55 last 74.82) - open/close 42855 (time 1713533727.50 total 603.56 last 65.53) - open/close 43542 (time 1713533737.50 total 613.57 last 68.64) - open/close 44170 (time 1713533747.51 total 623.57 last 62.79) - open/close 44792 (time 1713533757.51 total 633.57 last 62.16) - open/close 45402 (time 1713533767.52 total 643.59 last 60.93) - open/close 46137 (time 1713533777.53 total 653.59 last 73.43) - open/close 46793 (time 1713533787.54 total 663.60 last 65.54) - open/close 47341 (time 1713533797.56 total 673.62 last 54.72) - open/close 47890 (time 1713533807.57 total 683.63 last 54.84) - open/close 48437 (time 1713533817.57 total 693.63 last 54.67) - open/close 48972 (time 1713533827.58 total 703.64 last 53.45) - open/close 49536 (time 1713533837.60 total 713.66 last 56.28) open(/mnt/lustre/d101a.conf-sanity/f101a.conf-sanity-49632) error: No space left on device total: 49632 open/close in 715.64 seconds: 69.35 ops/second - unlinked 0 (time 1713533840 ; total 0 ; last 0) - unlinked 10000 (time 1713533986 ; total 146 ; last 146) - unlinked 20000 (time 1713534093 ; total 253 ; last 107) - unlinked 30000 (time 1713534161 ; total 321 ; last 68) - unlinked 40000 (time 1713534213 ; total 373 ; last 52) unlink(/mnt/lustre/d101a.conf-sanity/f101a.conf-sanity-49632) error: No such file or directory total: 49632 unlinks in 403 seconds: 123.156326 unlinks/second umount lustre on /mnt/lustre..... Stopping client oleg233-client.virtnet /mnt/lustre (opts:) stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server unloading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing unload_modules_local modules unloaded. pdsh@oleg233-client: oleg233-client: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-client: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 PASS 101a (1170s) error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='0': No such file or directory oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='0': No such file or directory error: get_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory oleg233-server: error: get_param: param_path 'debug_raw_pointers': No such file or directory pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory == conf-sanity test 101b: Race events DISCONNECT and ACTIVE in osp ========================================================== 09:44:26 (1713534266) start mds service on oleg233-server Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' quota/lquota options: 'hash_lqs_cur_bits=3' loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions oleg233-server: ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' oleg233-server: quota/lquota options: 'hash_lqs_cur_bits=3' Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre fail_loc=0x80002107 fail_val=20 stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 12 sec oleg233-client.virtnet: executing wait_import_state (FULL|IDLE) osc.lustre-OST0000-osc-ffff88012aaf6800.ost_server_uuid 50 osc.lustre-OST0000-osc-ffff88012aaf6800.ost_server_uuid in FULL state after 0 sec umount lustre on /mnt/lustre..... Stopping client oleg233-client.virtnet /mnt/lustre (opts:) stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server unloading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing unload_modules_local modules unloaded. pdsh@oleg233-client: oleg233-client: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-client: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 PASS 101b (95s) error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='0': No such file or directory oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='0': No such file or directory SKIP: conf-sanity test_102 skipping excluded test 102 error: get_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory oleg233-server: error: get_param: param_path 'debug_raw_pointers': No such file or directory pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory == conf-sanity test 103: rename filesystem name ========== 09:46:03 (1713534363) Checking servers environments Checking clients oleg233-client.virtnet environments Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' quota/lquota options: 'hash_lqs_cur_bits=3' loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions oleg233-server: ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' oleg233-server: quota/lquota options: 'hash_lqs_cur_bits=3' Setup mgs, mdt, osts Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 Starting ost2: -o localrecov /dev/mapper/ost2_flakey /mnt/lustre-ost2 seq.cli-lustre-OST0001-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/ost2_flakey Started lustre-OST0001 mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre Starting client oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre Started clients oleg233-client.virtnet: 192.168.202.133@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Using TIMEOUT=20 osc.lustre-OST0000-osc-ffff8800aa6b5000.idle_timeout=debug osc.lustre-OST0001-osc-ffff8800aa6b5000.idle_timeout=debug setting jobstats to procname_uid Setting lustre.sys.jobid_var from disable to procname_uid Waiting 90s for 'procname_uid' Updated after 3s: want 'procname_uid' got 'procname_uid' disable quota as required oleg233-server: Pool lustre.pool1 created oleg233-server: Pool lustre.lustre created oleg233-server: OST lustre-OST0000_UUID added to pool lustre.lustre Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:) Stopping client oleg233-client.virtnet /mnt/lustre opts: Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:) Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server Stopping /mnt/lustre-ost2 (opts:-f) on oleg233-server rename lustre to mylustre checking for existing Lustre data: found Read previous values: Target: lustre-MDT0000 Index: 0 Lustre FS: lustre Mount type: ldiskfs Flags: 0x5 (MDT MGS ) Persistent mount opts: user_xattr,errors=remount-ro Parameters: sys.timeout=20 mdt.identity_upcall=/home/green/git/lustre-release/lustre/utils/l_getidentity Permanent disk data: Target: mylustre-MDT0000 Index: 0 Lustre FS: mylustre Mount type: ldiskfs Flags: 0x5 (MDT MGS ) Persistent mount opts: user_xattr,errors=remount-ro Parameters: sys.timeout=20 mdt.identity_upcall=/home/green/git/lustre-release/lustre/utils/l_getidentity cmd: tune2fs -f -L 'mylustre-MDT0000' '/dev/mapper/mds1_flakey' >/dev/null 2>&1 Writing CONFIGS/mountdata checking for existing Lustre data: found Read previous values: Target: lustre-MDT0001 Index: 1 Lustre FS: lustre Mount type: ldiskfs Flags: 0x1 (MDT ) Persistent mount opts: user_xattr,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 mdt.identity_upcall=/home/green/git/lustre-release/lustre/utils/l_getidentity Permanent disk data: Target: mylustre-MDT0001 Index: 1 Lustre FS: mylustre Mount type: ldiskfs Flags: 0x1 (MDT ) Persistent mount opts: user_xattr,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 mdt.identity_upcall=/home/green/git/lustre-release/lustre/utils/l_getidentity cmd: tune2fs -f -L 'mylustre-MDT0001' '/dev/mapper/mds2_flakey' >/dev/null 2>&1 Writing CONFIGS/mountdata checking for existing Lustre data: found Read previous values: Target: lustre-OST0000 Index: 0 Lustre FS: lustre Mount type: ldiskfs Flags: 0x2 (OST ) Persistent mount opts: ,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 Permanent disk data: Target: mylustre-OST0000 Index: 0 Lustre FS: mylustre Mount type: ldiskfs Flags: 0x2 (OST ) Persistent mount opts: ,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 cmd: tune2fs -f -L 'mylustre-OST0000' '/dev/mapper/ost1_flakey' >/dev/null 2>&1 Writing CONFIGS/mountdata checking for existing Lustre data: found Read previous values: Target: lustre-OST0001 Index: 1 Lustre FS: lustre Mount type: ldiskfs Flags: 0x2 (OST ) Persistent mount opts: ,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 Permanent disk data: Target: mylustre-OST0001 Index: 1 Lustre FS: mylustre Mount type: ldiskfs Flags: 0x2 (OST ) Persistent mount opts: ,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 cmd: tune2fs -f -L 'mylustre-OST0001' '/dev/mapper/ost2_flakey' >/dev/null 2>&1 Writing CONFIGS/mountdata Checking servers environments Checking clients oleg233-client.virtnet environments Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions Setup mgs, mdt, osts Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started mylustre-MDT0000 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started mylustre-MDT0001 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-mylustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started mylustre-OST0000 Starting ost2: -o localrecov /dev/mapper/ost2_flakey /mnt/lustre-ost2 seq.cli-mylustre-OST0001-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started mylustre-OST0001 mount mylustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/mylustre /mnt/lustre Starting client oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/mylustre /mnt/lustre Started clients oleg233-client.virtnet: 192.168.202.133@tcp:/mylustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Using TIMEOUT=20 osc.mylustre-OST0000-osc-ffff8800a907e000.idle_timeout=debug osc.mylustre-OST0001-osc-ffff8800a907e000.idle_timeout=debug disable quota as required File: '/mnt/lustre/d103.conf-sanity/test-framework.sh' Size: 289857 Blocks: 568 IO Block: 4194304 regular file Device: c3aa56ceh/3282720462d Inode: 162129704428503042 Links: 1 Access: (0755/-rwxr-xr-x) Uid: ( 0/ root) Gid: ( 0/ root) Access: 2024-04-19 09:46:41.000000000 -0400 Modify: 2024-04-19 09:46:41.000000000 -0400 Change: 2024-04-19 09:46:41.000000000 -0400 Birth: - Pool: mylustre.pool1 Pool: mylustre.lustre mylustre-OST0000_UUID mylustre-OST0000_UUID oleg233-server: OST mylustre-OST0001_UUID added to pool mylustre.lustre Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:) Stopping client oleg233-client.virtnet /mnt/lustre opts: Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:) Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server Stopping /mnt/lustre-ost2 (opts:-f) on oleg233-server rename mylustre to tfs checking for existing Lustre data: found Read previous values: Target: mylustre-MDT0000 Index: 0 Lustre FS: mylustre Mount type: ldiskfs Flags: 0x5 (MDT MGS ) Persistent mount opts: user_xattr,errors=remount-ro Parameters: sys.timeout=20 mdt.identity_upcall=/home/green/git/lustre-release/lustre/utils/l_getidentity Permanent disk data: Target: tfs-MDT0000 Index: 0 Lustre FS: tfs Mount type: ldiskfs Flags: 0x5 (MDT MGS ) Persistent mount opts: user_xattr,errors=remount-ro Parameters: sys.timeout=20 mdt.identity_upcall=/home/green/git/lustre-release/lustre/utils/l_getidentity cmd: tune2fs -f -L 'tfs-MDT0000' '/dev/mapper/mds1_flakey' >/dev/null 2>&1 Writing CONFIGS/mountdata checking for existing Lustre data: found Read previous values: Target: mylustre-MDT0001 Index: 1 Lustre FS: mylustre Mount type: ldiskfs Flags: 0x1 (MDT ) Persistent mount opts: user_xattr,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 mdt.identity_upcall=/home/green/git/lustre-release/lustre/utils/l_getidentity Permanent disk data: Target: tfs-MDT0001 Index: 1 Lustre FS: tfs Mount type: ldiskfs Flags: 0x1 (MDT ) Persistent mount opts: user_xattr,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 mdt.identity_upcall=/home/green/git/lustre-release/lustre/utils/l_getidentity cmd: tune2fs -f -L 'tfs-MDT0001' '/dev/mapper/mds2_flakey' >/dev/null 2>&1 Writing CONFIGS/mountdata checking for existing Lustre data: found Read previous values: Target: mylustre-OST0000 Index: 0 Lustre FS: mylustre Mount type: ldiskfs Flags: 0x2 (OST ) Persistent mount opts: ,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 Permanent disk data: Target: tfs-OST0000 Index: 0 Lustre FS: tfs Mount type: ldiskfs Flags: 0x2 (OST ) Persistent mount opts: ,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 cmd: tune2fs -f -L 'tfs-OST0000' '/dev/mapper/ost1_flakey' >/dev/null 2>&1 Writing CONFIGS/mountdata checking for existing Lustre data: found Read previous values: Target: mylustre-OST0001 Index: 1 Lustre FS: mylustre Mount type: ldiskfs Flags: 0x2 (OST ) Persistent mount opts: ,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 Permanent disk data: Target: tfs-OST0001 Index: 1 Lustre FS: tfs Mount type: ldiskfs Flags: 0x2 (OST ) Persistent mount opts: ,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 cmd: tune2fs -f -L 'tfs-OST0001' '/dev/mapper/ost2_flakey' >/dev/null 2>&1 Writing CONFIGS/mountdata Checking servers environments Checking clients oleg233-client.virtnet environments Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions Setup mgs, mdt, osts Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started tfs-MDT0000 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started tfs-MDT0001 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-tfs-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started tfs-OST0000 Starting ost2: -o localrecov /dev/mapper/ost2_flakey /mnt/lustre-ost2 seq.cli-tfs-OST0001-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started tfs-OST0001 mount tfs on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/tfs /mnt/lustre Starting client oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/tfs /mnt/lustre Started clients oleg233-client.virtnet: 192.168.202.133@tcp:/tfs on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Using TIMEOUT=20 osc.tfs-OST0000-osc-ffff8800af2ca800.idle_timeout=debug osc.tfs-OST0001-osc-ffff8800af2ca800.idle_timeout=debug disable quota as required File: '/mnt/lustre/d103.conf-sanity/test-framework.sh' Size: 289857 Blocks: 568 IO Block: 4194304 regular file Device: 32e2fa5ah/853736026d Inode: 162129704428503042 Links: 1 Access: (0755/-rwxr-xr-x) Uid: ( 0/ root) Gid: ( 0/ root) Access: 2024-04-19 09:46:41.000000000 -0400 Modify: 2024-04-19 09:46:41.000000000 -0400 Change: 2024-04-19 09:46:41.000000000 -0400 Birth: - Pool: tfs.pool1 Pool: tfs.lustre tfs-OST0000_UUID tfs-OST0001_UUID tfs-OST0000_UUID Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:) Stopping client oleg233-client.virtnet /mnt/lustre opts: Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:) Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server Stopping /mnt/lustre-ost2 (opts:-f) on oleg233-server rename tfs to lustre checking for existing Lustre data: found Read previous values: Target: tfs-MDT0000 Index: 0 Lustre FS: tfs Mount type: ldiskfs Flags: 0x5 (MDT MGS ) Persistent mount opts: user_xattr,errors=remount-ro Parameters: sys.timeout=20 mdt.identity_upcall=/home/green/git/lustre-release/lustre/utils/l_getidentity Permanent disk data: Target: lustre-MDT0000 Index: 0 Lustre FS: lustre Mount type: ldiskfs Flags: 0x5 (MDT MGS ) Persistent mount opts: user_xattr,errors=remount-ro Parameters: sys.timeout=20 mdt.identity_upcall=/home/green/git/lustre-release/lustre/utils/l_getidentity cmd: tune2fs -f -L 'lustre-MDT0000' '/dev/mapper/mds1_flakey' >/dev/null 2>&1 Writing CONFIGS/mountdata checking for existing Lustre data: found Read previous values: Target: tfs-MDT0001 Index: 1 Lustre FS: tfs Mount type: ldiskfs Flags: 0x1 (MDT ) Persistent mount opts: user_xattr,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 mdt.identity_upcall=/home/green/git/lustre-release/lustre/utils/l_getidentity Permanent disk data: Target: lustre-MDT0001 Index: 1 Lustre FS: lustre Mount type: ldiskfs Flags: 0x1 (MDT ) Persistent mount opts: user_xattr,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 mdt.identity_upcall=/home/green/git/lustre-release/lustre/utils/l_getidentity cmd: tune2fs -f -L 'lustre-MDT0001' '/dev/mapper/mds2_flakey' >/dev/null 2>&1 Writing CONFIGS/mountdata checking for existing Lustre data: found Read previous values: Target: tfs-OST0000 Index: 0 Lustre FS: tfs Mount type: ldiskfs Flags: 0x2 (OST ) Persistent mount opts: ,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 Permanent disk data: Target: lustre-OST0000 Index: 0 Lustre FS: lustre Mount type: ldiskfs Flags: 0x2 (OST ) Persistent mount opts: ,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 cmd: tune2fs -f -L 'lustre-OST0000' '/dev/mapper/ost1_flakey' >/dev/null 2>&1 Writing CONFIGS/mountdata checking for existing Lustre data: found Read previous values: Target: tfs-OST0001 Index: 1 Lustre FS: tfs Mount type: ldiskfs Flags: 0x2 (OST ) Persistent mount opts: ,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 Permanent disk data: Target: lustre-OST0001 Index: 1 Lustre FS: lustre Mount type: ldiskfs Flags: 0x2 (OST ) Persistent mount opts: ,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 cmd: tune2fs -f -L 'lustre-OST0001' '/dev/mapper/ost2_flakey' >/dev/null 2>&1 Writing CONFIGS/mountdata Checking servers environments Checking clients oleg233-client.virtnet environments Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions Setup mgs, mdt, osts Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 Starting ost2: -o localrecov /dev/mapper/ost2_flakey /mnt/lustre-ost2 seq.cli-lustre-OST0001-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0001 mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre Starting client oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre Started clients oleg233-client.virtnet: 192.168.202.133@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Using TIMEOUT=20 osc.lustre-OST0000-osc-ffff8800b6f72800.idle_timeout=debug osc.lustre-OST0001-osc-ffff8800b6f72800.idle_timeout=debug disable quota as required PASS 103 (229s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 104a: Make sure user defined options are reflected in mount ========================================================== 09:49:53 (1713534593) mountfsopt: acl,user_xattr Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:-f) Stopping client oleg233-client.virtnet /mnt/lustre opts:-f Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:-f) Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server Stopping /mnt/lustre-ost2 (opts:-f) on oleg233-server oleg233-server: oleg233-server.virtnet: executing set_hostid Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions libkmod: kmod_module_get_holders: could not open '/sys/module/intel_rapl/holders': No such file or directory loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions oleg233-server: libkmod: kmod_module_get_holders: could not open '/sys/module/pcc_cpufreq/holders': No such file or directory Formatting mgs, mds, osts Format mds1: /dev/mapper/mds1_flakey Format mds2: /dev/mapper/mds2_flakey Format ost1: /dev/mapper/ost1_flakey Format ost2: /dev/mapper/ost2_flakey Starting mds1: -o localrecov,noacl /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/mds1_flakey Started lustre-MDT0000 Starting mds2: -o localrecov,noacl /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/mds2_flakey Started lustre-MDT0001 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/ost1_flakey Started lustre-OST0000 Starting ost2: -o localrecov /dev/mapper/ost2_flakey /mnt/lustre-ost2 seq.cli-lustre-OST0001-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/ost2_flakey Started lustre-OST0001 mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre setfacl: /mnt/lustre: Operation not supported PASS 104a (78s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 104b: Mount uses last flock argument ========================================================== 09:51:13 (1713534673) mount lustre with opts flock,localflock on /mnt/lustre3..... Starting client: oleg233-client.virtnet: -o flock,localflock oleg233-server@tcp:/lustre /mnt/lustre3 192.168.202.133@tcp:/lustre on /mnt/lustre3 type lustre (rw,checksum,localflock,nouser_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) umount lustre on /mnt/lustre3..... Stopping client oleg233-client.virtnet /mnt/lustre3 (opts:) mount lustre with opts localflock,flock on /mnt/lustre3..... Starting client: oleg233-client.virtnet: -o localflock,flock oleg233-server@tcp:/lustre /mnt/lustre3 192.168.202.133@tcp:/lustre on /mnt/lustre3 type lustre (rw,checksum,flock,nouser_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) umount lustre on /mnt/lustre3..... Stopping client oleg233-client.virtnet /mnt/lustre3 (opts:) mount lustre with opts localflock,flock,noflock on /mnt/lustre3..... Starting client: oleg233-client.virtnet: -o localflock,flock,noflock oleg233-server@tcp:/lustre /mnt/lustre3 umount lustre on /mnt/lustre3..... Stopping client oleg233-client.virtnet /mnt/lustre3 (opts:) PASS 104b (3s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 105: check file creation for ro and rw bind mnt pt ========================================================== 09:51:18 (1713534678) umount lustre on /mnt/lustre..... Stopping client oleg233-client.virtnet /mnt/lustre (opts:-f) stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server unloading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing unload_modules_local oleg233-server: rmmod: ERROR: Module lustre is in use pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 modules unloaded. Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:-f) Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:-f) Stopping /mnt/lustre-ost2 (opts:-f) on oleg233-server oleg233-server: oleg233-server.virtnet: executing set_hostid Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions ../libcfs/libcfs/libcfs options: 'cpu_npartitions=2' ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' quota/lquota options: 'hash_lqs_cur_bits=3' loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions oleg233-server: libkmod: kmod_module_get_holders: could not open '/sys/module/acpi_cpufreq/holders': No such file or directory oleg233-server: libkmod: kmod_module_get_holders: could not open '/sys/module/intel_rapl/holders': No such file or directory Formatting mgs, mds, osts Format mds1: /dev/mapper/mds1_flakey Format mds2: /dev/mapper/mds2_flakey Format ost1: /dev/mapper/ost1_flakey Format ost2: /dev/mapper/ost2_flakey start mds service on oleg233-server Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/mds1_flakey Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/mds2_flakey Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/ost1_flakey Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre touch: cannot touch '/tmp/d105.conf-sanity/f105.conf-sanity': Read-only file system umount lustre on /mnt/lustre..... Stopping client oleg233-client.virtnet /mnt/lustre (opts:) stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server unloading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing unload_modules_local modules unloaded. pdsh@oleg233-client: oleg233-client: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-client: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 PASS 105 (101s) error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='0': No such file or directory oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='0': No such file or directory SKIP: conf-sanity test_106 skipping SLOW test 106 error: get_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory oleg233-server: error: get_param: param_path 'debug_raw_pointers': No such file or directory pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory == conf-sanity test 107: Unknown config param should not fail target mounting ========================================================== 09:53:02 (1713534782) start mds service on oleg233-server Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' quota/lquota options: 'hash_lqs_cur_bits=3' loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions oleg233-server: ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' oleg233-server: quota/lquota options: 'hash_lqs_cur_bits=3' Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server unloading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing unload_modules_local modules unloaded. Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions ../libcfs/libcfs/libcfs options: 'cpu_npartitions=2' ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' quota/lquota options: 'hash_lqs_cur_bits=3' loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions oleg233-server: libkmod: kmod_module_get_holders: could not open '/sys/module/pcc_cpufreq/holders': No such file or directory oleg233-server: ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' oleg233-server: quota/lquota options: 'hash_lqs_cur_bits=3' start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid umount lustre on /mnt/lustre..... stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server stop mds service on oleg233-server stop mds service on oleg233-server unloading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing unload_modules_local modules unloaded. pdsh@oleg233-client: oleg233-client: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-client: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 PASS 107 (266s) error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='0': No such file or directory oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='0': No such file or directory error: get_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory oleg233-server: error: get_param: param_path 'debug_raw_pointers': No such file or directory pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory == conf-sanity test 108a: migrate from ldiskfs to ZFS ==== 09:57:30 (1713535050) SKIP: conf-sanity test_108a zfs only test SKIP 108a (1s) error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='0': No such file or directory oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='0': No such file or directory error: get_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory oleg233-server: error: get_param: param_path 'debug_raw_pointers': No such file or directory pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory == conf-sanity test 108b: migrate from ZFS to ldiskfs ==== 09:57:34 (1713535054) Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:) Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:) Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' quota/lquota options: 'hash_lqs_cur_bits=3' loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions oleg233-server: ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' oleg233-server: quota/lquota options: 'hash_lqs_cur_bits=3' oleg233-server: 1+0 records in oleg233-server: 1+0 records out oleg233-server: 1048576 bytes (1.0 MB) copied, 0.00393205 s, 267 MB/s oleg233-server: 1+0 records in oleg233-server: 1+0 records out oleg233-server: 1048576 bytes (1.0 MB) copied, 0.0038755 s, 271 MB/s oleg233-server: 1+0 records in oleg233-server: 1+0 records out oleg233-server: 1048576 bytes (1.0 MB) copied, 0.003797 s, 276 MB/s oleg233-server: 1+0 records in oleg233-server: 1+0 records out oleg233-server: 1048576 bytes (1.0 MB) copied, 0.00396594 s, 264 MB/s Permanent disk data: Target: lustre-MDT0000 Index: 0 Lustre FS: lustre Mount type: ldiskfs Flags: 0x45 (MDT MGS update ) Persistent mount opts: user_xattr,errors=remount-ro Parameters: formatting backing filesystem ldiskfs on /dev/loop0 target name lustre-MDT0000 kilobytes 200000 options -I 1024 -i 2560 -q -O uninit_bg,^extents,dirdata,dir_nlink,quota,project,huge_file,ea_inode,large_dir,^fast_commit,flex_bg -E lazy_journal_init,lazy_itable_init="0",packed_meta_blocks -F mkfs_cmd = mke2fs -j -b 4096 -L lustre-MDT0000 -I 1024 -i 2560 -q -O uninit_bg,^extents,dirdata,dir_nlink,quota,project,huge_file,ea_inode,large_dir,^fast_commit,flex_bg -E lazy_journal_init,lazy_itable_init="0",packed_meta_blocks -F /dev/loop0 200000k Writing CONFIGS/mountdata Permanent disk data: Target: lustre-MDT0001 Index: 1 Lustre FS: lustre Mount type: ldiskfs Flags: 0x41 (MDT update ) Persistent mount opts: user_xattr,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp formatting backing filesystem ldiskfs on /dev/loop0 target name lustre-MDT0001 kilobytes 200000 options -I 1024 -i 2560 -q -O uninit_bg,^extents,dirdata,dir_nlink,quota,project,huge_file,ea_inode,large_dir,^fast_commit,flex_bg -E lazy_journal_init,lazy_itable_init="0",packed_meta_blocks -F mkfs_cmd = mke2fs -j -b 4096 -L lustre-MDT0001 -I 1024 -i 2560 -q -O uninit_bg,^extents,dirdata,dir_nlink,quota,project,huge_file,ea_inode,large_dir,^fast_commit,flex_bg -E lazy_journal_init,lazy_itable_init="0",packed_meta_blocks -F /dev/loop0 200000k Writing CONFIGS/mountdata Permanent disk data: Target: lustre-OST0000 Index: 0 Lustre FS: lustre Mount type: ldiskfs Flags: 0x42 (OST update ) Persistent mount opts: ,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp formatting backing filesystem ldiskfs on /dev/loop0 target name lustre-OST0000 kilobytes 200000 options -I 512 -q -O uninit_bg,extents,dir_nlink,quota,project,huge_file,large_dir,^fast_commit,flex_bg -G 256 -E resize="4290772992",lazy_journal_init,lazy_itable_init="0",packed_meta_blocks -F mkfs_cmd = mke2fs -j -b 4096 -L lustre-OST0000 -I 512 -q -O uninit_bg,extents,dir_nlink,quota,project,huge_file,large_dir,^fast_commit,flex_bg -G 256 -E resize="4290772992",lazy_journal_init,lazy_itable_init="0",packed_meta_blocks -F /dev/loop0 200000k Writing CONFIGS/mountdata Permanent disk data: Target: lustre-OST0001 Index: 1 Lustre FS: lustre Mount type: ldiskfs Flags: 0x42 (OST update ) Persistent mount opts: ,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp formatting backing filesystem ldiskfs on /dev/loop0 target name lustre-OST0001 kilobytes 200000 options -I 512 -q -O uninit_bg,extents,dir_nlink,quota,project,huge_file,large_dir,^fast_commit,flex_bg -G 256 -E resize="4290772992",lazy_journal_init,lazy_itable_init="0",packed_meta_blocks -F mkfs_cmd = mke2fs -j -b 4096 -L lustre-OST0001 -I 512 -q -O uninit_bg,extents,dir_nlink,quota,project,huge_file,large_dir,^fast_commit,flex_bg -G 256 -E resize="4290772992",lazy_journal_init,lazy_itable_init="0",packed_meta_blocks -F /dev/loop0 200000k Writing CONFIGS/mountdata changing server nid... mounting mdt1 from backup... mounting mdt2 from backup... mounting ost1 from backup... mounting ost2 from backup... Started LFSCK on the device lustre-MDT0000: scrub Started LFSCK on the device lustre-MDT0001: scrub Started LFSCK on the device lustre-OST0000: scrub Started LFSCK on the device lustre-OST0001: scrub mounting client... check list total 12 drwxr-xr-x 2 root root 4096 Jan 20 2018 d1 -rw-r--r-- 1 root root 0 Jan 20 2018 f0 -rw-r--r-- 1 root root 4067 Jan 20 2018 README -rw-r--r-- 1 root root 331 Jan 20 2018 regression check truncate && write check create check read && write && append verify data done. cleanup... PASS 108b (78s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 109a: test lctl clear_conf fsname ==== 09:58:53 (1713535133) Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:) Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:) Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:-f) Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:-f) oleg233-server: oleg233-server.virtnet: executing set_hostid Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions libkmod: kmod_module_get_holders: could not open '/sys/module/acpi_cpufreq/holders': No such file or directory loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions Formatting mgs, mds, osts Format mds1: /dev/mapper/mds1_flakey Format mds2: /dev/mapper/mds2_flakey Format ost1: /dev/mapper/ost1_flakey Format ost2: /dev/mapper/ost2_flakey start mds service on oleg233-server Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/mds1_flakey Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/mds2_flakey Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/ost1_flakey Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre Setting lustre-MDT0000.mdd.atime_diff from 60 to 62 Waiting 90s for '62' Updated after 4s: want '62' got '62' Setting lustre-MDT0000.mdd.atime_diff from 62 to 63 Waiting 90s for '63' Updated after 8s: want '63' got '63' Setting lustre.llite.max_read_ahead_mb from 256 to 32 Waiting 90s for '32' Updated after 2s: want '32' got '32' Setting lustre.llite.max_read_ahead_mb from 32 to 64 Waiting 90s for '64' Updated after 6s: want '64' got '64' oleg233-server: Pool lustre.pool1 created Waiting 90s for '' oleg233-server: OST lustre-OST0000_UUID added to pool lustre.pool1 oleg233-server: OST lustre-OST0000_UUID removed from pool lustre.pool1 oleg233-server: OST lustre-OST0000_UUID added to pool lustre.pool1 umount lustre on /mnt/lustre..... Stopping client oleg233-client.virtnet /mnt/lustre (opts:) stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server start mds service on oleg233-server Starting mds1: -o localrecov -o nosvc /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all Start /dev/mapper/mds1_flakey without service Started lustre-MDT0000 oleg233-server: debugfs 1.46.2.wc5 (26-Mar-2022) oleg233-server: /dev/mapper/mds1_flakey: catastrophic mode - not reading inode or group bitmaps stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server oleg233-server: debugfs 1.46.2.wc5 (26-Mar-2022) oleg233-server: /dev/mapper/mds1_flakey: catastrophic mode - not reading inode or group bitmaps start mds service on oleg233-server Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre Destroy the created pools: pool1 lustre.pool1 oleg233-server: OST lustre-OST0000_UUID removed from pool lustre.pool1 oleg233-server: Pool lustre.pool1 destroyed umount lustre on /mnt/lustre..... Stopping client oleg233-client.virtnet /mnt/lustre (opts:) stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server unloading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing unload_modules_local modules unloaded. pdsh@oleg233-client: oleg233-client: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-client: ssh exited with exit code 2 PASS 109a (157s) error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='0': No such file or directory oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='0': No such file or directory error: get_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory oleg233-server: error: get_param: param_path 'debug_raw_pointers': No such file or directory pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory == conf-sanity test 109b: test lctl clear_conf one config ========================================================== 10:01:32 (1713535292) Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:) Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:) Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:-f) Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:-f) pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 oleg233-server: oleg233-server.virtnet: executing set_hostid Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' quota/lquota options: 'hash_lqs_cur_bits=3' loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions oleg233-server: libkmod: kmod_module_get_holders: could not open '/sys/module/acpi_cpufreq/holders': No such file or directory oleg233-server: libkmod: kmod_module_get_holders: could not open '/sys/module/pcc_cpufreq/holders': No such file or directory oleg233-server: ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' oleg233-server: quota/lquota options: 'hash_lqs_cur_bits=3' Formatting mgs, mds, osts Format mds1: /dev/mapper/mds1_flakey Format mds2: /dev/mapper/mds2_flakey Format ost1: /dev/mapper/ost1_flakey Format ost2: /dev/mapper/ost2_flakey start mds service on oleg233-server Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/mds1_flakey Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/mds2_flakey Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/ost1_flakey Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre Setting lustre-MDT0000.mdd.atime_diff from 60 to 62 Waiting 90s for '62' Setting lustre-MDT0000.mdd.atime_diff from 62 to 63 Waiting 90s for '63' Updated after 7s: want '63' got '63' Setting lustre.llite.max_read_ahead_mb from 256 to 32 Waiting 90s for '32' Updated after 5s: want '32' got '32' Setting lustre.llite.max_read_ahead_mb from 32 to 64 Waiting 90s for '64' Updated after 7s: want '64' got '64' oleg233-server: Pool lustre.pool1 created oleg233-server: OST lustre-OST0000_UUID added to pool lustre.pool1 oleg233-server: OST lustre-OST0000_UUID removed from pool lustre.pool1 oleg233-server: OST lustre-OST0000_UUID added to pool lustre.pool1 umount lustre on /mnt/lustre..... Stopping client oleg233-client.virtnet /mnt/lustre (opts:) stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server start mds service on oleg233-server Starting mds1: -o localrecov -o nosvc /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all Start /dev/mapper/mds1_flakey without service Started lustre-MDT0000 oleg233-server: debugfs 1.46.2.wc5 (26-Mar-2022) oleg233-server: /dev/mapper/mds1_flakey: catastrophic mode - not reading inode or group bitmaps stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server oleg233-server: debugfs 1.46.2.wc5 (26-Mar-2022) oleg233-server: /dev/mapper/mds1_flakey: catastrophic mode - not reading inode or group bitmaps start mds service on oleg233-server Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre Destroy the created pools: pool1 lustre.pool1 oleg233-server: OST lustre-OST0000_UUID removed from pool lustre.pool1 oleg233-server: Pool lustre.pool1 destroyed umount lustre on /mnt/lustre..... Stopping client oleg233-client.virtnet /mnt/lustre (opts:) stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server unloading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing unload_modules_local modules unloaded. pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-client: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-client: ssh exited with exit code 2 PASS 109b (212s) error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='0': No such file or directory oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='0': No such file or directory SKIP: conf-sanity test_110 skipping ALWAYS excluded test 110 SKIP: conf-sanity test_111 skipping SLOW test 111 error: get_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory oleg233-server: error: get_param: param_path 'debug_raw_pointers': No such file or directory pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory == conf-sanity test 112a: mount OST with no_create option ========================================================== 10:05:07 (1713535507) start mds service on oleg233-server Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' quota/lquota options: 'hash_lqs_cur_bits=3' loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions oleg233-server: ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' oleg233-server: quota/lquota options: 'hash_lqs_cur_bits=3' Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid start ost2 service on oleg233-server Starting ost2: -o localrecov,no_create /dev/mapper/ost2_flakey /mnt/lustre-ost2 seq.cli-lustre-OST0001-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/ost2_flakey Started lustre-OST0001 mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre oleg233-server: oleg233-server.virtnet: executing wait_import_state FULL os[cp].lustre-OST0000-osc-MDT0000.ost_server_uuid 50 oleg233-server: os[cp].lustre-OST0000-osc-MDT0000.ost_server_uuid in FULL state after 0 sec oleg233-client.virtnet: executing wait_import_state (FULL|IDLE) osc.lustre-OST0000-osc-ffff8800a9079800.ost_server_uuid 50 osc.lustre-OST0000-osc-ffff8800a9079800.ost_server_uuid in FULL state after 0 sec oleg233-client.virtnet: executing wait_import_state (FULL|IDLE) osc.lustre-OST0001-osc-ffff8800a9079800.ost_server_uuid 50 osc.lustre-OST0001-osc-ffff8800a9079800.ost_server_uuid in FULL state after 0 sec /mnt/lustre/f112a.conf-sanity.1 lmm_stripe_count: 1 lmm_stripe_size: 4194304 lmm_pattern: raid0 lmm_layout_gen: 0 lmm_stripe_offset: 0 obdidx objid objid group 0 67 0x43 0x280000401 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 95248 1704 84888 2% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 95248 1540 85052 2% /mnt/lustre[MDT:1] lustre-OST0000_UUID 142216 1528 126688 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 142216 1396 126820 2% /mnt/lustre[OST:1] N filesystem_summary: 284432 2924 253508 2% /mnt/lustre obdfilter.lustre-OST0001.no_create=0 stop ost2 service on oleg233-server Stopping /mnt/lustre-ost2 (opts:-f) on oleg233-server umount lustre on /mnt/lustre..... Stopping client oleg233-client.virtnet /mnt/lustre (opts:) stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server unloading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing unload_modules_local modules unloaded. pdsh@oleg233-client: oleg233-client: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-client: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 PASS 112a (74s) error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='0': No such file or directory oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='0': No such file or directory error: get_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory oleg233-server: error: get_param: param_path 'debug_raw_pointers': No such file or directory pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory == conf-sanity test 112b: mount MDT with no_create option ========================================================== 10:06:22 (1713535582) start mds service on oleg233-server Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions libkmod: kmod_module_get_holders: could not open '/sys/module/pcc_cpufreq/holders': No such file or directory ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' quota/lquota options: 'hash_lqs_cur_bits=3' loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions oleg233-server: ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' oleg233-server: quota/lquota options: 'hash_lqs_cur_bits=3' Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid start mds service on oleg233-server Starting mds2: -o localrecov -o no_create /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid start ost2 service on oleg233-server Starting ost2: -o localrecov /dev/mapper/ost2_flakey /mnt/lustre-ost2 seq.cli-lustre-OST0001-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0001 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0001-osc-[-0-9a-f]*.ost_server_uuid mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre oleg233-server: oleg233-server.virtnet: executing wait_import_state (FULL|IDLE) os[cp].lustre-OST0001-osc-MDT0001.ost_server_uuid 50 oleg233-server: os[cp].lustre-OST0001-osc-MDT0001.ost_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 95248 1704 84888 2% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 95248 1544 85048 2% /mnt/lustre[MDT:1] N lustre-OST0000_UUID 142216 1532 126684 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 142216 1532 126684 2% /mnt/lustre[OST:1] filesystem_summary: 284432 3064 253368 2% /mnt/lustre 100 0 mdt.lustre-MDT0001.no_create=0 1 0 99 1 stop ost2 service on oleg233-server Stopping /mnt/lustre-ost2 (opts:-f) on oleg233-server umount lustre on /mnt/lustre..... Stopping client oleg233-client.virtnet /mnt/lustre (opts:) stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server unloading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing unload_modules_local modules unloaded. pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-client: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-client: ssh exited with exit code 2 PASS 112b (139s) error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='0': No such file or directory oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='0': No such file or directory error: get_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory oleg233-server: error: get_param: param_path 'debug_raw_pointers': No such file or directory pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory == conf-sanity test 113: Shadow mountpoint correctly report ro/rw for mounts ========================================================== 10:08:43 (1713535723) Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:) Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:) Checking servers environments Checking clients oleg233-client.virtnet environments Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions libkmod: kmod_module_get_holders: could not open '/sys/module/acpi_cpufreq/holders': No such file or directory ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' quota/lquota options: 'hash_lqs_cur_bits=3' loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions oleg233-server: ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' oleg233-server: quota/lquota options: 'hash_lqs_cur_bits=3' Setup mgs, mdt, osts Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 Starting ost2: -o localrecov /dev/mapper/ost2_flakey /mnt/lustre-ost2 seq.cli-lustre-OST0001-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0001 mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre Starting client oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre Started clients oleg233-client.virtnet: 192.168.202.133@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Using TIMEOUT=20 osc.lustre-OST0000-osc-ffff8800a907a000.idle_timeout=debug osc.lustre-OST0001-osc-ffff8800a907a000.idle_timeout=debug setting jobstats to procname_uid Setting lustre.sys.jobid_var from disable to procname_uid Waiting 90s for 'procname_uid' Updated after 6s: want 'procname_uid' got 'procname_uid' disable quota as required /dev/mapper/mds1_flakey on /mnt/lustre-mds1 type lustre (rw,svname=lustre-MDT0000,mgs,osd=osd-ldiskfs,user_xattr,errors=remount-ro) /dev/mapper/mds2_flakey on /mnt/lustre-mds2 type lustre (rw,svname=lustre-MDT0001,mgsnode=192.168.202.133@tcp,osd=osd-ldiskfs) /dev/mapper/ost1_flakey on /mnt/lustre-ost1 type lustre (rw,svname=lustre-OST0000,mgsnode=192.168.202.133@tcp,osd=osd-ldiskfs) /dev/mapper/ost2_flakey on /mnt/lustre-ost2 type lustre (rw,svname=lustre-OST0001,mgsnode=192.168.202.133@tcp,osd=osd-ldiskfs) /dev/mapper/ost1_flakey on /mnt/lustre-ost1 type lustre (rw,svname=lustre-OST0000,mgsnode=192.168.202.133@tcp,osd=osd-ldiskfs) /dev/mapper/ost2_flakey on /mnt/lustre-ost2 type lustre (rw,svname=lustre-OST0001,mgsnode=192.168.202.133@tcp,osd=osd-ldiskfs) Shadow Mountpoint correctly reports rw for ldiskfs Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:) Stopping client oleg233-client.virtnet /mnt/lustre opts: Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:) Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server Stopping /mnt/lustre-ost2 (opts:-f) on oleg233-server Checking servers environments Checking clients oleg233-client.virtnet environments Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions libkmod: kmod_module_get_holders: could not open '/sys/module/intel_rapl/holders': No such file or directory loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions oleg233-server: libkmod: kmod_module_get_holders: could not open '/sys/module/pcc_cpufreq/holders': No such file or directory oleg233-server: libkmod: kmod_module_get_holders: could not open '/sys/module/acpi_cpufreq/holders': No such file or directory Setup mgs, mdt, osts Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 Starting ost2: -o localrecov /dev/mapper/ost2_flakey /mnt/lustre-ost2 seq.cli-lustre-OST0001-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0001 mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre Starting client oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre Started clients oleg233-client.virtnet: 192.168.202.133@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Using TIMEOUT=20 osc.lustre-OST0000-osc-ffff8800b3f01800.idle_timeout=debug osc.lustre-OST0001-osc-ffff8800b3f01800.idle_timeout=debug disable quota as required Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:) Stopping client oleg233-client.virtnet /mnt/lustre opts: Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:) Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server Stopping /mnt/lustre-ost2 (opts:-f) on oleg233-server PASS 113 (147s) debug_raw_pointers=0 debug_raw_pointers=0 SKIP: conf-sanity test_114 skipping SLOW test 114 SKIP: conf-sanity test_115 skipping excluded test 115 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 116: big size MDT support ============ 10:11:13 (1713535873) /usr/sbin/mkfs.xfs Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:) Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:) Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions oleg233-server: libkmod: kmod_module_get_holders: could not open '/sys/module/pcc_cpufreq/holders': No such file or directory meta-data=/tmp/f116.conf-sanity-mdt0 isize=512 agcount=4, agsize=67108864 blks = sectsz=512 attr=2, projid32bit=1 = crc=1 finobt=0, sparse=0 data = bsize=4096 blocks=268435456, imaxpct=5 = sunit=0 swidth=0 blks naming =version 2 bsize=4096 ascii-ci=0 ftype=1 log =internal log bsize=4096 blocks=131072, version=2 = sectsz=512 sunit=0 blks, lazy-count=1 realtime =none extsz=4096 blocks=0, rtextents=0 Permanent disk data: Target: lustre:MDT0000 Index: 0 Lustre FS: lustre Mount type: ldiskfs Flags: 0x61 (MDT first_time update ) Persistent mount opts: user_xattr,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 mdt.identity_upcall=/home/green/git/lustre-release/lustre/utils/l_getidentity checking for existing Lustre data: not found formatting backing filesystem ldiskfs on /dev/loop1 target name lustre:MDT0000 kilobytes 18253611008 options -i 16777216 -b 4096 -J size=4096 -I 1024 -q -O uninit_bg,extents,dirdata,dir_nlink,quota,project,huge_file,64bit,^resize_inode,ea_inode,large_dir,^fast_commit,flex_bg -E lazy_itable_init,lazy_journal_init,lazy_itable_init,packed_meta_blocks -F mkfs_cmd = mke2fs -j -b 4096 -L lustre:MDT0000 -i 16777216 -b 4096 -J size=4096 -I 1024 -q -O uninit_bg,extents,dirdata,dir_nlink,quota,project,huge_file,64bit,^resize_inode,ea_inode,large_dir,^fast_commit,flex_bg -E lazy_itable_init,lazy_journal_init,lazy_itable_init,packed_meta_blocks -F /dev/loop1 18253611008k Writing CONFIGS/mountdata Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:-f) Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:-f) oleg233-server: oleg233-server.virtnet: executing set_hostid Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions Formatting mgs, mds, osts Format mds1: /dev/mapper/mds1_flakey Format mds2: /dev/mapper/mds2_flakey Format ost1: /dev/mapper/ost1_flakey Format ost2: /dev/mapper/ost2_flakey start mds service on oleg233-server Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/mds1_flakey Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/mds2_flakey Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/ost1_flakey Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid oleg233-server: oleg233-server.virtnet: executing wait_import_state FULL os[cp].lustre-OST0000-osc-MDT0000.ost_server_uuid 50 oleg233-server: os[cp].lustre-OST0000-osc-MDT0000.ost_server_uuid in FULL state after 0 sec oleg233-server: oleg233-server.virtnet: executing wait_import_state FULL os[cp].lustre-OST0000-osc-MDT0001.ost_server_uuid 50 oleg233-server: os[cp].lustre-OST0000-osc-MDT0001.ost_server_uuid in FULL state after 0 sec stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server PASS 116 (68s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 117: lctl get_param return errors properly ========================================================== 10:12:23 (1713535943) start mds service on oleg233-server Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre ost.OSS.ost_io.nrs_policies=fifo oleg233-server: error: read_param: '/sys/kernel/debug/lustre/ost/OSS/ost_io/nrs_tbf_rule': No such device pdsh@oleg233-client: oleg233-server: ssh exited with exit code 19 umount lustre on /mnt/lustre..... Stopping client oleg233-client.virtnet /mnt/lustre (opts:) stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server unloading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing unload_modules_local modules unloaded. pdsh@oleg233-client: oleg233-client: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-client: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 PASS 117 (35s) error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='0': No such file or directory oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='0': No such file or directory error: get_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory oleg233-server: error: get_param: param_path 'debug_raw_pointers': No such file or directory pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory == conf-sanity test 119: writeconf on slave mdt shouldn't duplicate mdc/osp and crash ========================================================== 10:13:01 (1713535981) oleg233-server: error: get_param: param_path 'debug': No such file or directory pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 oleg233-server: error: set_param: param_path 'debug': No such file or directory oleg233-server: error: set_param: setting 'debug'='+config': No such file or directory pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 oleg233-server: opening /dev/lnet failed: No such file or directory oleg233-server: hint: the kernel modules may not be loaded oleg233-server: IOC_LIBCFS_CLEAR_DEBUG failed: No such file or directory pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 start mds service on oleg233-server Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' quota/lquota options: 'hash_lqs_cur_bits=3' loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions oleg233-server: ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' oleg233-server: quota/lquota options: 'hash_lqs_cur_bits=3' Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server debug_mb=84 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 300s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 290s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 280s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 270s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 260s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 250s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 240s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 230s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 210s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 200s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 190s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 180s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 170s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 160s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 150s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 140s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 130s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 120s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 110s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 100s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 90s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 80s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 70s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 60s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 50s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 30s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 10s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Update not seen after 300s: want '1' got '0' stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server debug_mb=84 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 300s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 290s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 280s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 270s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 260s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 240s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 220s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 210s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 200s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 190s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 180s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 170s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 160s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 150s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 130s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 110s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 100s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 90s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 80s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 70s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 60s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 40s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 20s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 0s for '1' Update not seen after 300s: want '1' got '0' stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server debug_mb=84 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 300s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 290s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 270s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 250s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 240s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 230s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 220s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 210s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 200s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 180s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 160s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 150s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 140s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 130s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 110s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 90s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 70s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 60s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 50s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 40s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 30s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 20s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 10s for '1' pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Waiting 0s for '1' Update not seen after 300s: want '1' got '0' debug_mb=21 debug_mb=21 debug=-config Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:-f) Stopping client oleg233-client.virtnet /mnt/lustre opts:-f Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:-f) Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server oleg233-server: oleg233-server.virtnet: executing set_hostid Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions Formatting mgs, mds, osts Format mds1: /dev/mapper/mds1_flakey Format mds2: /dev/mapper/mds2_flakey Format ost1: /dev/mapper/ost1_flakey Format ost2: /dev/mapper/ost2_flakey PASS 119 (1013s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 120: cross-target rename should not create bad symlinks ========================================================== 10:29:56 (1713536996) start mds service on oleg233-server Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/mds1_flakey Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/mds2_flakey Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/ost1_flakey Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:) Stopping client oleg233-client.virtnet /mnt/lustre opts: Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:) Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server e2fsck -d -v -t -t -f -n /dev/mapper/mds1_flakey -m8 oleg233-server: e2fsck 1.46.2.wc5 (26-Mar-2022) oleg233-server: Use max possible thread num: 1 instead Pass 1: Checking inodes, blocks, and sizes [Thread 0] Scan group range [0, 3) [Thread 0] jumping to group 0 [Thread 0] e2fsck_pass1_run:2564: increase inode 81 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 82 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 83 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 84 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 85 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 86 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 87 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 88 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 89 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 90 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 91 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 92 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 93 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 94 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 95 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 96 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 97 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 98 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 99 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 100 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 101 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 102 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 103 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 104 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 105 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 106 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 107 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 108 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 109 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 110 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 111 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 112 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 113 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 114 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 115 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 116 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 117 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 118 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 119 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 120 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 121 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 122 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 123 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 124 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 125 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 126 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 127 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 128 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 129 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 130 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 131 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 132 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 133 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 134 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 135 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 136 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 137 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 138 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 139 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 140 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 141 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 142 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 143 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 144 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 145 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 146 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 147 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 148 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 149 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 150 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 151 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 152 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 153 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 154 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 155 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 156 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 157 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 158 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 159 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 160 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 162 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 163 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 164 badness 0 to 2 for 10084 [Thread 0] group 1 finished [Thread 0] e2fsck_pass1_run:2564: increase inode 26697 badness 0 to 2 for 10084 [Thread 0] group 2 finished [Thread 0] e2fsck_pass1_run:2564: increase inode 53372 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 53373 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 53374 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 53375 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 53376 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 53378 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 53379 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 53380 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 53381 badness 0 to 2 for 10084 [Thread 0] e2fsck_pass1_run:2564: increase inode 53382 badness 0 to 2 for 10084 [Thread 0] group 3 finished [Thread 0] Pass 1: Memory used: 268k/0k (140k/129k), time: 0.00/ 0.00/ 0.00 [Thread 0] Pass 1: I/O read: 1MB, write: 0MB, rate: 203.21MB/s [Thread 0] Scanned group range [0, 3), inodes 280 Pass 2: Checking directory structure Pass 2: Memory used: 268k/0k (97k/172k), time: 0.01/ 0.00/ 0.00 Pass 2: I/O read: 1MB, write: 0MB, rate: 179.79MB/s Pass 3: Checking directory connectivity Peak memory: Memory used: 268k/0k (97k/172k), time: 0.02/ 0.01/ 0.01 Pass 3: Memory used: 268k/0k (96k/173k), time: 0.00/ 0.00/ 0.00 Pass 3: I/O read: 0MB, write: 0MB, rate: 0.00MB/s Pass 4: Checking reference counts Pass 4: Memory used: 268k/0k (67k/202k), time: 0.00/ 0.00/ 0.00 Pass 4: I/O read: 0MB, write: 0MB, rate: 0.00MB/s Pass 5: Checking group summary information Pass 5: Memory used: 268k/0k (67k/202k), time: 0.01/ 0.00/ 0.00 Pass 5: I/O read: 1MB, write: 0MB, rate: 187.58MB/s 279 inodes used (0.35%, out of 79992) 5 non-contiguous files (1.8%) 0 non-contiguous directories (0.0%) # of inodes with ind/dind/tind blocks: 0/0/0 24583 blocks used (49.17%, out of 50000) 0 bad blocks 1 large file 149 regular files 119 directories 0 character device files 0 block device files 0 fifos 0 links 1 symbolic link (1 fast symbolic link) 0 sockets ------------ 269 files Memory used: 268k/0k (66k/203k), time: 0.03/ 0.02/ 0.01 I/O read: 1MB, write: 0MB, rate: 32.56MB/s PASS 120 (46s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 121: failover MGS ==================== 10:30:45 (1713537045) Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:) Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:) start mds service on oleg233-server Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid Failing mgs on oleg233-server Stopping /mnt/lustre-mds1 (opts:) on oleg233-server 10:31:01 (1713537061) shut down Failover mgs to oleg233-server mount facets: mgs Starting mgs: /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 10:31:15 (1713537075) targets are mounted 10:31:15 (1713537075) facet_failover done pdsh@oleg233-client: oleg233-client: ssh exited with exit code 95 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mgc.*.mgs_server_uuid pdsh@oleg233-client: oleg233-client: ssh exited with exit code 95 stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server start mds service on oleg233-server Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid Failing mgs on oleg233-server Stopping /mnt/lustre-mds1 (opts:) on oleg233-server 10:31:39 (1713537099) shut down Failover mgs to oleg233-server mount facets: mgs Starting mgs: /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 10:31:54 (1713537114) targets are mounted 10:31:54 (1713537114) facet_failover done pdsh@oleg233-client: oleg233-client: ssh exited with exit code 95 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mgc.*.mgs_server_uuid pdsh@oleg233-client: oleg233-client: ssh exited with exit code 95 stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server PASS 121 (83s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 122a: Check OST sequence update ====== 10:32:10 (1713537130) Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:-f) Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:-f) oleg233-server: oleg233-server.virtnet: executing set_hostid Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions Formatting mgs, mds, osts Format mds1: /dev/mapper/mds1_flakey Format mds2: /dev/mapper/mds2_flakey Format ost1: /dev/mapper/ost1_flakey Format ost2: /dev/mapper/ost2_flakey Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions fail_loc=0x00001e0 start mds service on oleg233-server Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/mds1_flakey Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/mds2_flakey Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/ost1_flakey Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre fail_loc=0 total: 1000 open/close in 4.31 seconds: 231.89 ops/second umount lustre on /mnt/lustre..... Stopping client oleg233-client.virtnet /mnt/lustre (opts:) stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server unloading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing unload_modules_local modules unloaded. pdsh@oleg233-client: oleg233-client: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-client: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 PASS 122a (85s) error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='0': No such file or directory oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='0': No such file or directory error: get_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory oleg233-server: error: get_param: param_path 'debug_raw_pointers': No such file or directory pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory == conf-sanity test 123aa: llog_print works with FIDs and simple names ========================================================== 10:33:38 (1713537218) start mds service on oleg233-server Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' quota/lquota options: 'hash_lqs_cur_bits=3' loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions oleg233-server: ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' oleg233-server: quota/lquota options: 'hash_lqs_cur_bits=3' Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre 1 UP mgs MGS MGS 7 - { index: 2, event: attach, device: lustre-clilov, type: lov, UUID: lustre-clilov_UUID } - { index: 3, event: setup, device: lustre-clilov, UUID: } - { index: 6, event: attach, device: lustre-clilmv, type: lmv, UUID: lustre-clilmv_UUID } - { index: 7, event: setup, device: lustre-clilmv, UUID: } - { index: 10, event: new_profile, name: lustre-client, lov: lustre-clilov, lmv: lustre-clilmv } - { index: 2, event: attach, device: lustre-clilov, type: lov, UUID: lustre-clilov_UUID } - { index: 3, event: setup, device: lustre-clilov, UUID: } - { index: 6, event: attach, device: lustre-clilmv, type: lmv, UUID: lustre-clilmv_UUID } - { index: 7, event: setup, device: lustre-clilmv, UUID: } - { index: 10, event: new_profile, name: lustre-client, lov: lustre-clilov, lmv: lustre-clilmv } - { index: 13, event: add_uuid, nid: 192.168.202.133@tcp(0x20000c0a8ca85), node: 192.168.202.133@tcp } - { index: 14, event: attach, device: lustre-MDT0000-mdc, type: mdc, UUID: lustre-clilmv_UUID } - { index: 15, event: setup, device: lustre-MDT0000-mdc, UUID: lustre-MDT0000_UUID, node: 192.168.202.133@tcp } - { index: 16, event: add_mdc, device: lustre-clilmv, mdt: lustre-MDT0000_UUID, index: 0, gen: 1, UUID: lustre-MDT0000-mdc_UUID } - { index: 22, event: add_uuid, nid: 192.168.202.133@tcp(0x20000c0a8ca85), node: 192.168.202.133@tcp } - { index: 23, event: attach, device: lustre-MDT0001-mdc, type: mdc, UUID: lustre-clilmv_UUID } - { index: 24, event: setup, device: lustre-MDT0001-mdc, UUID: lustre-MDT0001_UUID, node: 192.168.202.133@tcp } - { index: 25, event: add_mdc, device: lustre-clilmv, mdt: lustre-MDT0001_UUID, index: 1, gen: 1, UUID: lustre-MDT0001-mdc_UUID } - { index: 31, event: add_uuid, nid: 192.168.202.133@tcp(0x20000c0a8ca85), node: 192.168.202.133@tcp } - { index: 32, event: attach, device: lustre-OST0000-osc, type: osc, UUID: lustre-clilov_UUID } - { index: 33, event: setup, device: lustre-OST0000-osc, UUID: lustre-OST0000_UUID, node: 192.168.202.133@tcp } - { index: 34, event: add_osc, device: lustre-clilov, ost: lustre-OST0000_UUID, index: 0, gen: 1 } - { index: 37, event: set_timeout, num: 0x000014, parameter: sys.timeout=20 } PASS 123aa (37s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 123ab: llog_print params output values from set_param -P ========================================================== 10:34:17 (1713537257) PASS 123ab (3s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 123ac: llog_print with --start and --end ========================================================== 10:34:23 (1713537263) PASS 123ac (3s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 123ad: llog_print shows all records == 10:34:28 (1713537268) PASS 123ad (3s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 123ae: llog_cancel can cancel requested record ========================================================== 10:34:33 (1713537273) - { index: 11, event: set_param, device: general, parameter: osc.*.max_dirty_mb, value: 467 } - { index: 46, event: conf_param, device: lustre-OST0000-osc, parameter: osc.max_dirty_mb=467 } - { index: 2, event: attach, device: lustre-clilov, type: lov, UUID: lustre-clilov_UUID } - { index: 3, event: setup, device: lustre-clilov, UUID: } - { index: 6, event: attach, device: lustre-clilmv, type: lmv, UUID: lustre-clilmv_UUID } - { index: 7, event: setup, device: lustre-clilmv, UUID: } - { index: 10, event: new_profile, name: lustre-client, lov: lustre-clilov, lmv: lustre-clilmv } - { index: 13, event: add_uuid, nid: 192.168.202.133@tcp(0x20000c0a8ca85), node: 192.168.202.133@tcp } - { index: 14, event: attach, device: lustre-MDT0000-mdc, type: mdc, UUID: lustre-clilmv_UUID } - { index: 15, event: setup, device: lustre-MDT0000-mdc, UUID: lustre-MDT0000_UUID, node: 192.168.202.133@tcp } - { index: 16, event: add_mdc, device: lustre-clilmv, mdt: lustre-MDT0000_UUID, index: 0, gen: 1, UUID: lustre-MDT0000-mdc_UUID } - { index: 22, event: add_uuid, nid: 192.168.202.133@tcp(0x20000c0a8ca85), node: 192.168.202.133@tcp } - { index: 23, event: attach, device: lustre-MDT0001-mdc, type: mdc, UUID: lustre-clilmv_UUID } - { index: 24, event: setup, device: lustre-MDT0001-mdc, UUID: lustre-MDT0001_UUID, node: 192.168.202.133@tcp } - { index: 25, event: add_mdc, device: lustre-clilmv, mdt: lustre-MDT0001_UUID, index: 1, gen: 1, UUID: lustre-MDT0001-mdc_UUID } - { index: 31, event: add_uuid, nid: 192.168.202.133@tcp(0x20000c0a8ca85), node: 192.168.202.133@tcp } - { index: 32, event: attach, device: lustre-OST0000-osc, type: osc, UUID: lustre-clilov_UUID } - { index: 33, event: setup, device: lustre-OST0000-osc, UUID: lustre-OST0000_UUID, node: 192.168.202.133@tcp } - { index: 34, event: add_osc, device: lustre-clilov, ost: lustre-OST0000_UUID, index: 0, gen: 1 } - { index: 37, event: set_timeout, num: 0x000014, parameter: sys.timeout=20 } - { index: 43, event: conf_param, device: lustre-OST0000-osc, parameter: osc.max_pages_per_rpc=1024 } - { index: 46, event: conf_param, device: lustre-OST0000-osc, parameter: osc.max_dirty_mb=467 } - { index: 2, event: attach, device: lustre-clilov, type: lov, UUID: lustre-clilov_UUID } - { index: 3, event: setup, device: lustre-clilov, UUID: } - { index: 6, event: attach, device: lustre-clilmv, type: lmv, UUID: lustre-clilmv_UUID } - { index: 7, event: setup, device: lustre-clilmv, UUID: } - { index: 10, event: new_profile, name: lustre-client, lov: lustre-clilov, lmv: lustre-clilmv } - { index: 13, event: add_uuid, nid: 192.168.202.133@tcp(0x20000c0a8ca85), node: 192.168.202.133@tcp } - { index: 14, event: attach, device: lustre-MDT0000-mdc, type: mdc, UUID: lustre-clilmv_UUID } - { index: 15, event: setup, device: lustre-MDT0000-mdc, UUID: lustre-MDT0000_UUID, node: 192.168.202.133@tcp } - { index: 16, event: add_mdc, device: lustre-clilmv, mdt: lustre-MDT0000_UUID, index: 0, gen: 1, UUID: lustre-MDT0000-mdc_UUID } - { index: 22, event: add_uuid, nid: 192.168.202.133@tcp(0x20000c0a8ca85), node: 192.168.202.133@tcp } - { index: 23, event: attach, device: lustre-MDT0001-mdc, type: mdc, UUID: lustre-clilmv_UUID } - { index: 24, event: setup, device: lustre-MDT0001-mdc, UUID: lustre-MDT0001_UUID, node: 192.168.202.133@tcp } - { index: 25, event: add_mdc, device: lustre-clilmv, mdt: lustre-MDT0001_UUID, index: 1, gen: 1, UUID: lustre-MDT0001-mdc_UUID } - { index: 31, event: add_uuid, nid: 192.168.202.133@tcp(0x20000c0a8ca85), node: 192.168.202.133@tcp } - { index: 32, event: attach, device: lustre-OST0000-osc, type: osc, UUID: lustre-clilov_UUID } - { index: 33, event: setup, device: lustre-OST0000-osc, UUID: lustre-OST0000_UUID, node: 192.168.202.133@tcp } - { index: 34, event: add_osc, device: lustre-clilov, ost: lustre-OST0000_UUID, index: 0, gen: 1 } - { index: 37, event: set_timeout, num: 0x000014, parameter: sys.timeout=20 } - { index: 43, event: conf_param, device: lustre-OST0000-osc, parameter: osc.max_pages_per_rpc=1024 } PASS 123ae (9s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 123af: llog_catlist can show all config files correctly ========================================================== 10:34:44 (1713537284) lctl --device MGS llog_catlist ... orig_clist: lustre-OST0000 lustre-MDT0001 lustre-client lustre-MDT0000 fail_loc=0x131b fail_val=2 new_clist: lustre-MDT0001 lustre-client lustre-MDT0000 fail_loc=0 done lctl --device lustre-MDT0000 llog_catlist ... orig_clist: [0x1:0x2:0x0] fail_loc=0x131b fail_val=2 new_clist: fail_loc=0 done fail_loc=0 PASS 123af (5s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 123ag: llog_print skips values deleted by set_param -P -d ========================================================== 10:34:51 (1713537291) PASS 123ag (5s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 123ah: del_ost cancels config log entries correctly ========================================================== 10:34:59 (1713537299) del_ost: dry run for target lustre-OST0000 config_log: lustre-MDT0001 [DRY RUN] cancel catalog 'lustre-MDT0001:38':"- { index: 38, event: conf_param, device: lustre-OST0000-osc-MDT0001, parameter: osc.max_dirty_mb=467 }" [DRY RUN] cancel catalog 'lustre-MDT0001:26':"- { index: 26, event: add_osc, device: lustre-MDT0001-mdtlov, ost: lustre-OST0000_UUID, index: 0, gen: 1 }" [DRY RUN] cancel catalog 'lustre-MDT0001:25':"- { index: 25, event: setup, device: lustre-OST0000-osc-MDT0001, UUID: lustre-OST0000_UUID, node: 192.168.202.133@tcp }" [DRY RUN] cancel catalog 'lustre-MDT0001:24':"- { index: 24, event: attach, device: lustre-OST0000-osc-MDT0001, type: osc, UUID: lustre-MDT0001-mdtlov_UUID }" del_ost: no catalog entry deleted config_log: lustre-client [DRY RUN] cancel catalog 'lustre-client:34':"- { index: 34, event: add_osc, device: lustre-clilov, ost: lustre-OST0000_UUID, index: 0, gen: 1 }" [DRY RUN] cancel catalog 'lustre-client:33':"- { index: 33, event: setup, device: lustre-OST0000-osc, UUID: lustre-OST0000_UUID, node: 192.168.202.133@tcp }" [DRY RUN] cancel catalog 'lustre-client:32':"- { index: 32, event: attach, device: lustre-OST0000-osc, type: osc, UUID: lustre-clilov_UUID }" del_ost: no catalog entry deleted config_log: lustre-MDT0000 [DRY RUN] cancel catalog 'lustre-MDT0000:41':"- { index: 41, event: conf_param, device: lustre-OST0000-osc-MDT0000, parameter: osc.max_dirty_mb=467 }" [DRY RUN] cancel catalog 'lustre-MDT0000:29':"- { index: 29, event: add_osc, device: lustre-MDT0000-mdtlov, ost: lustre-OST0000_UUID, index: 0, gen: 1 }" [DRY RUN] cancel catalog 'lustre-MDT0000:28':"- { index: 28, event: setup, device: lustre-OST0000-osc-MDT0000, UUID: lustre-OST0000_UUID, node: 192.168.202.133@tcp }" [DRY RUN] cancel catalog 'lustre-MDT0000:27':"- { index: 27, event: attach, device: lustre-OST0000-osc-MDT0000, type: osc, UUID: lustre-MDT0000-mdtlov_UUID }" del_ost: no catalog entry deleted config_log: lustre-MDT0001 cancel catalog lustre-MDT0001 log_idx 38: done cancel catalog lustre-MDT0001 log_idx 26: done cancel catalog lustre-MDT0001 log_idx 25: done cancel catalog lustre-MDT0001 log_idx 24: done del_ost: cancelled 4 catalog entries config_log: lustre-client cancel catalog lustre-client log_idx 34: done cancel catalog lustre-client log_idx 33: done cancel catalog lustre-client log_idx 32: done del_ost: cancelled 3 catalog entries config_log: lustre-MDT0000 cancel catalog lustre-MDT0000 log_idx 41: done cancel catalog lustre-MDT0000 log_idx 29: done cancel catalog lustre-MDT0000 log_idx 28: done cancel catalog lustre-MDT0000 log_idx 27: done del_ost: cancelled 4 catalog entries umount lustre on /mnt/lustre..... Stopping client oleg233-client.virtnet /mnt/lustre (opts:) mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre umount lustre on /mnt/lustre..... Stopping client oleg233-client.virtnet /mnt/lustre (opts:) stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server unloading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing unload_modules_local modules unloaded. Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:-f) Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:-f) pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 oleg233-server: oleg233-server.virtnet: executing set_hostid Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions ../libcfs/libcfs/libcfs options: 'cpu_npartitions=2' libkmod: kmod_module_get_holders: could not open '/sys/module/intel_rapl/holders': No such file or directory ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' quota/lquota options: 'hash_lqs_cur_bits=3' loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions oleg233-server: ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' oleg233-server: quota/lquota options: 'hash_lqs_cur_bits=3' Formatting mgs, mds, osts Format mds1: /dev/mapper/mds1_flakey Format mds2: /dev/mapper/mds2_flakey Format ost1: /dev/mapper/ost1_flakey Format ost2: /dev/mapper/ost2_flakey start mds service on oleg233-server Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/mds1_flakey Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/mds2_flakey Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Commit the device label on /dev/mapper/ost1_flakey Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid oleg233-server: oleg233-server.virtnet: executing wait_import_state FULL os[cp].lustre-OST0000-osc-MDT0000.ost_server_uuid 50 oleg233-server: os[cp].lustre-OST0000-osc-MDT0000.ost_server_uuid in FULL state after 0 sec oleg233-server: oleg233-server.virtnet: executing wait_import_state FULL os[cp].lustre-OST0000-osc-MDT0001.ost_server_uuid 50 oleg233-server: os[cp].lustre-OST0000-osc-MDT0001.ost_server_uuid in FULL state after 0 sec stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server fail_loc=0 PASS 123ah (85s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 123ai: llog_print display all non skipped records ========================================================== 10:36:27 (1713537387) start mds service on oleg233-server Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre oleg233-server: params: OBD_IOC_LLOG_PRINT failed: No such file or directory pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 - { index: 394, event: set_param, device: general, parameter: timeout, value: 129 } cleanup test 123ai timeout=20 timeout=20 PASS 123ai (63s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 123F: clear and reset all parameters using set_param -F ========================================================== 10:37:32 (1713537452) oleg233-server: rm: cannot remove '/tmp/f123F.conf-sanity.yaml': No such file or directory pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Unmounting FS Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:) Stopping client oleg233-client.virtnet /mnt/lustre opts: Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:) Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server Writeconf checking for existing Lustre data: found Read previous values: Target: lustre-MDT0000 Index: 0 Lustre FS: lustre Mount type: ldiskfs Flags: 0x5 (MDT MGS ) Persistent mount opts: user_xattr,errors=remount-ro Parameters: sys.timeout=20 mdt.identity_upcall=/home/green/git/lustre-release/lustre/utils/l_getidentity Permanent disk data: Target: lustre=MDT0000 Index: 0 Lustre FS: lustre Mount type: ldiskfs Flags: 0x105 (MDT MGS writeconf ) Persistent mount opts: user_xattr,errors=remount-ro Parameters: sys.timeout=20 mdt.identity_upcall=/home/green/git/lustre-release/lustre/utils/l_getidentity checking for existing Lustre data: found Read previous values: Target: lustre-MDT0001 Index: 1 Lustre FS: lustre Mount type: ldiskfs Flags: 0x1 (MDT ) Persistent mount opts: user_xattr,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 mdt.identity_upcall=/home/green/git/lustre-release/lustre/utils/l_getidentity Permanent disk data: Target: lustre=MDT0001 Index: 1 Lustre FS: lustre Mount type: ldiskfs Flags: 0x101 (MDT writeconf ) Persistent mount opts: user_xattr,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 mdt.identity_upcall=/home/green/git/lustre-release/lustre/utils/l_getidentity checking for existing Lustre data: found Read previous values: Target: lustre-OST0000 Index: 0 Lustre FS: lustre Mount type: ldiskfs Flags: 0x2 (OST ) Persistent mount opts: ,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 Permanent disk data: Target: lustre=OST0000 Index: 0 Lustre FS: lustre Mount type: ldiskfs Flags: 0x102 (OST writeconf ) Persistent mount opts: ,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 checking for existing Lustre data: found Read previous values: Target: lustre-OST0001 Index: 1 Lustre FS: lustre Mount type: ldiskfs Flags: 0x62 (OST first_time update ) Persistent mount opts: ,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 Permanent disk data: Target: lustre=OST0001 Index: 1 Lustre FS: lustre Mount type: ldiskfs Flags: 0x162 (OST first_time update writeconf ) Persistent mount opts: ,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 Remounting start mds service on oleg233-server Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre Setting configuration parameters This option left for backward compatibility, please use 'lctl apply_yaml' instead set_param: mdt.lustre-MDT0000.identity_upcall=/home/green/git/lustre-release/lustre/utils/l_getidentity set_param: mdt.lustre-MDT0001.identity_upcall=/home/green/git/lustre-release/lustre/utils/l_getidentity set_param: jobid_var=TESTNAME umount lustre on /mnt/lustre..... Stopping client oleg233-client.virtnet /mnt/lustre (opts:) stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server unloading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing unload_modules_local modules unloaded. pdsh@oleg233-client: oleg233-client: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-client: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 PASS 123F (86s) error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='0': No such file or directory oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='0': No such file or directory error: get_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory oleg233-server: error: get_param: param_path 'debug_raw_pointers': No such file or directory pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory == conf-sanity test 123G: clear and reset all parameters using apply_yaml ========================================================== 10:39:00 (1713537540) start mds service on oleg233-server Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' quota/lquota options: 'hash_lqs_cur_bits=3' loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions oleg233-server: libkmod: kmod_module_get_holders: could not open '/sys/module/acpi_cpufreq/holders': No such file or directory oleg233-server: ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' oleg233-server: quota/lquota options: 'hash_lqs_cur_bits=3' Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre oleg233-server: rm: cannot remove '/tmp/f123G.conf-sanity.yaml': No such file or directory pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Unmounting FS Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:) Stopping client oleg233-client.virtnet /mnt/lustre opts: Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:) Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server Writeconf checking for existing Lustre data: found Read previous values: Target: lustre-MDT0000 Index: 0 Lustre FS: lustre Mount type: ldiskfs Flags: 0x5 (MDT MGS ) Persistent mount opts: user_xattr,errors=remount-ro Parameters: sys.timeout=20 mdt.identity_upcall=/home/green/git/lustre-release/lustre/utils/l_getidentity Permanent disk data: Target: lustre=MDT0000 Index: 0 Lustre FS: lustre Mount type: ldiskfs Flags: 0x105 (MDT MGS writeconf ) Persistent mount opts: user_xattr,errors=remount-ro Parameters: sys.timeout=20 mdt.identity_upcall=/home/green/git/lustre-release/lustre/utils/l_getidentity checking for existing Lustre data: found Read previous values: Target: lustre-MDT0001 Index: 1 Lustre FS: lustre Mount type: ldiskfs Flags: 0x1 (MDT ) Persistent mount opts: user_xattr,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 mdt.identity_upcall=/home/green/git/lustre-release/lustre/utils/l_getidentity Permanent disk data: Target: lustre=MDT0001 Index: 1 Lustre FS: lustre Mount type: ldiskfs Flags: 0x101 (MDT writeconf ) Persistent mount opts: user_xattr,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 mdt.identity_upcall=/home/green/git/lustre-release/lustre/utils/l_getidentity checking for existing Lustre data: found Read previous values: Target: lustre-OST0000 Index: 0 Lustre FS: lustre Mount type: ldiskfs Flags: 0x2 (OST ) Persistent mount opts: ,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 Permanent disk data: Target: lustre=OST0000 Index: 0 Lustre FS: lustre Mount type: ldiskfs Flags: 0x102 (OST writeconf ) Persistent mount opts: ,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 checking for existing Lustre data: found Read previous values: Target: lustre-OST0001 Index: 1 Lustre FS: lustre Mount type: ldiskfs Flags: 0x142 (OST update writeconf ) Persistent mount opts: ,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 Permanent disk data: Target: lustre=OST0001 Index: 1 Lustre FS: lustre Mount type: ldiskfs Flags: 0x142 (OST update writeconf ) Persistent mount opts: ,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 Remounting start mds service on oleg233-server Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre Setting configuration parameters conf_param: lustre-MDT0000.mdt.identity_upcall=/home/green/git/lustre-release/lustre/utils/l_getidentity conf_param: lustre-MDT0001.mdt.identity_upcall=/home/green/git/lustre-release/lustre/utils/l_getidentity set_param: mdt.lustre-MDT0000.identity_upcall=/home/green/git/lustre-release/lustre/utils/l_getidentity set_param: mdt.lustre-MDT0001.identity_upcall=/home/green/git/lustre-release/lustre/utils/l_getidentity set_param: jobid_var=TESTNAME umount lustre on /mnt/lustre..... Stopping client oleg233-client.virtnet /mnt/lustre (opts:) stop ost1 service on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server unloading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing unload_modules_local modules unloaded. pdsh@oleg233-client: oleg233-client: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-client: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 PASS 123G (136s) error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='0': No such file or directory oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='0': No such file or directory error: get_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory oleg233-server: error: get_param: param_path 'debug_raw_pointers': No such file or directory pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory == conf-sanity test 124: check failover after replace_nids ========================================================== 10:41:19 (1713537679) SKIP: conf-sanity test_124 needs MDT failover setup SKIP 124 (1s) error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='0': No such file or directory oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='0': No such file or directory error: get_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory oleg233-server: error: get_param: param_path 'debug_raw_pointers': No such file or directory pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory == conf-sanity test 125: check l_tunedisk only tunes OSTs and their slave devices ========================================================== 10:41:22 (1713537682) Before: mgs /dev/mapper/mds1_flakey 511 2147483647 After: mgs /dev/mapper/mds1_flakey 511 2147483647 Before: ost1 /dev/mapper/ost1_flakey 16383 2147483647 oleg233-server: l_tunedisk: increased '/sys/devices/virtual/block/dm-2/queue/max_sectors_kb' from 16383 to 16384 After: ost1 /dev/mapper/ost1_flakey 16384 2147483647 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 PASS 125 (14s) error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='0': No such file or directory oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='0': No such file or directory error: get_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: param_path 'debug_raw_pointers': No such file or directory error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory oleg233-server: error: get_param: param_path 'debug_raw_pointers': No such file or directory pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 oleg233-server: error: set_param: param_path 'debug_raw_pointers': No such file or directory oleg233-server: error: set_param: setting 'debug_raw_pointers'='Y': No such file or directory == conf-sanity test 126: mount in parallel shouldn't cause a crash ========================================================== 10:41:38 (1713537698) umount lustre on /mnt/lustre..... stop ost1 service on oleg233-server stop mds service on oleg233-server stop mds service on oleg233-server LNET unconfigure error 22: (null) unloading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing unload_modules_local oleg233-server: LNET unconfigure error 22: (null) modules unloaded. oleg233-server: oleg233-server.virtnet: executing load_module ../libcfs/libcfs/libcfs fail_loc=0x60d oleg233-server: oleg233-server.virtnet: executing load_modules oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions oleg233-server: ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' oleg233-server: quota/lquota options: 'hash_lqs_cur_bits=3' Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions libkmod: kmod_module_get_holders: could not open '/sys/module/acpi_cpufreq/holders': No such file or directory ../libcfs/libcfs/libcfs options: 'cpu_npartitions=2' ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' quota/lquota options: 'hash_lqs_cur_bits=3' loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 clearing fail_loc on mds1 fail_loc=0 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 PASS 126 (35s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 127: direct io overwrite on full ost ========================================================== 10:42:15 (1713537735) umount lustre on /mnt/lustre..... stop ost1 service on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server stop mds service on oleg233-server unloading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing unload_modules_local modules unloaded. start mds service on oleg233-server Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions ../libcfs/libcfs/libcfs options: 'cpu_npartitions=2' ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' quota/lquota options: 'hash_lqs_cur_bits=3' loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions oleg233-server: ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' oleg233-server: quota/lquota options: 'hash_lqs_cur_bits=3' Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid start ost1 service on oleg233-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 oleg233-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre Stopping clients: /mnt/lustre (opts:) pdsh@oleg233-client: no remote hosts specified check osc.lustre-OST0000-osc-MDT0000.active target updated after 0 sec (got 1) check osc.lustre-OST0000-osc-MDT0001.active target updated after 0 sec (got 1) dd: error writing '/mnt/lustre/f127.conf-sanity': No space left on device 124+0 records in 123+0 records out 128974848 bytes (129 MB) copied, 4.84751 s, 26.6 MB/s 123+0 records in 123+0 records out 128974848 bytes (129 MB) copied, 4.65145 s, 27.7 MB/s PASS 127 (58s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 128: Force using remote logs with --nolocallogs ========================================================== 10:43:15 (1713537795) SKIP: conf-sanity test_128 need separate mgs device SKIP 128 (2s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 129: attempt to connect an OST with the same index should fail ========================================================== 10:43:19 (1713537799) Stopping clients: oleg233-client.virtnet /mnt/lustre (opts:) Stopping client oleg233-client.virtnet /mnt/lustre opts: Stopping clients: oleg233-client.virtnet /mnt/lustre2 (opts:) Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg233-server start mds service on oleg233-server Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg233-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0001-mdc-*.mds_server_uuid Format ost1: /dev/mapper/ost1_flakey Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 oleg233-server: mount.lustre: mount /dev/mapper/ost1_flakey at /mnt/lustre-ost1 failed: Address already in use oleg233-server: The target service's index is already in use. (/dev/mapper/ost1_flakey) pdsh@oleg233-client: oleg233-server: ssh exited with exit code 98 oleg233-server: error: set_param: param_path 'seq/cli-lustre:OST0000-super/width': No such file or directory oleg233-server: error: set_param: setting 'seq/cli-lustre:OST0000-super/width'='65536': No such file or directory pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 Start of /dev/mapper/ost1_flakey on ost1 failed 98 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 oleg233-server: mount.lustre: mount /dev/mapper/ost1_flakey at /mnt/lustre-ost1 failed: Address already in use oleg233-server: The target service's index is already in use. (/dev/mapper/ost1_flakey) pdsh@oleg233-client: oleg233-server: ssh exited with exit code 98 oleg233-server: error: set_param: param_path 'seq/cli-lustre:OST0000-super/width': No such file or directory oleg233-server: error: set_param: setting 'seq/cli-lustre:OST0000-super/width'='65536': No such file or directory pdsh@oleg233-client: oleg233-server: ssh exited with exit code 2 Start of /dev/mapper/ost1_flakey on ost1 failed 98 checking for existing Lustre data: found Read previous values: Target: lustre-OST0000 Index: 0 Lustre FS: lustre Mount type: ldiskfs Flags: 0x22 (OST first_time ) Persistent mount opts: ,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 Permanent disk data: Target: lustre=OST0000 Index: 0 Lustre FS: lustre Mount type: ldiskfs Flags: 0x122 (OST first_time writeconf ) Persistent mount opts: ,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 Writing CONFIGS/mountdata Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 Stopping /mnt/lustre-ost1 (opts:) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds1 (opts:-f) on oleg233-server stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server PASS 129 (58s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == conf-sanity test 130: re-register an MDT after writeconf ========================================================== 10:44:19 (1713537859) Checking servers environments Checking clients oleg233-client.virtnet environments Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions loading modules on: 'oleg233-server' oleg233-server: oleg233-server.virtnet: executing load_modules_local oleg233-server: Loading modules from /home/green/git/lustre-release/lustre oleg233-server: detected 4 online CPUs by sysfs oleg233-server: Force libcfs to create 2 CPU partitions Setup mgs, mdt, osts Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0000 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0000 Starting ost2: -o localrecov /dev/mapper/ost2_flakey /mnt/lustre-ost2 seq.cli-lustre-OST0001-super.width=65536 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-OST0001 mount lustre on /mnt/lustre..... Starting client: oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre Starting client oleg233-client.virtnet: -o user_xattr,flock oleg233-server@tcp:/lustre /mnt/lustre Started clients oleg233-client.virtnet: 192.168.202.133@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Using TIMEOUT=20 osc.lustre-OST0000-osc-ffff88012ff42800.idle_timeout=debug osc.lustre-OST0001-osc-ffff88012ff42800.idle_timeout=debug setting jobstats to procname_uid Setting lustre.sys.jobid_var from disable to procname_uid Waiting 90s for 'procname_uid' Updated after 2s: want 'procname_uid' got 'procname_uid' disable quota as required stop mds service on oleg233-server Stopping /mnt/lustre-mds2 (opts:-f) on oleg233-server checking for existing Lustre data: found Read previous values: Target: lustre-MDT0001 Index: 1 Lustre FS: lustre Mount type: ldiskfs Flags: 0x1 (MDT ) Persistent mount opts: user_xattr,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 mdt.identity_upcall=/home/green/git/lustre-release/lustre/utils/l_getidentity Permanent disk data: Target: lustre=MDT0001 Index: 1 Lustre FS: lustre Mount type: ldiskfs Flags: 0x101 (MDT writeconf ) Persistent mount opts: user_xattr,errors=remount-ro Parameters: mgsnode=192.168.202.133@tcp sys.timeout=20 mdt.identity_upcall=/home/green/git/lustre-release/lustre/utils/l_getidentity Writing CONFIGS/mountdata start mds service on oleg233-server Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg233-server: oleg233-server.virtnet: executing set_default_debug -1 all pdsh@oleg233-client: oleg233-server: ssh exited with exit code 1 Started lustre-MDT0001 PASS 130 (50s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y