-----============= acceptance-small: replay-single ============----- Fri Apr 19 08:49:39 EDT 2024 excepting tests: 110f 131b 59 36 === replay-single: start setup 08:49:41 (1713530981) === oleg119-client.virtnet: executing check_config_client /mnt/lustre oleg119-client.virtnet: Checking config lustre mounted on /mnt/lustre Checking servers environments Checking clients oleg119-client.virtnet environments Using TIMEOUT=20 osc.lustre-OST0000-osc-ffff8800b6d86800.idle_timeout=debug osc.lustre-OST0001-osc-ffff8800b6d86800.idle_timeout=debug disable quota as required oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all === replay-single: finish setup 08:49:48 (1713530988) === osp.lustre-OST0001-osc-MDT0000.prealloc_force_new_seq=1 osp.lustre-OST0000-osc-MDT0000.prealloc_force_new_seq=1 Creating to objid 33 on ost lustre-OST0000... Creating to objid 33 on ost lustre-OST0001... total: 33 open/close in 0.48 seconds: 69.23 ops/second total: 33 open/close in 0.49 seconds: 67.71 ops/second osp.lustre-OST0001-osc-MDT0000.prealloc_force_new_seq=0 osp.lustre-OST0000-osc-MDT0000.prealloc_force_new_seq=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 0a: empty replay =================== 08:50:01 (1713531001) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 08:50:04 (1713531004) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 08:50:16 (1713531016) targets are mounted 08:50:16 (1713531016) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 0a (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 0b: ensure object created after recover exists. (3284) ========================================================== 08:50:22 (1713531022) Failing ost1 on oleg119-server Stopping /mnt/lustre-ost1 (opts:) on oleg119-server 08:50:24 (1713531024) shut down Failover ost1 to oleg119-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-OST0000 08:50:37 (1713531037) targets are mounted 08:50:37 (1713531037) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec total: 20 open/close in 0.08 seconds: 241.82 ops/second - unlinked 0 (time 1713531041 ; total 0 ; last 0) total: 20 unlinks in 0 seconds: inf unlinks/second PASS 0b (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 0c: check replay-barrier =========== 08:50:43 (1713531043) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 7168 7531520 1% /mnt/lustre Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 08:50:46 (1713531046) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 08:50:59 (1713531059) targets are mounted 08:50:59 (1713531059) facet_failover done Starting client: oleg119-client.virtnet: -o user_xattr,flock oleg119-server@tcp:/lustre /mnt/lustre rm: cannot remove '/mnt/lustre/f0c.replay-single': No such file or directory PASS 0c (78s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 0d: expired recovery with no clients ========================================================== 08:52:03 (1713531123) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 7168 7531520 1% /mnt/lustre Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 08:52:05 (1713531125) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 08:52:17 (1713531137) targets are mounted 08:52:17 (1713531137) facet_failover done Starting client: oleg119-client.virtnet: -o user_xattr,flock oleg119-server@tcp:/lustre /mnt/lustre PASS 0d (77s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 1: simple create =================== 08:53:21 (1713531201) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 7168 7531520 1% /mnt/lustre Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 08:53:24 (1713531204) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 08:53:36 (1713531216) targets are mounted 08:53:36 (1713531216) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f1.replay-single has type file OK PASS 1 (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 2a: touch ========================== 08:53:42 (1713531222) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 7168 7531520 1% /mnt/lustre Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 08:53:45 (1713531225) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 08:53:57 (1713531237) targets are mounted 08:53:57 (1713531237) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f2a.replay-single has type file OK PASS 2a (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 2b: touch ========================== 08:54:03 (1713531243) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 08:54:06 (1713531246) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 08:54:18 (1713531258) targets are mounted 08:54:18 (1713531258) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f2b.replay-single has type file OK PASS 2b (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 2c: setstripe replay =============== 08:54:24 (1713531264) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 08:54:26 (1713531266) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 08:54:39 (1713531279) targets are mounted 08:54:39 (1713531279) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f2c.replay-single has type file OK PASS 2c (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 2d: setdirstripe replay ============ 08:54:45 (1713531285) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 08:54:48 (1713531288) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 08:55:00 (1713531300) targets are mounted 08:55:00 (1713531300) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d2d.replay-single has type dir OK PASS 2d (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 2e: O_CREAT|O_EXCL create replay === 08:55:06 (1713531306) fail_loc=0x8000013b UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 08:55:10 (1713531310) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 08:55:22 (1713531322) targets are mounted 08:55:22 (1713531322) facet_failover done Succeed in opening file "/mnt/lustre/f2e.replay-single"(flags=O_CREAT) oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f2e.replay-single has type file OK PASS 2e (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 3a: replay failed open(O_DIRECTORY) ========================================================== 08:55:28 (1713531328) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre Error in opening file "/mnt/lustre/f3a.replay-single"(flags=O_DIRECTORY) 20: Not a directory Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 08:55:31 (1713531331) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 08:55:43 (1713531343) targets are mounted 08:55:43 (1713531343) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f3a.replay-single has type file OK PASS 3a (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 3b: replay failed open -ENOMEM ===== 08:55:49 (1713531349) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre fail_loc=0x80000114 touch: cannot touch '/mnt/lustre/f3b.replay-single': Cannot allocate memory fail_loc=0 Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 08:55:52 (1713531352) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 08:56:04 (1713531364) targets are mounted 08:56:04 (1713531364) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/f3b.replay-single: No such file or directory PASS 3b (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 3c: replay failed open -ENOMEM ===== 08:56:10 (1713531370) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre fail_loc=0x80000128 touch: cannot touch '/mnt/lustre/f3c.replay-single': Cannot allocate memory fail_loc=0 Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 08:56:13 (1713531373) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 08:56:25 (1713531385) targets are mounted 08:56:25 (1713531385) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/f3c.replay-single: No such file or directory PASS 3c (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 4a: |x| 10 open(O_CREAT)s ========== 08:56:31 (1713531391) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 08:56:34 (1713531394) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 08:56:46 (1713531406) targets are mounted 08:56:46 (1713531406) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 4a (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 4b: |x| rm 10 files ================ 08:56:52 (1713531412) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 08:56:55 (1713531415) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 08:57:07 (1713531427) targets are mounted 08:57:07 (1713531427) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/f4b.replay-single-*: No such file or directory PASS 4b (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 5: |x| 220 open(O_CREAT) =========== 08:57:13 (1713531433) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 08:57:16 (1713531436) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 08:57:29 (1713531449) targets are mounted 08:57:29 (1713531449) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 5 (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 6a: mkdir + contained create ======= 08:57:40 (1713531460) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 08:57:43 (1713531463) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 08:57:55 (1713531475) targets are mounted 08:57:55 (1713531475) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d6a.replay-single has type dir OK /mnt/lustre/d6a.replay-single/f6a.replay-single has type file OK PASS 6a (22s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 6b: |X| rmdir ====================== 08:58:03 (1713531483) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 08:58:06 (1713531486) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 08:58:19 (1713531499) targets are mounted 08:58:19 (1713531499) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/d6b.replay-single: No such file or directory PASS 6b (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 7: mkdir |X| contained create ====== 08:58:25 (1713531505) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 08:58:27 (1713531507) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 08:58:39 (1713531519) targets are mounted 08:58:39 (1713531519) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d7.replay-single has type dir OK /mnt/lustre/d7.replay-single/f7.replay-single has type file OK PASS 7 (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 8: creat open |X| close ============ 08:58:46 (1713531526) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre multiop /mnt/lustre/f8.replay-single vmo_c TMPPIPE=/tmp/multiop_open_wait_pipe.6753 Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 08:58:48 (1713531528) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 08:59:01 (1713531541) targets are mounted 08:59:01 (1713531541) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f8.replay-single /mnt/lustre/f8.replay-single has type file OK PASS 8 (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 9: |X| create (same inum/gen) ====== 08:59:07 (1713531547) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 08:59:09 (1713531549) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 08:59:22 (1713531562) targets are mounted 08:59:22 (1713531562) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec old_inum == 144115305935798546, new_inum == 144115305935798546 old_inum and new_inum match PASS 9 (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 10: create |X| rename unlink ======= 08:59:29 (1713531569) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 08:59:31 (1713531571) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 08:59:43 (1713531583) targets are mounted 08:59:43 (1713531583) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/f10.replay-single: No such file or directory PASS 10 (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 11: create open write rename |X| create-old-name read ========================================================== 08:59:49 (1713531589) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre new old Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 08:59:52 (1713531592) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:00:05 (1713531605) targets are mounted 09:00:05 (1713531605) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec new old PASS 11 (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 12: open, unlink |X| close ========= 09:00:11 (1713531611) multiop /mnt/lustre/f12.replay-single vo_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6753 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:00:13 (1713531613) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:00:26 (1713531626) targets are mounted 09:00:26 (1713531626) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 12 (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 13: open chmod 0 |x| write close === 09:00:32 (1713531632) multiop /mnt/lustre/f13.replay-single vO_wc TMPPIPE=/tmp/multiop_open_wait_pipe.6753 /mnt/lustre/f13.replay-single has perms 00 OK UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:00:35 (1713531635) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:00:47 (1713531647) targets are mounted 09:00:47 (1713531647) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f13.replay-single has perms 00 OK /mnt/lustre/f13.replay-single has size 1 OK PASS 13 (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 14: open(O_CREAT), unlink |X| close ========================================================== 09:00:53 (1713531653) multiop /mnt/lustre/f14.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6753 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:00:56 (1713531656) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:01:08 (1713531668) targets are mounted 09:01:08 (1713531668) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 14 (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 15: open(O_CREAT), unlink |X| touch new, close ========================================================== 09:01:14 (1713531674) multiop /mnt/lustre/f15.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6753 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:01:17 (1713531677) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:01:30 (1713531690) targets are mounted 09:01:30 (1713531690) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 15 (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 16: |X| open(O_CREAT), unlink, touch new, unlink new ========================================================== 09:01:36 (1713531696) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:01:38 (1713531698) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:01:50 (1713531710) targets are mounted 09:01:50 (1713531710) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 16 (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 17: |X| open(O_CREAT), |replay| close ========================================================== 09:01:56 (1713531716) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre multiop /mnt/lustre/f17.replay-single vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.6753 Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:01:59 (1713531719) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:02:11 (1713531731) targets are mounted 09:02:11 (1713531731) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f17.replay-single has type file OK PASS 17 (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 18: open(O_CREAT), unlink, touch new, close, touch, unlink ========================================================== 09:02:17 (1713531737) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre multiop /mnt/lustre/f18.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6753 pid: 21600 will close Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:02:20 (1713531740) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:02:32 (1713531752) targets are mounted 09:02:32 (1713531752) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 18 (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 19: mcreate, open, write, rename === 09:02:38 (1713531758) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3328 2205184 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre old Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:02:41 (1713531761) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:02:53 (1713531773) targets are mounted 09:02:53 (1713531773) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec old PASS 19 (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 20a: |X| open(O_CREAT), unlink, replay, close (test mds_cleanup_orphans) ========================================================== 09:03:00 (1713531780) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre multiop /mnt/lustre/f20a.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6753 Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:03:03 (1713531783) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:03:15 (1713531795) targets are mounted 09:03:15 (1713531795) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 20a (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 20b: write, unlink, eviction, replay (test mds_cleanup_orphans) ========================================================== 09:03:21 (1713531801) /mnt/lustre/f20b.replay-single lmm_stripe_count: 1 lmm_stripe_size: 4194304 lmm_pattern: raid0 lmm_layout_gen: 0 lmm_stripe_offset: 0 obdidx objid objid group 0 1122 0x462 0x240000401 10000+0 records in 10000+0 records out 40960000 bytes (41 MB) copied, 1.27403 s, 32.1 MB/s Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:03:25 (1713531805) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:03:38 (1713531818) targets are mounted 09:03:38 (1713531818) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec affected facets: mds1 oleg119-server: oleg119-server.virtnet: executing _wait_recovery_complete *.lustre-MDT0000.recovery_status 1475 oleg119-server: *.lustre-MDT0000.recovery_status status: COMPLETE sleep 5 for ZFS zfs Waiting for MDT destroys to complete before 8192, after 8192 PASS 20b (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 20c: check that client eviction does not affect file content ========================================================== 09:03:52 (1713531832) multiop /mnt/lustre/f20c.replay-single vOw_c TMPPIPE=/tmp/multiop_open_wait_pipe.6753 -rw-r--r-- 1 root root 1 Apr 19 09:03 /mnt/lustre/f20c.replay-single PASS 20c (4s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 21: |X| open(O_CREAT), unlink touch new, replay, close (test mds_cleanup_orphans) ========================================================== 09:03:57 (1713531837) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3763200 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7528448 1% /mnt/lustre multiop /mnt/lustre/f21.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6753 Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:04:00 (1713531840) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:04:12 (1713531852) targets are mounted 09:04:12 (1713531852) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 21 (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 22: open(O_CREAT), |X| unlink, replay, close (test mds_cleanup_orphans) ========================================================== 09:04:19 (1713531859) multiop /mnt/lustre/f22.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6753 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3763200 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7528448 1% /mnt/lustre Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:04:22 (1713531862) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:04:35 (1713531875) targets are mounted 09:04:35 (1713531875) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 22 (22s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 23: open(O_CREAT), |X| unlink touch new, replay, close (test mds_cleanup_orphans) ========================================================== 09:04:43 (1713531883) multiop /mnt/lustre/f23.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6753 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3763200 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7528448 1% /mnt/lustre Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:04:46 (1713531886) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:04:58 (1713531898) targets are mounted 09:04:58 (1713531898) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 23 (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 24: open(O_CREAT), replay, unlink, close (test mds_cleanup_orphans) ========================================================== 09:05:05 (1713531905) multiop /mnt/lustre/f24.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6753 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3763200 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7528448 1% /mnt/lustre Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:05:08 (1713531908) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:05:22 (1713531922) targets are mounted 09:05:22 (1713531922) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 24 (21s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 25: open(O_CREAT), unlink, replay, close (test mds_cleanup_orphans) ========================================================== 09:05:28 (1713531928) multiop /mnt/lustre/f25.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6753 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3763200 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7528448 1% /mnt/lustre Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:05:31 (1713531931) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:05:44 (1713531944) targets are mounted 09:05:44 (1713531944) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 25 (22s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 26: |X| open(O_CREAT), unlink two, close one, replay, close one (test mds_cleanup_orphans) ========================================================== 09:05:52 (1713531952) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3763200 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7528448 1% /mnt/lustre multiop /mnt/lustre/f26.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6753 multiop /mnt/lustre/f26.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6753 Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:05:55 (1713531955) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:06:08 (1713531968) targets are mounted 09:06:08 (1713531968) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 26 (22s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 27: |X| open(O_CREAT), unlink two, replay, close two (test mds_cleanup_orphans) ========================================================== 09:06:16 (1713531976) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3763200 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7528448 1% /mnt/lustre multiop /mnt/lustre/f27.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6753 multiop /mnt/lustre/f27.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6753 Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:06:19 (1713531979) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:06:31 (1713531991) targets are mounted 09:06:31 (1713531991) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 27 (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 28: open(O_CREAT), |X| unlink two, close one, replay, close one (test mds_cleanup_orphans) ========================================================== 09:06:37 (1713531997) multiop /mnt/lustre/f28.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6753 multiop /mnt/lustre/f28.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6753 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3763200 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7528448 1% /mnt/lustre Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:06:40 (1713532000) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:06:53 (1713532013) targets are mounted 09:06:53 (1713532013) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 28 (22s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 29: open(O_CREAT), |X| unlink two, replay, close two (test mds_cleanup_orphans) ========================================================== 09:07:01 (1713532021) multiop /mnt/lustre/f29.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6753 multiop /mnt/lustre/f29.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6753 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3763200 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7528448 1% /mnt/lustre Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:07:04 (1713532024) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:07:18 (1713532038) targets are mounted 09:07:18 (1713532038) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 29 (22s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 30: open(O_CREAT) two, unlink two, replay, close two (test mds_cleanup_orphans) ========================================================== 09:07:25 (1713532045) multiop /mnt/lustre/f30.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6753 multiop /mnt/lustre/f30.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6753 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3763200 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7528448 1% /mnt/lustre Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:07:27 (1713532047) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:07:40 (1713532060) targets are mounted 09:07:40 (1713532060) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 30 (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 31: open(O_CREAT) two, unlink one, |X| unlink one, close two (test mds_cleanup_orphans) ========================================================== 09:07:47 (1713532067) multiop /mnt/lustre/f31.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6753 multiop /mnt/lustre/f31.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6753 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3763200 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7528448 1% /mnt/lustre Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:07:51 (1713532071) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:08:04 (1713532084) targets are mounted 09:08:04 (1713532084) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 31 (23s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 32: close() notices client eviction; close() after client eviction ========================================================== 09:08:12 (1713532092) multiop /mnt/lustre/f32.replay-single vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.6753 multiop /mnt/lustre/f32.replay-single vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.6753 pdsh@oleg119-client: oleg119-client: ssh exited with exit code 5 PASS 32 (5s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 33a: fid seq shouldn't be reused after abort recovery ========================================================== 09:08:19 (1713532099) total: 10 open/close in 0.07 seconds: 136.55 ops/second Replay barrier on lustre-MDT0000 Stopping /mnt/lustre-mds1 (opts:) on oleg119-server Failover mds1 to oleg119-server Starting mds1: -o localrecov -o abort_recovery lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg119-client: error: invalid path '/mnt/lustre': Input/output error pdsh@oleg119-client: oleg119-client: ssh exited with exit code 5 first stat failed: 5 total: 10 open/close in 0.08 seconds: 131.98 ops/second PASS 33a (12s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 33b: test fid seq allocation ======= 09:08:33 (1713532113) fail_loc=0x1311 total: 10 open/close in 0.07 seconds: 137.23 ops/second Replay barrier on lustre-MDT0000 Stopping /mnt/lustre-mds1 (opts:) on oleg119-server Failover mds1 to oleg119-server Starting mds1: -o localrecov -o abort_recovery lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 total: 10 open/close in 0.06 seconds: 156.89 ops/second PASS 33b (12s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 34: abort recovery before client does replay (test mds_cleanup_orphans) ========================================================== 09:08:47 (1713532127) multiop /mnt/lustre/f34.replay-single vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.6753 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3328 2205184 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3763200 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7528448 1% /mnt/lustre Stopping /mnt/lustre-mds1 (opts:) on oleg119-server Failover mds1 to oleg119-server Starting mds1: -o localrecov -o abort_recovery lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 PASS 34 (11s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 35: test recovery from llog for unlink op ========================================================== 09:09:01 (1713532141) fail_loc=0x80000119 Stopping /mnt/lustre-mds1 (opts:) on oleg119-server Failover mds1 to oleg119-server Starting mds1: -o localrecov -o abort_recovery lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 rm: cannot remove '/mnt/lustre/f35.replay-single': Input/output error Started lustre-MDT0000 Can't lstat /mnt/lustre/f35.replay-single: No such file or directory PASS 35 (12s) debug_raw_pointers=0 debug_raw_pointers=0 SKIP: replay-single test_36 skipping ALWAYS excluded test 36 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 37: abort recovery before client does replay (test mds_cleanup_orphans for directories) ========================================================== 09:09:16 (1713532156) multiop /mnt/lustre/d37.replay-single/f37.replay-single vdD_c TMPPIPE=/tmp/multiop_open_wait_pipe.6753 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3328 2205184 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3763200 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7528448 1% /mnt/lustre Stopping /mnt/lustre-mds1 (opts:) on oleg119-server Failover mds1 to oleg119-server Starting mds1: -o localrecov -o abort_recovery lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 PASS 37 (13s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 38: test recovery from unlink llog (test llog_gen_rec) ========================================================== 09:09:31 (1713532171) total: 800 open/close in 4.55 seconds: 175.89 ops/second - unlinked 0 (time 1713532178 ; total 0 ; last 0) total: 400 unlinks in 2 seconds: 200.000000 unlinks/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3712 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3763200 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7528448 1% /mnt/lustre Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:09:43 (1713532183) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:09:57 (1713532197) targets are mounted 09:09:57 (1713532197) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec - unlinked 0 (time 1713532202 ; total 0 ; last 0) total: 400 unlinks in 2 seconds: 200.000000 unlinks/second Can't lstat /mnt/lustre/f38.replay-single-*: No such file or directory PASS 38 (37s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 39: test recovery from unlink llog (test llog_gen_rec) ========================================================== 09:10:10 (1713532210) total: 800 open/close in 4.31 seconds: 185.74 ops/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3584 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3763200 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7528448 1% /mnt/lustre - unlinked 0 (time 1713532218 ; total 0 ; last 0) total: 400 unlinks in 2 seconds: 200.000000 unlinks/second Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:10:22 (1713532222) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:10:35 (1713532235) targets are mounted 09:10:35 (1713532235) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec - unlinked 0 (time 1713532240 ; total 0 ; last 0) total: 400 unlinks in 3 seconds: 133.333328 unlinks/second Can't lstat /mnt/lustre/f39.replay-single-*: No such file or directory PASS 39 (37s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 41: read from a valid osc while other oscs are invalid ========================================================== 09:10:49 (1713532249) 1+0 records in 1+0 records out 4096 bytes (4.1 kB) copied, 0.00254726 s, 1.6 MB/s 1+0 records in 1+0 records out 4096 bytes (4.1 kB) copied, 0.00695773 s, 589 kB/s PASS 41 (4s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 42: recovery after ost failure ===== 09:10:55 (1713532255) total: 800 open/close in 4.52 seconds: 177.14 ops/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3840 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre - unlinked 0 (time 1713532263 ; total 0 ; last 0) total: 400 unlinks in 1 seconds: 400.000000 unlinks/second debug=-1 Failing ost1 on oleg119-server Stopping /mnt/lustre-ost1 (opts:) on oleg119-server 09:11:07 (1713532267) shut down Failover ost1 to oleg119-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-OST0000 09:11:21 (1713532281) targets are mounted 09:11:21 (1713532281) facet_failover done wait for MDS to timeout and recover debug=super ioctl neterror warning dlmtrace error emerg ha rpctrace vfstrace config console lfsck debug=super ioctl neterror warning dlmtrace error emerg ha rpctrace vfstrace config console lfsck - unlinked 0 (time 1713532322 ; total 0 ; last 0) total: 400 unlinks in 2 seconds: 200.000000 unlinks/second Can't lstat /mnt/lustre/f42.replay-single-*: No such file or directory PASS 42 (71s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 43: mds osc import failure during recovery; don't LBUG ========================================================== 09:12:08 (1713532328) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3584 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre fail_loc=0x80000204 Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:12:12 (1713532332) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:12:25 (1713532345) targets are mounted 09:12:25 (1713532345) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 43 (33s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 44a: race in target handle connect ========================================================== 09:12:43 (1713532363) at_max=40 1 of 10 (1713532365) service : cur 5 worst 5 (at 1713532345, 21s ago) 4 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3584 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre 2 of 10 (1713532371) service : cur 6 worst 6 (at 1713532372, 0s ago) 6 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3584 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre 3 of 10 (1713532376) service : cur 6 worst 6 (at 1713532372, 5s ago) 6 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3584 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre 4 of 10 (1713532382) service : cur 6 worst 6 (at 1713532372, 11s ago) 6 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3584 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre 5 of 10 (1713532387) service : cur 6 worst 6 (at 1713532372, 16s ago) 6 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3584 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre 6 of 10 (1713532393) service : cur 6 worst 6 (at 1713532372, 22s ago) 6 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3584 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre 7 of 10 (1713532398) service : cur 6 worst 6 (at 1713532372, 28s ago) 6 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3584 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre 8 of 10 (1713532404) service : cur 6 worst 6 (at 1713532372, 33s ago) 6 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3584 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre 9 of 10 (1713532409) service : cur 6 worst 6 (at 1713532372, 39s ago) 6 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3584 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre 10 of 10 (1713532415) service : cur 6 worst 6 (at 1713532372, 44s ago) 6 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3584 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre fail_loc=0 at_max=600 PASS 44a (60s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 44b: race in target handle connect ========================================================== 09:13:44 (1713532424) 1 of 10 (1713532425) service : cur 6 worst 6 (at 1713532372, 54s ago) 6 0 0 0 fail_loc=0x80000704 error: recover: Connection timed out Filesystem 1K-blocks Used Available Use% Mounted on 192.168.201.119@tcp:/lustre 7542784 8192 7530496 1% /mnt/lustre 2 of 10 (1713532465) service : cur 40 worst 40 (at 1713532466, 0s ago) 40 0 0 0 fail_loc=0x80000704 error: recover: Connection timed out df: '/mnt/lustre': Operation already in progress 3 of 10 (1713532486) service : cur 40 worst 40 (at 1713532466, 21s ago) 40 0 0 0 fail_loc=0x80000704 error: recover: Connection timed out df: '/mnt/lustre': Operation already in progress 4 of 10 (1713532506) service : cur 40 worst 40 (at 1713532466, 42s ago) 40 40 0 0 fail_loc=0x80000704 error: recover: Connection timed out df: '/mnt/lustre': Operation already in progress 5 of 10 (1713532527) service : cur 40 worst 40 (at 1713532466, 62s ago) 40 40 0 0 fail_loc=0x80000704 error: recover: Connection timed out df: '/mnt/lustre': Operation already in progress 6 of 10 (1713532548) service : cur 40 worst 40 (at 1713532466, 83s ago) 40 40 0 0 fail_loc=0x80000704 error: recover: Connection timed out df: '/mnt/lustre': Operation already in progress 7 of 10 (1713532568) service : cur 40 worst 40 (at 1713532466, 104s ago) 40 40 0 0 fail_loc=0x80000704 error: recover: Connection timed out df: '/mnt/lustre': Operation already in progress 8 of 10 (1713532589) service : cur 40 worst 40 (at 1713532466, 124s ago) 40 40 0 0 fail_loc=0x80000704 error: recover: Connection timed out df: '/mnt/lustre': Operation already in progress 9 of 10 (1713532610) service : cur 40 worst 40 (at 1713532466, 145s ago) 40 40 0 0 fail_loc=0x80000704 error: recover: Connection timed out df: '/mnt/lustre': Operation already in progress 10 of 10 (1713532630) service : cur 40 worst 40 (at 1713532466, 166s ago) 40 40 0 0 fail_loc=0x80000704 error: recover: Connection timed out df: '/mnt/lustre': Operation already in progress PASS 44b (229s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 44c: race in target handle connect ========================================================== 09:17:35 (1713532655) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3584 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre total: 100 create in 0.35 seconds: 285.00 ops/second fail_loc=0x80000712 Stopping /mnt/lustre-mds1 (opts:) on oleg119-server Failover mds1 to oleg119-server Starting mds1: -o localrecov -o abort_recovery lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 unlink(/mnt/lustre/f44c.replay-single-0) error: No such file or directory total: 0 unlinks in 0 seconds: -nan unlinks/second Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:17:47 (1713532667) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:17:59 (1713532679) targets are mounted 09:17:59 (1713532679) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec unlink(/mnt/lustre/f44c.replay-single-0) error: No such file or directory total: 0 unlinks in 0 seconds: -nan unlinks/second PASS 44c (29s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 45: Handle failed close ============ 09:18:06 (1713532686) multiop /mnt/lustre/f45.replay-single vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.6753 /mnt/lustre/f45.replay-single has type file OK PASS 45 (3s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 46: Don't leak file handle after open resend (3325) ========================================================== 09:18:10 (1713532690) fail_loc=0x122 fail_loc=0 Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:18:29 (1713532709) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:18:41 (1713532721) targets are mounted 09:18:41 (1713532721) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec lfs path2fid: cannot get fid for 'f46.replay-single': No such file or directory PASS 46 (37s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 47: MDS->OSC failure during precreate cleanup (2824) ========================================================== 09:18:49 (1713532729) total: 20 open/close in 0.15 seconds: 131.50 ops/second Failing ost1 on oleg119-server Stopping /mnt/lustre-ost1 (opts:) on oleg119-server 09:18:52 (1713532732) shut down Failover ost1 to oleg119-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-OST0000 09:19:05 (1713532745) targets are mounted 09:19:05 (1713532745) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec fail_loc=0x80000204 total: 20 open/close in 0.13 seconds: 159.59 ops/second - unlinked 0 (time 1713532810 ; total 0 ; last 0) total: 20 unlinks in 0 seconds: inf unlinks/second PASS 47 (83s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 48: MDS->OSC failure during precreate cleanup (2824) ========================================================== 09:20:14 (1713532814) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3584 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre total: 20 open/close in 0.12 seconds: 167.53 ops/second Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:20:18 (1713532818) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:20:30 (1713532830) targets are mounted 09:20:30 (1713532830) facet_failover done fail_loc=0x80000216 total: 20 open/close in 0.07 seconds: 291.35 ops/second - unlinked 0 (time 1713532892 ; total 0 ; last 0) total: 40 unlinks in 0 seconds: inf unlinks/second PASS 48 (79s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 50: Double OSC recovery, don't LASSERT (3812) ========================================================== 09:21:35 (1713532895) PASS 50 (8s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 52: time out lock replay (3764) ==== 09:21:45 (1713532905) multiop /mnt/lustre/f52.replay-single vs_s TMPPIPE=/tmp/multiop_open_wait_pipe.6753 fail_loc=0x80000157 Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:21:47 (1713532907) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:22:01 (1713532921) targets are mounted 09:22:01 (1713532921) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec fail_loc=0x0 fail_loc=0x0 PASS 52 (37s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53a: |X| close request while two MDC requests in flight ========================================================== 09:22:24 (1713532944) fail_loc=0x80000115 fail_loc=0 Replay barrier on lustre-MDT0000 Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:22:29 (1713532949) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:22:42 (1713532962) targets are mounted 09:22:42 (1713532962) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d53a.replay-single-1/f has type file OK /mnt/lustre/d53a.replay-single-2/f has type file OK PASS 53a (24s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53b: |X| open request while two MDC requests in flight ========================================================== 09:22:50 (1713532970) multiop /mnt/lustre/d53b.replay-single-1/f vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.6753 fail_loc=0x80000107 fail_loc=0 Replay barrier on lustre-MDT0000 Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:22:55 (1713532975) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:23:07 (1713532987) targets are mounted 09:23:07 (1713532987) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d53b.replay-single-1/f has type file OK /mnt/lustre/d53b.replay-single-2/f has type file OK PASS 53b (22s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53c: |X| open request and close request while two MDC requests in flight ========================================================== 09:23:14 (1713532994) fail_loc=0x80000107 fail_loc=0x80000115 Replay barrier on lustre-MDT0000 Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:23:19 (1713532999) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:23:33 (1713533013) targets are mounted 09:23:33 (1713533013) facet_failover done fail_loc=0 /mnt/lustre/d53c.replay-single-1/f has type file OK /mnt/lustre/d53c.replay-single-2/f has type file OK PASS 53c (23s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53d: close reply while two MDC requests in flight ========================================================== 09:23:39 (1713533019) fail_loc=0x8000013b fail_loc=0 Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:23:42 (1713533022) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:23:55 (1713533035) targets are mounted 09:23:55 (1713533035) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d53d.replay-single-1/f has type file OK /mnt/lustre/d53d.replay-single-2/f has type file OK PASS 53d (21s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53e: |X| open reply while two MDC requests in flight ========================================================== 09:24:01 (1713533041) fail_loc=0x119 fail_loc=0 Replay barrier on lustre-MDT0000 Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:24:06 (1713533046) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:24:19 (1713533059) targets are mounted 09:24:19 (1713533059) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d53e.replay-single-1/f has type file OK /mnt/lustre/d53e.replay-single-2/f has type file OK PASS 53e (23s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53f: |X| open reply and close reply while two MDC requests in flight ========================================================== 09:24:26 (1713533066) fail_loc=0x119 fail_loc=0x8000013b Replay barrier on lustre-MDT0000 Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:24:31 (1713533071) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:24:43 (1713533083) targets are mounted 09:24:43 (1713533083) facet_failover done fail_loc=0 /mnt/lustre/d53f.replay-single-1/f has type file OK /mnt/lustre/d53f.replay-single-2/f has type file OK PASS 53f (21s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53g: |X| drop open reply and close request while close and open are both in flight ========================================================== 09:24:49 (1713533089) fail_loc=0x119 fail_loc=0x80000115 fail_loc=0 Replay barrier on lustre-MDT0000 Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:24:53 (1713533093) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:25:06 (1713533106) targets are mounted 09:25:06 (1713533106) facet_failover done /mnt/lustre/d53g.replay-single-1/f has type file OK /mnt/lustre/d53g.replay-single-2/f has type file OK PASS 53g (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53h: open request and close reply while two MDC requests in flight ========================================================== 09:25:11 (1713533111) fail_loc=0x80000107 fail_loc=0x8000013b Replay barrier on lustre-MDT0000 Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:25:16 (1713533116) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:25:29 (1713533129) targets are mounted 09:25:29 (1713533129) facet_failover done fail_loc=0 /mnt/lustre/d53h.replay-single-1/f has type file OK /mnt/lustre/d53h.replay-single-2/f has type file OK PASS 53h (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 55: let MDS_CHECK_RESENT return the original return code instead of 0 ========================================================== 09:25:41 (1713533141) fail_loc=0x8000012b fail_loc=0x0 rm: touch: cannot remove '/mnt/lustre/f55.replay-single'cannot touch '/mnt/lustre/f55.replay-single': No such file or directory : No such file or directory PASS 55 (18s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 56: don't replay a symlink open request (3440) ========================================================== 09:26:01 (1713533161) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3584 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:26:04 (1713533164) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:26:17 (1713533177) targets are mounted 09:26:17 (1713533177) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 56 (31s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 57: test recovery from llog for setattr op ========================================================== 09:26:34 (1713533194) fail_loc=0x8000012c UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3584 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:26:37 (1713533197) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:26:51 (1713533211) targets are mounted 09:26:51 (1713533211) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec affected facets: mds1 oleg119-server: oleg119-server.virtnet: executing _wait_recovery_complete *.lustre-MDT0000.recovery_status 1475 oleg119-server: *.lustre-MDT0000.recovery_status status: COMPLETE Waiting for orphan cleanup... osp.lustre-OST0000-osc-MDT0000.old_sync_processed osp.lustre-OST0001-osc-MDT0000.old_sync_processed wait 40 secs maximumly for oleg119-server mds-ost sync done. /mnt/lustre/f57.replay-single has type file OK fail_loc=0x0 PASS 57 (25s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 58a: test recovery from llog for setattr op (test llog_gen_rec) ========================================================== 09:27:01 (1713533221) fail_loc=0x8000012c total: 2500 open/close in 7.53 seconds: 331.93 ops/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 4224 2204288 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:27:13 (1713533233) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:27:26 (1713533246) targets are mounted 09:27:26 (1713533246) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec fail_loc=0x0 - unlinked 0 (time 1713533262 ; total 0 ; last 0) total: 2500 unlinks in 5 seconds: 500.000000 unlinks/second PASS 58a (48s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 58b: test replay of setxattr op ==== 09:27:51 (1713533271) Starting client: oleg119-client.virtnet: -o user_xattr,flock oleg119-server@tcp:/lustre /mnt/lustre2 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 4096 2204416 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:27:54 (1713533274) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:28:07 (1713533287) targets are mounted 09:28:07 (1713533287) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Stopping client oleg119-client.virtnet /mnt/lustre2 (opts:) oleg119-client.virtnet: executing wait_import_state_mount FULL mgc.*.mgs_server_uuid mgc.*.mgs_server_uuid in FULL state after 0 sec PASS 58b (22s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 58c: resend/reconstruct setxattr op ========================================================== 09:28:15 (1713533295) Starting client: oleg119-client.virtnet: -o user_xattr,flock oleg119-server@tcp:/lustre /mnt/lustre2 fail_val=0 fail_loc=0x123 fail_loc=0 fail_loc=0x119 fail_loc=0 Stopping client oleg119-client.virtnet /mnt/lustre2 (opts:) PASS 58c (41s) debug_raw_pointers=0 debug_raw_pointers=0 SKIP: replay-single test_59 skipping ALWAYS excluded test 59 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 60: test llog post recovery init vs llog unlink ========================================================== 09:28:59 (1713533339) total: 200 open/close in 0.83 seconds: 242.19 ops/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3712 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre - unlinked 0 (time 1713533343 ; total 0 ; last 0) total: 100 unlinks in 0 seconds: inf unlinks/second Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:29:05 (1713533345) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:29:18 (1713533358) targets are mounted 09:29:18 (1713533358) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec - unlinked 0 (time 1713533362 ; total 0 ; last 0) total: 100 unlinks in 0 seconds: inf unlinks/second PASS 60 (25s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 61a: test race llog recovery vs llog cleanup ========================================================== 09:29:26 (1713533366) total: 800 open/close in 2.36 seconds: 339.41 ops/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3968 2204544 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre - unlinked 0 (time 1713533372 ; total 0 ; last 0) total: 800 unlinks in 1 seconds: 800.000000 unlinks/second fail_val=0 fail_loc=0x80000221 Failing ost1 on oleg119-server Stopping /mnt/lustre-ost1 (opts:) on oleg119-server 09:29:35 (1713533375) shut down Failover ost1 to oleg119-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-OST0000 09:29:48 (1713533388) targets are mounted 09:29:48 (1713533388) facet_failover done Failing ost1 on oleg119-server Stopping /mnt/lustre-ost1 (opts:) on oleg119-server 09:29:59 (1713533399) shut down Failover ost1 to oleg119-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-OST0000 09:30:12 (1713533412) targets are mounted 09:30:12 (1713533412) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec fail_val=0 fail_loc=0x0 Can't lstat /mnt/lustre/d61a.replay-single/f61a.replay-single-*: No such file or directory PASS 61a (82s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 61b: test race mds llog sync vs llog cleanup ========================================================== 09:30:50 (1713533450) fail_loc=0x8000013a Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:30:52 (1713533452) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:31:05 (1713533465) targets are mounted 09:31:05 (1713533465) facet_failover done Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:31:16 (1713533476) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:31:29 (1713533489) targets are mounted 09:31:29 (1713533489) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec 1+0 records in 1+0 records out 4096 bytes (4.1 kB) copied, 0.00336136 s, 1.2 MB/s PASS 61b (44s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 61c: test race mds llog sync vs llog cleanup ========================================================== 09:31:36 (1713533496) fail_val=0 fail_loc=0x80000222 Failing ost1 on oleg119-server Stopping /mnt/lustre-ost1 (opts:) on oleg119-server 09:31:47 (1713533507) shut down Failover ost1 to oleg119-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-OST0000 09:32:01 (1713533521) targets are mounted 09:32:01 (1713533521) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec fail_val=0 fail_loc=0x0 PASS 61c (30s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 61d: error in llog_setup should cleanup the llog context correctly ========================================================== 09:32:08 (1713533528) Stopping /mnt/lustre-mds1 (opts:) on oleg119-server fail_loc=0x80000605 Starting mgs: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: mount.lustre: mount lustre-mdt1/mdt1 at /mnt/lustre-mds1 failed: Operation not supported pdsh@oleg119-client: oleg119-server: ssh exited with exit code 95 Start of lustre-mdt1/mdt1 on mgs failed 95 fail_loc=0 Starting mgs: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 PASS 61d (8s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 62: don't mis-drop resent replay === 09:32:18 (1713533538) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3712 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre total: 25 open/close in 0.10 seconds: 246.66 ops/second fail_loc=0x80000707 Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:32:22 (1713533542) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:32:34 (1713533554) targets are mounted 09:32:34 (1713533554) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec fail_loc=0 - unlinked 0 (time 1713533573 ; total 0 ; last 0) total: 25 unlinks in 0 seconds: inf unlinks/second PASS 62 (37s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 65a: AT: verify early replies ====== 09:32:56 (1713533576) at_history=8 at_history=8 debug=other fail_val=11000 fail_loc=0x8000050a 00000100:00001000:3.0:1713533606.258349:0:6955:0:(client.c:537:ptlrpc_at_recv_early_reply()) @@@ Early reply #1, new deadline in 35s (25s) req@ffff8800a8c3a300 x1796767408128448/t0(0) o101->lustre-MDT0000-mdc-ffff8800a9887800@192.168.201.119@tcp:12/10 lens 664/66320 e 1 to 0 dl 1713533641 ref 2 fl Rpc:PQr/200/ffffffff rc 0/-1 job:'createmany.0' uid:0 gid:0 portal 12 : cur 36 worst 40 (at 1713532465, 1146s ago) 36 5 5 5 portal 23 : cur 5 worst 10 (at 1713531238, 2373s ago) 5 10 5 5 portal 30 : cur 5 worst 5 (at 1713531202, 2409s ago) 5 5 0 0 portal 17 : cur 5 worst 5 (at 1713531265, 2346s ago) 5 5 5 5 portal 13 : cur 10 worst 10 (at 1713531583, 2028s ago) 10 10 0 0 portal 12 : cur 5 worst 40 (at 1713532465, 1155s ago) 36 5 5 5 portal 23 : cur 5 worst 10 (at 1713531238, 2382s ago) 5 10 5 5 portal 30 : cur 5 worst 5 (at 1713531202, 2418s ago) 5 5 0 0 portal 17 : cur 5 worst 5 (at 1713531265, 2355s ago) 5 5 5 5 portal 13 : cur 10 worst 10 (at 1713531583, 2037s ago) 10 10 0 0 PASS 65a (46s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 65b: AT: verify early replies on packed reply / bulk ========================================================== 09:33:43 (1713533623) at_history=8 at_history=8 debug=other trace fail_val=11 fail_loc=0x224 fail_loc=0 00000100:00001000:3.0:1713533653.650114:0:1827:0:(client.c:537:ptlrpc_at_recv_early_reply()) @@@ Early reply #1, new deadline in 35s (25s) req@ffff8800a8c38000 x1796767408138304/t0(0) o4->lustre-OST0000-osc-ffff8800a9887800@192.168.201.119@tcp:6/4 lens 4584/448 e 1 to 0 dl 1713533688 ref 2 fl Rpc:Qr/200/ffffffff rc 0/-1 job:'multiop.0' uid:0 gid:0 debug=super ioctl neterror warning dlmtrace error emerg ha rpctrace vfstrace config console lfsck debug=super ioctl neterror warning dlmtrace error emerg ha rpctrace vfstrace config console lfsck portal 28 : cur 5 worst 5 (at 1713531197, 2462s ago) 5 5 5 5 portal 7 : cur 5 worst 5 (at 1713531199, 2460s ago) 5 5 5 5 portal 17 : cur 5 worst 5 (at 1713531263, 2396s ago) 5 0 5 5 portal 6 : cur 36 worst 36 (at 1713533658, 1s ago) 36 5 0 0 PASS 65b (37s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 66a: AT: verify MDT service time adjusts with no early replies ========================================================== 09:34:22 (1713533662) at_history=8 at_history=8 portal 12 : cur 5 worst 40 (at 1713532465, 1220s ago) 36 5 5 5 fail_val=5000 fail_loc=0x8000050a portal 12 : cur 5 worst 40 (at 1713532465, 1225s ago) 36 5 5 5 fail_val=10000 fail_loc=0x8000050a portal 12 : cur 36 worst 40 (at 1713532465, 1236s ago) 36 5 5 5 fail_loc=0 portal 12 : cur 5 worst 40 (at 1713532465, 1246s ago) 36 5 5 5 Current MDT timeout 5, worst 40 PASS 66a (50s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 66b: AT: verify net latency adjusts ========================================================== 09:35:14 (1713533714) at_history=8 at_history=8 fail_val=10 fail_loc=0x50c fail_loc=0 network timeout orig 5, cur 10, worst 10 PASS 66b (84s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 67a: AT: verify slow request processing doesn't induce reconnects ========================================================== 09:36:40 (1713533800) at_history=8 at_history=8 fail_val=400 fail_loc=0x50a fail_loc=0 0 osc reconnect attempts on gradual slow PASS 67a (74s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 67b: AT: verify instant slowdown doesn't induce reconnects ========================================================== 09:37:56 (1713533876) at_history=8 at_history=8 Creating to objid 5473 on ost lustre-OST0000... fail_val=20000 fail_loc=0x80000223 total: 17 open/close in 0.08 seconds: 202.77 ops/second Connected clients: oleg119-client.virtnet oleg119-client.virtnet service : cur 5 worst 5 (at 1713530960, 2941s ago) 1 0 1 0 phase 2 0 osc reconnect attempts on instant slow fail_loc=0x80000223 fail_loc=0 Connected clients: oleg119-client.virtnet oleg119-client.virtnet service : cur 5 worst 5 (at 1713530960, 2942s ago) 1 0 1 0 0 osc reconnect attempts on 2nd slow PASS 67b (27s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 68: AT: verify slowing locks ======= 09:38:25 (1713533905) at_history=8 at_history=8 /home/green/git/lustre-release/lustre/tests/replay-single.sh: line 1970: $ldlm_enqueue_min: ambiguous redirect fail_val=19 fail_loc=0x80000312 fail_val=25 fail_loc=0x80000312 fail_loc=0 /home/green/git/lustre-release/lustre/tests/replay-single.sh: line 1985: $ldlm_enqueue_min: ambiguous redirect PASS 68 (70s) debug_raw_pointers=0 debug_raw_pointers=0 Cleaning up AT ... debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 70a: check multi client t-f ======== 09:39:37 (1713533977) SKIP: replay-single test_70a Need two or more clients, have 1 SKIP 70a (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 70b: dbench 1mdts recovery; 1 clients ========================================================== 09:39:40 (1713533980) Starting client oleg119-client.virtnet: -o user_xattr,flock oleg119-server@tcp:/lustre /mnt/lustre Started clients oleg119-client.virtnet: 192.168.201.119@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) + MISSING_DBENCH_OK= + PATH=/opt/iozone/bin:/opt/iozone/bin:/home/green/git/lustre-release/lustre/tests/mpi:/home/green/git/lustre-release/lustre/tests/racer:/home/green/git/lustre-release/lustre/../lustre-iokit/sgpdd-survey:/home/green/git/lustre-release/lustre/tests:/home/green/git/lustre-release/lustre/utils/gss:/home/green/git/lustre-release/lustre/utils:/opt/iozone/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin::/home/green/git/lustre-release/lustre/utils:/home/green/git/lustre-release/lustre/tests:/sbin:/usr/sbin:/home/green/git/lustre-release/lustre/utils:/home/green/git/lustre-release/lustre/tests/: + DBENCH_LIB= + TESTSUITE=replay-single + TESTNAME=test_70b + MOUNT=/mnt/lustre ++ hostname + DIR=/mnt/lustre/d70b.replay-single/oleg119-client.virtnet + LCTL=/home/green/git/lustre-release/lustre/utils/lctl + rundbench 1 -t 300 dbench: no process found Started rundbench load pid=14013 ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3968 2204544 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 32768 3684352 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3715072 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 36864 7399424 1% /mnt/lustre test_70b fail mds1 1 times Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:39:48 (1713533988) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:40:02 (1713534002) targets are mounted 09:40:02 (1713534002) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec oleg119-client.virtnet: looking for dbench program oleg119-client.virtnet: /usr/bin/dbench oleg119-client.virtnet: creating output directory /mnt/lustre/d70b.replay-single/oleg119-client.virtnet oleg119-client.virtnet: mkdir: created directory '/mnt/lustre/d70b.replay-single' oleg119-client.virtnet: mkdir: created directory '/mnt/lustre/d70b.replay-single/oleg119-client.virtnet' oleg119-client.virtnet: found dbench client file /usr/share/dbench/client.txt oleg119-client.virtnet: '/usr/share/dbench/client.txt' -> 'client.txt' oleg119-client.virtnet: running 'dbench 1 -t 300' on /mnt/lustre/d70b.replay-single/oleg119-client.virtnet at Fri Apr 19 09:39:41 EDT 2024 oleg119-client.virtnet: waiting for dbench pid 14037 oleg119-client.virtnet: dbench version 4.00 - Copyright Andrew Tridgell 1999-2004 oleg119-client.virtnet: oleg119-client.virtnet: Running for 300 seconds with load 'client.txt' and minimum warmup 60 secs oleg119-client.virtnet: failed to create barrier semaphore oleg119-client.virtnet: 0 of 1 processes prepared for launch 0 sec oleg119-client.virtnet: 1 of 1 processes prepared for launch 0 sec oleg119-client.virtnet: releasing clients oleg119-client.virtnet: 1 316 11.48 MB/sec warmup 1 sec latency 23.246 ms oleg119-client.virtnet: 1 671 10.58 MB/sec warmup 2 sec latency 53.431 ms oleg119-client.virtnet: 1 980 7.30 MB/sec warmup 3 sec latency 42.730 ms oleg119-client.virtnet: 1 1481 6.51 MB/sec warmup 4 sec latency 55.644 ms oleg119-client.virtnet: 1 2033 5.31 MB/sec warmup 5 sec latency 65.624 ms oleg119-client.virtnet: 1 2416 4.68 MB/sec warmup 6 sec latency 323.403 ms oleg119-client.virtnet: 1 2416 4.01 MB/sec warmup 7 sec latency 1323.689 ms oleg119-client.virtnet: 1 2416 3.51 MB/sec warmup 8 sec latency 2323.937 ms oleg119-client.virtnet: 1 2416 3.12 MB/sec warmup 9 sec latency 3324.164 ms oleg119-client.virtnet: 1 2416 2.81 MB/sec warmup 10 sec latency 4324.434 ms oleg119-client.virtnet: 1 2416 2.55 MB/sec warmup 11 sec latency 5324.649 ms oleg119-client.virtnet: 1 2416 2.34 MB/sec warmup 12 sec latency 6324.845 ms oleg119-client.virtnet: 1 2416 2.16 MB/sec warmup 13 sec latency 7325.090 ms oleg119-client.virtnet: 1 2416 2.01 MB/sec warmup 14 sec latency 8325.292 ms oleg119-client.virtnet: 1 2416 1.87 MB/sec warmup 15 sec latency 9325.620 ms oleg119-client.virtnet: 1 2416 1.76 MB/sec warmup 16 sec latency 10325.911 ms oleg119-client.virtnet: 1 2416 1.65 MB/sec warmup 17 sec latency 11326.125 ms oleg119-client.virtnet: 1 2416 1.56 MB/sec warmup 18 sec latency 12326.341 ms oleg119-client.virtnet: 1 2416 1.48 MB/sec warmup 19 sec latency 13326.546 ms oleg119-client.virtnet: 1 2416 1.40 MB/sec warmup 20 sec latency 14326.750 ms oleg119-client.virtnet: 1 2416 1.34 MB/sec warmup 21 sec latency 15327.019 ms oleg119-client.virtnet: 1 2416 1.28 MB/sec warmup 22 sec latency 16327.325 ms oleg119-client.virtnet: 1 2416 1.22 MB/sec warmup 23 sec latency 17327.556 ms oleg119-client.virtnet: 1 2416 1.17 MB/sec warmup 24 sec latency 18327.783 ms oleg119-client.virtnet: 1 2416 1.12 MB/sec warmup 25 sec latency 19327.976 ms oleg119-client.virtnet: 1 2416 1.08 MB/sec warmup 26 sec latency 20328.183 ms oleg119-client.virtnet: 1 2416 1.04 MB/sec warmup 27 sec latency 21328.384 ms oleg119-client.virtnet: 1 2416 1.00 MB/sec warmup 28 sec latency 22328.587 ms oleg119-client.virtnet: 1 2416 0.97 MB/sec warmup 29 sec latency 23328.795 ms oleg119-client.virtnet: 1 2865 1.07 MB/sec warmup 30 sec latency 23551.706 ms oleg119-client.virtnet: 1 3378 1.14 MB/sec warmup 31 sec latency 43.762 ms oleg119-client.virtnet: 1 3856 1.24 MB/sec warmup 32 sec laUUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3968 2204544 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 33792 3672064 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 14336 3737600 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 48128 7409664 1% /mnt/lustre test_70b fail mds1 2 times Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:40:19 (1713534019) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:40:33 (1713534033) targets are mounted 09:40:33 (1713534033) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3968 2204544 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 40960 3699712 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 14336 3741696 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 55296 7441408 1% /mnt/lustre test_70b fail mds1 3 times Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:40:44 (1713534044) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:40:58 (1713534058) targets are mounted 09:40:58 (1713534058) facet_failover done tency 15.074 ms oleg119-client.virtnet: 1 4135 1.21 MB/sec warmup 33 sec latency 15.672 ms oleg119-client.virtnet: 1 4462 1.20 MB/sec warmup 34 sec latency 65.848 ms oleg119-client.virtnet: 1 4965 1.29 MB/sec warmup 35 sec latency 43.384 ms oleg119-client.virtnet: 1 5515 1.27 MB/sec warmup 36 sec latency 60.579 ms oleg119-client.virtnet: 1 5976 1.28 MB/sec warmup 37 sec latency 318.561 ms oleg119-client.virtnet: 1 5976 1.24 MB/sec warmup 38 sec latency 1318.819 ms oleg119-client.virtnet: 1 5976 1.21 MB/sec warmup 39 sec latency 2319.225 ms oleg119-client.virtnet: 1 5976 1.18 MB/sec warmup 40 sec latency 3319.559 ms oleg119-client.virtnet: 1 5976 1.15 MB/sec warmup 41 sec latency 4319.863 ms oleg119-client.virtnet: 1 5976 1.12 MB/sec warmup 42 sec latency 5320.232 ms oleg119-client.virtnet: 1 5976 1.10 MB/sec warmup 43 sec latency 6320.552 ms oleg119-client.virtnet: 1 5976 1.07 MB/sec warmup 44 sec latency 7320.838 ms oleg119-client.virtnet: 1 5976 1.05 MB/sec warmup 45 sec latency 8321.254 ms oleg119-client.virtnet: 1 5976 1.03 MB/sec warmup 46 sec latency 9321.537 ms oleg119-client.virtnet: 1 5976 1.00 MB/sec warmup 47 sec latency 10321.828 ms oleg119-client.virtnet: 1 5976 0.98 MB/sec warmup 48 sec latency 11322.133 ms oleg119-client.virtnet: 1 5976 0.96 MB/sec warmup 49 sec latency 12322.382 ms oleg119-client.virtnet: 1 5976 0.94 MB/sec warmup 50 sec latency 13322.618 ms oleg119-client.virtnet: 1 5976 0.93 MB/sec warmup 51 sec latency 14322.822 ms oleg119-client.virtnet: 1 5976 0.91 MB/sec warmup 52 sec latency 15323.045 ms oleg119-client.virtnet: 1 5976 0.89 MB/sec warmup 53 sec latency 16323.326 ms oleg119-client.virtnet: 1 5976 0.87 MB/sec warmup 54 sec latency 17323.599 ms oleg119-client.virtnet: 1 6419 0.93 MB/sec warmup 55 sec latency 17608.121 ms oleg119-client.virtnet: 1 7047 1.03 MB/sec warmup 56 sec latency 42.207 ms oleg119-client.virtnet: 1 7424 1.03 MB/sec warmup 57 sec latency 17.782 ms oleg119-client.virtnet: 1 7778 1.02 MB/sec warmup 58 sec latency 24.785 ms oleg119-client.virtnet: 1 8030 1.02 MB/sec warmup 59 sec latency 65.087 ms oleg119-client.virtnet: 1 9247 0.63 MB/sec execute 1 sec latency 51.171 ms oleg119-client.virtnet: 1 9582 1.34 MB/sec execute 2 sec latency 201.502 ms oleg119-client.virtnet: 1 9582 0.89 MB/sec execute 3 sec latency 1201.758 ms oleg119-client.virtnet: 1 9582 0.67 MB/sec execute 4 sec latency 2202.028 ms oleg119-client.virtnet: 1 9582 0.53 MB/sec execute 5 sec latency 3202.335 ms oleg119-client.virtnet: 1 9582 0.45 MB/sec execute 6 sec latency 4202.612 ms oleg119-client.virtnet: 1 9582 0.38 MB/sec execute 7 sec latency 5202.906 ms oleg119-client.virtnet: 1 9582 0.33 MB/sec execute 8 sec latency 6203.180 ms oleg119-client.virtnet: 1 9582 0.30 MB/sec execute 9 sec latency 7203.461 ms oleg119-client.virtnet: 1 9582 0.27 MB/sec execute 10 sec latency 8203.805 ms oleg119-client.virtnet: 1 9582 0.24 MB/sec execute 11 sec latency 9204.084 ms oleg119-client.virtnet: 1 9582 0.22 MB/sec execute 12 sec latency 10204.335 ms oleg119-client.virtnet: 1 9582 0.21 MB/sec execute 13 sec latency 11204.553 ms oleg119-client.virtnet: 1 9582 0.19 MB/sec execute 14 sec latency 12204.757 ms oleg119-client.virtnet: 1 9582 0.18 MB/sec execute 15 sec latency 13204.933 ms oleg119-client.virtnet: 1 9582 0.17 MB/sec execute 16 sec latency 14205.188 ms oleg119-client.virtnet: 1 9582 0.16 MB/sec execute 17 sec latency 15205.522 ms oleg119-client.virtnoleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3968 2204544 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 41984 3688448 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 14336 3742720 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 56320 7431168 1% /mnt/lustre test_70b fail mds1 4 times Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:41:09 (1713534069) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:41:22 (1713534082) targets are mounted 09:41:22 (1713534082) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3840 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 41984 3706880 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 14336 3741696 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 56320 7448576 1% /mnt/lustre test_70b fail mds1 5 times Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:41:34 (1713534094) shut down et: 1 9582 0.15 MB/sec execute 18 sec latency 16205.772 ms oleg119-client.virtnet: 1 9582 0.14 MB/sec execute 19 sec latency 17205.963 ms oleg119-client.virtnet: 1 10129 0.38 MB/sec execute 20 sec latency 17457.111 ms oleg119-client.virtnet: 1 10915 0.65 MB/sec execute 21 sec latency 26.497 ms oleg119-client.virtnet: 1 11261 0.64 MB/sec execute 22 sec latency 13.758 ms oleg119-client.virtnet: 1 11686 0.69 MB/sec execute 23 sec latency 67.243 ms oleg119-client.virtnet: 1 12243 0.80 MB/sec execute 24 sec latency 52.612 ms oleg119-client.virtnet: 1 13002 0.84 MB/sec execute 25 sec latency 46.348 ms oleg119-client.virtnet: 1 13531 0.97 MB/sec execute 26 sec latency 43.853 ms oleg119-client.virtnet: 1 13839 1.01 MB/sec execute 27 sec latency 486.722 ms oleg119-client.virtnet: 1 13839 0.98 MB/sec execute 28 sec latency 1486.978 ms oleg119-client.virtnet: 1 13839 0.94 MB/sec execute 29 sec latency 2487.205 ms oleg119-client.virtnet: 1 13839 0.91 MB/sec execute 30 sec latency 3487.526 ms oleg119-client.virtnet: 1 13839 0.88 MB/sec execute 31 sec latency 4487.870 ms oleg119-client.virtnet: 1 13839 0.85 MB/sec execute 32 sec latency 5488.167 ms oleg119-client.virtnet: 1 13839 0.83 MB/sec execute 33 sec latency 6488.523 ms oleg119-client.virtnet: 1 13839 0.80 MB/sec execute 34 sec latency 7488.753 ms oleg119-client.virtnet: 1 13839 0.78 MB/sec execute 35 sec latency 8488.981 ms oleg119-client.virtnet: 1 13839 0.76 MB/sec execute 36 sec latency 9489.305 ms oleg119-client.virtnet: 1 13839 0.74 MB/sec execute 37 sec latency 10489.534 ms oleg119-client.virtnet: 1 13839 0.72 MB/sec execute 38 sec latency 11489.776 ms oleg119-client.virtnet: 1 13839 0.70 MB/sec execute 39 sec latency 12489.984 ms oleg119-client.virtnet: 1 13839 0.68 MB/sec execute 40 sec latency 13490.208 ms oleg119-client.virtnet: 1 13839 0.67 MB/sec execute 41 sec latency 14490.463 ms oleg119-client.virtnet: 1 13839 0.65 MB/sec execute 42 sec latency 15490.761 ms oleg119-client.virtnet: 1 13839 0.64 MB/sec execute 43 sec latency 16491.010 ms oleg119-client.virtnet: 1 13839 0.62 MB/sec execute 44 sec latency 17491.302 ms oleg119-client.virtnet: 1 14208 0.70 MB/sec execute 45 sec latency 17881.258 ms oleg119-client.virtnet: 1 14569 0.72 MB/sec execute 46 sec latency 16.153 ms oleg119-client.virtnet: 1 14872 0.71 MB/sec execute 47 sec latency 17.705 ms oleg119-client.virtnet: 1 15120 0.71 MB/sec execute 48 sec latency 72.336 ms oleg119-client.virtnet: 1 15683 0.78 MB/sec execute 49 sec latency 38.242 ms oleg119-client.virtnet: 1 16339 0.78 MB/sec execute 50 sec latency 50.148 ms oleg119-client.virtnet: 1 16772 0.81 MB/sec execute 51 sec latency 49.049 ms oleg119-client.virtnet: 1 17364 0.89 MB/sec execute 52 sec latency 173.413 ms oleg119-client.virtnet: 1 17364 0.87 MB/sec execute 53 sec latency 1173.618 ms oleg119-client.virtnet: 1 17364 0.86 MB/sec execute 54 sec latency 2173.854 ms oleg119-client.virtnet: 1 17364 0.84 MB/sec execute 55 sec latency 3174.077 ms oleg119-client.virtnet: 1 17364 0.83 MB/sec execute 56 sec latency 4174.256 ms oleg119-client.virtnet: 1 17364 0.81 MB/sec execute 57 sec latency 5174.428 ms oleg119-client.virtnet: 1 17364 0.80 MB/sec execute 58 sec latency 6174.658 ms oleg119-client.virtnet: 1 17364 0.78 MB/sec execute 59 sec latency 7174.866 ms oleg119-client.virtnet: 1 17364 0.77 MB/sec execute 60 sec latency 8175.070 ms oleg119-client.virtnet: 1 17364 0.76 MB/sec execute 61 sec latency 9175.257 ms oleg119-client.virtnet: 1 17364Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:41:48 (1713534108) targets are mounted 09:41:48 (1713534108) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3968 2204544 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 40960 3718144 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 14336 3734528 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 55296 7452672 1% /mnt/lustre test_70b fail mds1 6 times Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:41:59 (1713534119) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:42:13 (1713534133) targets are mounted 09:42:13 (1713534133) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3968 2204544 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 41984 3722240 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 17408 3741696 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 59392 7463936 1% /mnt/lustre test_70b fail mds1 7 times Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:42:24 (1713534144) shut down 0.75 MB/sec execute 62 sec latency 10175.428 ms oleg119-client.virtnet: 1 17364 0.74 MB/sec execute 63 sec latency 11175.643 ms oleg119-client.virtnet: 1 17364 0.72 MB/sec execute 64 sec latency 12175.837 ms oleg119-client.virtnet: 1 17364 0.71 MB/sec execute 65 sec latency 13176.053 ms oleg119-client.virtnet: 1 17364 0.70 MB/sec execute 66 sec latency 14176.256 ms oleg119-client.virtnet: 1 17364 0.69 MB/sec execute 67 sec latency 15176.562 ms oleg119-client.virtnet: 1 17364 0.68 MB/sec execute 68 sec latency 16176.796 ms oleg119-client.virtnet: 1 17364 0.67 MB/sec execute 69 sec latency 17176.987 ms oleg119-client.virtnet: 1 17814 0.73 MB/sec execute 70 sec latency 17501.487 ms oleg119-client.virtnet: 1 18132 0.74 MB/sec execute 71 sec latency 16.180 ms oleg119-client.virtnet: 1 18420 0.73 MB/sec execute 72 sec latency 56.664 ms oleg119-client.virtnet: 1 18680 0.73 MB/sec execute 73 sec latency 71.379 ms oleg119-client.virtnet: 1 19239 0.78 MB/sec execute 74 sec latency 53.721 ms oleg119-client.virtnet: 1 20013 0.78 MB/sec execute 75 sec latency 37.764 ms oleg119-client.virtnet: 1 20387 0.79 MB/sec execute 76 sec latency 53.474 ms oleg119-client.virtnet: 1 21050 0.87 MB/sec execute 77 sec latency 94.676 ms oleg119-client.virtnet: 1 21050 0.85 MB/sec execute 78 sec latency 1094.862 ms oleg119-client.virtnet: 1 21050 0.84 MB/sec execute 79 sec latency 2095.085 ms oleg119-client.virtnet: 1 21050 0.83 MB/sec execute 80 sec latency 3095.327 ms oleg119-client.virtnet: 1 21050 0.82 MB/sec execute 81 sec latency 4095.526 ms oleg119-client.virtnet: 1 21050 0.81 MB/sec execute 82 sec latency 5095.735 ms oleg119-client.virtnet: 1 21050 0.80 MB/sec execute 83 sec latency 6095.939 ms oleg119-client.virtnet: 1 21050 0.79 MB/sec execute 84 sec latency 7096.126 ms oleg119-client.virtnet: 1 21050 0.78 MB/sec execute 85 sec latency 8096.354 ms oleg119-client.virtnet: 1 21050 0.78 MB/sec execute 86 sec latency 9096.556 ms oleg119-client.virtnet: 1 21050 0.77 MB/sec execute 87 sec latency 10096.790 ms oleg119-client.virtnet: 1 21050 0.76 MB/sec execute 88 sec latency 11096.995 ms oleg119-client.virtnet: 1 21050 0.75 MB/sec execute 89 sec latency 12097.251 ms oleg119-client.virtnet: 1 21050 0.74 MB/sec execute 90 sec latency 13097.436 ms oleg119-client.virtnet: 1 21050 0.73 MB/sec execute 91 sec latency 14097.621 ms oleg119-client.virtnet: 1 21050 0.72 MB/sec execute 92 sec latency 15097.811 ms oleg119-client.virtnet: 1 21050 0.72 MB/sec execute 93 sec latency 16098.010 ms oleg119-client.virtnet: 1 21050 0.71 MB/sec execute 94 sec latency 17098.240 ms oleg119-client.virtnet: 1 21415 0.74 MB/sec execute 95 sec latency 17532.255 ms oleg119-client.virtnet: 1 21724 0.74 MB/sec execute 96 sec latency 20.943 ms oleg119-client.virtnet: 1 22006 0.74 MB/sec execute 97 sec latency 64.672 ms oleg119-client.virtnet: 1 22304 0.74 MB/sec execute 98 sec latency 44.159 ms oleg119-client.virtnet: 1 22834 0.77 MB/sec execute 99 sec latency 57.087 ms oleg119-client.virtnet: 1 23588 0.77 MB/sec execute 100 sec latency 49.191 ms oleg119-client.virtnet: 1 23933 0.79 MB/sec execute 101 sec latency 50.733 ms oleg119-client.virtnet: 1 24662 0.84 MB/sec execute 102 sec latency 46.654 ms oleg119-client.virtnet: 1 24680 0.83 MB/sec execute 103 sec latency 928.755 ms oleg119-client.virtnet: 1 24680 0.83 MB/sec execute 104 sec latency 1928.975 ms oleg119-client.virtnet: 1 24680 0.82 MB/sec execute 105 sec latency 2929.207 ms oleg119-client.virtnet: 1 24680 0.81 MB/seFailover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:42:38 (1713534158) targets are mounted 09:42:38 (1713534158) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3840 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 41984 3708928 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 17408 3723264 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 59392 7432192 1% /mnt/lustre test_70b fail mds1 8 times Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:42:49 (1713534169) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:43:03 (1713534183) targets are mounted 09:43:03 (1713534183) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3968 2204544 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 41984 3718144 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 17408 3744768 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 59392 7462912 1% /mnt/lustre c execute 106 sec latency 3929.471 ms oleg119-client.virtnet: 1 24680 0.80 MB/sec execute 107 sec latency 4929.709 ms oleg119-client.virtnet: 1 24680 0.80 MB/sec execute 108 sec latency 5929.976 ms oleg119-client.virtnet: 1 24680 0.79 MB/sec execute 109 sec latency 6930.255 ms oleg119-client.virtnet: 1 24680 0.78 MB/sec execute 110 sec latency 7930.522 ms oleg119-client.virtnet: 1 24680 0.77 MB/sec execute 111 sec latency 8930.833 ms oleg119-client.virtnet: 1 24680 0.77 MB/sec execute 112 sec latency 9931.098 ms oleg119-client.virtnet: 1 24680 0.76 MB/sec execute 113 sec latency 10931.346 ms oleg119-client.virtnet: 1 24680 0.75 MB/sec execute 114 sec latency 11931.552 ms oleg119-client.virtnet: 1 24680 0.75 MB/sec execute 115 sec latency 12931.789 ms oleg119-client.virtnet: 1 24680 0.74 MB/sec execute 116 sec latency 13931.972 ms oleg119-client.virtnet: 1 24680 0.73 MB/sec execute 117 sec latency 14932.197 ms oleg119-client.virtnet: 1 24680 0.73 MB/sec execute 118 sec latency 15932.404 ms oleg119-client.virtnet: 1 24680 0.72 MB/sec execute 119 sec latency 16932.612 ms oleg119-client.virtnet: 1 25030 0.74 MB/sec execute 120 sec latency 17348.014 ms oleg119-client.virtnet: 1 25307 0.75 MB/sec execute 121 sec latency 15.828 ms oleg119-client.virtnet: 1 25597 0.74 MB/sec execute 122 sec latency 70.362 ms oleg119-client.virtnet: 1 25877 0.74 MB/sec execute 123 sec latency 50.229 ms oleg119-client.virtnet: 1 26451 0.77 MB/sec execute 124 sec latency 48.332 ms oleg119-client.virtnet: 1 27221 0.78 MB/sec execute 125 sec latency 44.594 ms oleg119-client.virtnet: 1 27640 0.81 MB/sec execute 126 sec latency 48.847 ms oleg119-client.virtnet: 1 28235 0.83 MB/sec execute 127 sec latency 51.733 ms oleg119-client.virtnet: 1 28247 0.82 MB/sec execute 128 sec latency 951.003 ms oleg119-client.virtnet: 1 28247 0.81 MB/sec execute 129 sec latency 1951.226 ms oleg119-client.virtnet: 1 28247 0.81 MB/sec execute 130 sec latency 2951.527 ms oleg119-client.virtnet: 1 28247 0.80 MB/sec execute 131 sec latency 3951.760 ms oleg119-client.virtnet: 1 28247 0.80 MB/sec execute 132 sec latency 4952.002 ms oleg119-client.virtnet: 1 28247 0.79 MB/sec execute 133 sec latency 5952.249 ms oleg119-client.virtnet: 1 28247 0.78 MB/sec execute 134 sec latency 6952.429 ms oleg119-client.virtnet: 1 28247 0.78 MB/sec execute 135 sec latency 7952.568 ms oleg119-client.virtnet: 1 28247 0.77 MB/sec execute 136 sec latency 8952.745 ms oleg119-client.virtnet: 1 28247 0.77 MB/sec execute 137 sec latency 9952.918 ms oleg119-client.virtnet: 1 28247 0.76 MB/sec execute 138 sec latency 10953.037 ms oleg119-client.virtnet: 1 28247 0.76 MB/sec execute 139 sec latency 11953.205 ms oleg119-client.virtnet: 1 28247 0.75 MB/sec execute 140 sec latency 12953.394 ms oleg119-client.virtnet: 1 28247 0.75 MB/sec execute 141 sec latency 13953.554 ms oleg119-client.virtnet: 1 28247 0.74 MB/sec execute 142 sec latency 14953.758 ms oleg119-client.virtnet: 1 28247 0.73 MB/sec execute 143 sec latency 15954.020 ms oleg119-client.virtnet: 1 28247 0.73 MB/sec execute 144 sec latency 16954.256 ms oleg119-client.virtnet: 1 28636 0.75 MB/sec execute 145 sec latency 17378.909 ms oleg119-client.virtnet: 1 28875 0.75 MB/sec execute 146 sec latency 18.470 ms oleg119-client.virtnet: 1 29158 0.75 MB/sec execute 147 sec latency 64.430 ms oleg119-client.virtnet: 1 29527 0.75 MB/sec execute 148 sec latency 43.461 ms oleg119-client.virtnet: 1 30066 0.77 MB/sec execute 149 sec latency 51.899 ms oleg119-client.virtnet: 1 30820 0.78 MB/sec execute test_70b fail mds1 9 times Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:43:14 (1713534194) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:43:28 (1713534208) targets are mounted 09:43:28 (1713534208) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3968 2204544 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 45056 3704832 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 14336 3748864 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 59392 7453696 1% /mnt/lustre test_70b fail mds1 10 times Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:43:39 (1713534219) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:43:53 (1713534233) targets are mounted 09:43:53 (1713534233) facet_failover done 150 sec latency 40.884 ms oleg119-client.virtnet: 1 31340 0.81 MB/sec execute 151 sec latency 49.369 ms oleg119-client.virtnet: 1 32034 0.84 MB/sec execute 152 sec latency 56.173 ms oleg119-client.virtnet: 1 32034 0.83 MB/sec execute 153 sec latency 1056.397 ms oleg119-client.virtnet: 1 32034 0.83 MB/sec execute 154 sec latency 2056.611 ms oleg119-client.virtnet: 1 32034 0.82 MB/sec execute 155 sec latency 3056.859 ms oleg119-client.virtnet: 1 32034 0.82 MB/sec execute 156 sec latency 4057.162 ms oleg119-client.virtnet: 1 32034 0.81 MB/sec execute 157 sec latency 5057.471 ms oleg119-client.virtnet: 1 32034 0.81 MB/sec execute 158 sec latency 6057.701 ms oleg119-client.virtnet: 1 32034 0.80 MB/sec execute 159 sec latency 7057.923 ms oleg119-client.virtnet: 1 32034 0.80 MB/sec execute 160 sec latency 8058.147 ms oleg119-client.virtnet: 1 32034 0.79 MB/sec execute 161 sec latency 9058.421 ms oleg119-client.virtnet: 1 32034 0.79 MB/sec execute 162 sec latency 10058.617 ms oleg119-client.virtnet: 1 32034 0.78 MB/sec execute 163 sec latency 11058.840 ms oleg119-client.virtnet: 1 32034 0.78 MB/sec execute 164 sec latency 12059.077 ms oleg119-client.virtnet: 1 32034 0.77 MB/sec execute 165 sec latency 13059.305 ms oleg119-client.virtnet: 1 32034 0.77 MB/sec execute 166 sec latency 14059.489 ms oleg119-client.virtnet: 1 32034 0.76 MB/sec execute 167 sec latency 15059.678 ms oleg119-client.virtnet: 1 32034 0.76 MB/sec execute 168 sec latency 16059.865 ms oleg119-client.virtnet: 1 32034 0.75 MB/sec execute 169 sec latency 17060.065 ms oleg119-client.virtnet: 1 32206 0.76 MB/sec execute 170 sec latency 17602.043 ms oleg119-client.virtnet: 1 32440 0.75 MB/sec execute 171 sec latency 15.531 ms oleg119-client.virtnet: 1 32709 0.75 MB/sec execute 172 sec latency 61.777 ms oleg119-client.virtnet: 1 33098 0.76 MB/sec execute 173 sec latency 46.018 ms oleg119-client.virtnet: 1 33656 0.77 MB/sec execute 174 sec latency 46.498 ms oleg119-client.virtnet: 1 34367 0.78 MB/sec execute 175 sec latency 44.813 ms oleg119-client.virtnet: 1 34893 0.80 MB/sec execute 176 sec latency 49.401 ms oleg119-client.virtnet: 1 35587 0.83 MB/sec execute 177 sec latency 42.500 ms oleg119-client.virtnet: 1 35587 0.82 MB/sec execute 178 sec latency 1001.142 ms oleg119-client.virtnet: 1 35587 0.82 MB/sec execute 179 sec latency 2001.468 ms oleg119-client.virtnet: 1 35587 0.81 MB/sec execute 180 sec latency 3001.681 ms oleg119-client.virtnet: 1 35587 0.81 MB/sec execute 181 sec latency 4001.933 ms oleg119-client.virtnet: 1 35587 0.80 MB/sec execute 182 sec latency 5002.294 ms oleg119-client.virtnet: 1 35587 0.80 MB/sec execute 183 sec latency 6002.678 ms oleg119-client.virtnet: 1 35587 0.80 MB/sec execute 184 sec latency 7002.930 ms oleg119-client.virtnet: 1 35587 0.79 MB/sec execute 185 sec latency 8003.190 ms oleg119-client.virtnet: 1 35587 0.79 MB/sec execute 186 sec latency 9003.438 ms oleg119-client.virtnet: 1 35587 0.78 MB/sec execute 187 sec latency 10003.672 ms oleg119-client.virtnet: 1 35587 0.78 MB/sec execute 188 sec latency 11003.809 ms oleg119-client.virtnet: 1 35587 0.77 MB/sec execute 189 sec latency 12004.077 ms oleg119-client.virtnet: 1 35587 0.77 MB/sec execute 190 sec latency 13004.295 ms oleg119-client.virtnet: 1 35587 0.77 MB/sec execute 191 sec latency 14004.529 ms oleg119-client.virtnet: 1 35587 0.76 MB/sec execute 192 sec latency 15004.832 ms oleg119-client.virtnet: 1 35587 0.76 MB/sec execute 193 sec latency 16005.105 ms oleg119-client.virtnet: 1 35587 0.75 MB/sec execute 194 sec oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3968 2204544 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 44032 3702784 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 14336 3740672 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 58368 7443456 1% /mnt/lustre test_70b fail mds1 11 times Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:44:04 (1713534244) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:44:18 (1713534258) targets are mounted 09:44:18 (1713534258) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3840 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 43008 3707904 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 14336 3748864 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 57344 7456768 1% /mnt/lustre test_70b fail mds1 12 times Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:44:30 (1713534270) shut down Failover mds1 to oleg119-server mount facets: mds1 latency 17005.331 ms oleg119-client.virtnet: 1 35766 0.76 MB/sec execute 195 sec latency 17522.269 ms oleg119-client.virtnet: 1 36035 0.75 MB/sec execute 196 sec latency 15.816 ms oleg119-client.virtnet: 1 36300 0.75 MB/sec execute 197 sec latency 66.875 ms oleg119-client.virtnet: 1 36672 0.76 MB/sec execute 198 sec latency 47.370 ms oleg119-client.virtnet: 1 37212 0.77 MB/sec execute 199 sec latency 48.883 ms oleg119-client.virtnet: 1 37925 0.78 MB/sec execute 200 sec latency 55.008 ms oleg119-client.virtnet: 1 38477 0.80 MB/sec execute 201 sec latency 52.603 ms oleg119-client.virtnet: 1 39135 0.82 MB/sec execute 202 sec latency 44.381 ms oleg119-client.virtnet: 1 39169 0.82 MB/sec execute 203 sec latency 935.999 ms oleg119-client.virtnet: 1 39169 0.81 MB/sec execute 204 sec latency 1936.186 ms oleg119-client.virtnet: 1 39169 0.81 MB/sec execute 205 sec latency 2936.365 ms oleg119-client.virtnet: 1 39169 0.80 MB/sec execute 206 sec latency 3936.568 ms oleg119-client.virtnet: 1 39169 0.80 MB/sec execute 207 sec latency 4936.794 ms oleg119-client.virtnet: 1 39169 0.80 MB/sec execute 208 sec latency 5936.972 ms oleg119-client.virtnet: 1 39169 0.79 MB/sec execute 209 sec latency 6937.155 ms oleg119-client.virtnet: 1 39169 0.79 MB/sec execute 210 sec latency 7937.340 ms oleg119-client.virtnet: 1 39169 0.79 MB/sec execute 211 sec latency 8937.527 ms oleg119-client.virtnet: 1 39169 0.78 MB/sec execute 212 sec latency 9937.733 ms oleg119-client.virtnet: 1 39169 0.78 MB/sec execute 213 sec latency 10937.928 ms oleg119-client.virtnet: 1 39169 0.77 MB/sec execute 214 sec latency 11938.068 ms oleg119-client.virtnet: 1 39169 0.77 MB/sec execute 215 sec latency 12938.217 ms oleg119-client.virtnet: 1 39169 0.77 MB/sec execute 216 sec latency 13938.411 ms oleg119-client.virtnet: 1 39169 0.76 MB/sec execute 217 sec latency 14938.681 ms oleg119-client.virtnet: 1 39169 0.76 MB/sec execute 218 sec latency 15938.864 ms oleg119-client.virtnet: 1 39169 0.76 MB/sec execute 219 sec latency 16939.047 ms oleg119-client.virtnet: 1 39315 0.76 MB/sec execute 220 sec latency 17511.761 ms oleg119-client.virtnet: 1 39580 0.76 MB/sec execute 221 sec latency 18.905 ms oleg119-client.virtnet: 1 39838 0.75 MB/sec execute 222 sec latency 82.228 ms oleg119-client.virtnet: 1 40191 0.76 MB/sec execute 223 sec latency 46.898 ms oleg119-client.virtnet: 1 40705 0.77 MB/sec execute 224 sec latency 56.985 ms oleg119-client.virtnet: 1 41447 0.77 MB/sec execute 225 sec latency 49.317 ms oleg119-client.virtnet: 1 41972 0.79 MB/sec execute 226 sec latency 45.600 ms oleg119-client.virtnet: 1 42582 0.81 MB/sec execute 227 sec latency 52.767 ms oleg119-client.virtnet: 1 42721 0.81 MB/sec execute 228 sec latency 739.406 ms oleg119-client.virtnet: 1 42721 0.81 MB/sec execute 229 sec latency 1739.636 ms oleg119-client.virtnet: 1 42721 0.80 MB/sec execute 230 sec latency 2739.879 ms oleg119-client.virtnet: 1 42721 0.80 MB/sec execute 231 sec latency 3740.116 ms oleg119-client.virtnet: 1 42721 0.80 MB/sec execute 232 sec latency 4740.355 ms oleg119-client.virtnet: 1 42721 0.79 MB/sec execute 233 sec latency 5740.596 ms oleg119-client.virtnet: 1 42721 0.79 MB/sec execute 234 sec latency 6740.808 ms oleg119-client.virtnet: 1 42721 0.79 MB/sec execute 235 sec latency 7740.998 ms oleg119-client.virtnet: 1 42721 0.78 MB/sec execute 236 sec latency 8741.203 ms oleg119-client.virtnet: 1 42721 0.78 MB/sec execute 237 sec latency 9741.487 ms oleg119-client.virtnet: 1 42721 0.78 MB/sec execute 238 sec latency 10741.696 ms olStarting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:44:43 (1713534283) targets are mounted 09:44:43 (1713534283) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec eg119-client.virtnet: 1 42721 0.77 MB/sec execute 239 sec latency 11741.936 ms oleg119-client.virtnet: 1 42721 0.77 MB/sec execute 240 sec latency 12742.125 ms oleg119-client.virtnet: 1 42721 0.77 MB/sec execute 241 sec latency 13742.332 ms oleg119-client.virtnet: 1 42721 0.76 MB/sec execute 242 sec latency 14742.553 ms oleg119-client.virtnet: 1 42721 0.76 MB/sec execute 243 sec latency 15742.800 ms oleg119-client.virtnet: 1 42721 0.76 MB/sec execute 244 sec latency 16742.992 ms oleg119-client.virtnet: 1 42863 0.76 MB/sec execute 245 sec latency 17338.841 ms oleg119-client.virtnet: 1 43137 0.76 MB/sec execute 246 sec latency 17.521 ms oleg119-client.virtnet: 1 43405 0.76 MB/sec execute 247 sec latency 62.256 ms oleg119-client.virtnet: 1 43810 0.76 MB/sec execute 248 sec latency 44.410 ms oleg119-client.virtnet: 1 44458 0.77 MB/sec execute 249 sec latency 41.287 ms oleg119-client.virtnet: 1 45049 0.77 MB/sec execute 250 sec latency 42.936 ms oleg119-client.virtnet: 1 45610 0.79 MB/sec execute 251 sec latency 46.653 ms oleg119-client.virtnet: 1 46348 0.81 MB/sec execute 252 sec latency 41.343 ms oleg119-client.virtnet: 1 46591 0.81 MB/sec execute 253 sec latency 15.804 ms oleg119-client.virtnet: 1 46937 0.81 MB/sec execute 254 sec latency 68.598 ms oleg119-client.virtnet: 1 47302 0.81 MB/sec execute 255 sec latency 44.893 ms oleg119-client.virtnet: 1 47844 0.82 MB/sec execute 256 sec latency 53.813 ms oleg119-client.virtnet: 1 48561 0.83 MB/sec execute 257 sec latency 41.863 ms oleg119-client.virtnet: 1 49112 0.84 MB/sec execute 258 sec latency 49.023 ms oleg119-client.virtnet: 1 49783 0.86 MB/sec execute 259 sec latency 39.367 ms oleg119-client.virtnet: 1 50106 0.86 MB/sec execute 260 sec latency 16.145 ms oleg119-client.virtnet: 1 50430 0.86 MB/sec execute 261 sec latency 62.535 ms oleg119-client.virtnet: 1 50813 0.86 MB/sec execute 262 sec latency 43.224 ms oleg119-client.virtnet: 1 51345 0.87 MB/sec execute 263 sec latency 53.037 ms oleg119-client.virtnet: 1 52092 0.88 MB/sec execute 264 sec latency 40.373 ms oleg119-client.virtnet: 1 52630 0.89 MB/sec execute 265 sec latency 44.189 ms oleg119-client.virtnet: 1 53343 0.91 MB/sec execute 266 sec latency 51.835 ms oleg119-client.virtnet: 1 53642 0.91 MB/sec execute 267 sec latency 16.313 ms oleg119-client.virtnet: 1 53971 0.91 MB/sec execute 268 sec latency 72.508 ms oleg119-client.virtnet: 1 54319 0.91 MB/sec execute 269 sec latency 43.242 ms oleg119-client.virtnet: 1 54876 0.92 MB/sec execute 270 sec latency 46.787 ms oleg119-client.virtnet: 1 55614 0.92 MB/sec execute 271 sec latency 44.510 ms oleg119-client.virtnet: 1 56154 0.94 MB/sec execute 272 sec latency 51.305 ms oleg119-client.virtnet: 1 56782 0.96 MB/sec execute 273 sec latency 58.485 ms oleg119-client.virtnet: 1 57151 0.96 MB/sec execute 274 sec latency 16.267 ms oleg119-client.virtnet: 1 57462 0.96 MB/sec execute 275 sec latency 74.947 ms oleg119-client.virtnet: 1 57724 0.95 MB/sec execute 276 sec latency 52.070 ms oleg119-client.virtnet: 1 58273 0.97 MB/sec execute 277 sec latency 56.051 ms oleg119-client.virtnet: 1 59053 0.97 MB/sec execute 278 sec latency 44.880 ms oleg119-client.virtnet: 1 59497 0.98 MB/sec execute 279 sec latency 47.865 ms oleg119-client.virtnet: 1 60119 0.99 MB/sec execute 280 sec latency 44.991 ms oleg119-client.virtnet: 1 60615 1.00 MB/sec execute 281 sec latency 16.047 ms oleg119-client.virtnet: 1 60906 1.00 MB/sec execute 282 sec latency 16.466 ms oleg119-client.virtnet: 1 61212 1.00 MB/sec execute 283 sec latency 81.850 ms oleg119-client.virtnet: 1 61732 1.01 MB/sec execute 284 sec latency 44.430 ms oleg119-client.virtnet: 1 62365 1.01 MB/sec execute 285 sec latency 52.253 ms oleg119-client.virtnet: 1 62886 1.01 MB/sec execute 286 sec latency 47.502 ms oleg119-client.virtnet: 1 63523 1.03 MB/sec execute 287 sec latency 43.383 ms oleg119-client.virtnet: 1 64117 1.04 MB/sec execute 288 sec latency 45.147 ms oleg119-client.virtnet: 1 64354 1.04 MB/sec execute 289 sec latency 16.069 ms oleg119-client.virtnet: 1 64680 1.04 MB/sec execute 290 sec latency 70.798 ms oleg119-client.virtnet: 1 65100 1.04 MB/sec execute 291 sec latency 45.101 ms oleg119-client.virtnet: 1 65760 1.05 MB/sec execute 292 sec latency 48.490 ms oleg119-client.virtnet: 1 66349 1.05 MB/sec execute 293 sec latency 44.249 ms oleg119-client.virtnet: 1 67043 1.07 MB/sec execute 294 sec latency 43.474 ms oleg119-client.virtnet: 1 67649 1.08 MB/sec execute 295 sec latency 38.477 ms oleg119-client.virtnet: 1 67928 1.08 MB/sec execute 296 sec latency 16.161 ms oleg119-client.virtnet: 1 68344 1.08 MB/sec execute 297 sec latency 64.357 ms oleg119-client.virtnet: 1 68916 1.09 MB/sec execute 298 sec latency 37.537 ms oleg119-client.virtnet: 1 69887 1.10 MB/sec execute 299 sec latency 28.962 ms oleg119-client.virtnet: 1 cleanup 300 sec oleg119-client.virtnet: 0 cleanup 301 sec oleg119-client.virtnet: oleg119-client.virtnet: Operation Count AvgLat MaxLat oleg119-client.virtnet: ---------------------------------------- oleg119-client.virtnet: NTCreateX 10700 7.480 17881.233 oleg119-client.virtnet: Close 7877 4.986 17501.457 oleg119-client.virtnet: Rename 453 50.549 17601.992 oleg119-client.virtnet: Unlink 2136 11.962 17378.881 oleg119-client.virtnet: Qpathinfo 9736 7.709 17532.236 oleg119-client.virtnet: Qfileinfo 1711 0.198 2.484 oleg119-client.virtnet: Qfsinfo 1759 0.156 5.409 oleg119-client.virtnet: Sfileinfo 866 27.409 17457.098 oleg119-client.virtnet: Find 3739 0.655 12.655 oleg119-client.virtnet: WriteX 5300 1.198 9.706 oleg119-client.virtnet: ReadX 16906 0.024 3.138 oleg119-client.virtnet: LockX 36 1.942 2.200 oleg119-client.virtnet: UnlockX 36 2.011 2.317 oleg119-client.virtnet: Flush 745 30.903 82.216 oleg119-client.virtnet: oleg119-client.virtnet: Throughput 1.09572 MB/sec 1 clients 1 procs max_latency=17881.258 ms oleg119-client.virtnet: stopping dbench on /mnt/lustre/d70b.replay-single/oleg119-client.virtnet at Fri Apr 19 09:45:42 EDT 2024 with return code 0 oleg119-client.virtnet: clean dbench files on /mnt/lustre/d70b.replay-single/oleg119-client.virtnet oleg119-client.virtnet: /mnt/lustre/d70b.replay-single/oleg119-client.virtnet /mnt/lustre/d70b.replay-single/oleg119-client.virtnet oleg119-client.virtnet: removed directory: 'clients/client0/~dmtmp/WORD' oleg119-client.virtnet: removed directory: 'clients/client0/~dmtmp/SEED' oleg119-client.virtnet: removed directory: 'clients/client0/~dmtmp/EXCEL' oleg119-client.virtnet: removed directory: 'clients/client0/~dmtmp/WORDPRO' oleg119-client.virtnet: removed directory: 'clients/client0/~dmtmp/PM' oleg119-client.virtnet: removed directory: 'clients/client0/~dmtmp/ACCESS' oleg119-client.virtnet: removed directory: 'clients/client0/~dmtmp/COREL' oleg119-client.virtnet: removed directory: 'clients/client0/~dmtmp/PARADOX' oleg119-client.virtnet: removed directory: 'clients/client0/~dmtmp/PWRPNT' oleg119-client.virtnet: removed directory: 'clients/client0/~dmtmp' oleg119-client.virtnet: removed directory: 'clients/client0' oleg119-client.virtnet: removed directory: 'clients' oleg119-client.virtnet: removed 'client.txt' oleg119-client.virtnet: /mnt/lustre/d70b.replay-single/oleg119-client.virtnet oleg119-client.virtnet: dbench successfully finished PASS 70b (364s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 70c: tar 1mdts recovery ============ 09:45:46 (1713534346) Starting client oleg119-client.virtnet: -o user_xattr,flock oleg119-server@tcp:/lustre /mnt/lustre Started clients oleg119-client.virtnet: 192.168.201.119@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Started tar 27208 tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 12032 2196480 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 22528 3254272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 18432 3251200 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 40960 6505472 1% /mnt/lustre test_70c fail mds1 1 times Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:48:01 (1713534481) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:48:14 (1713534494) targets are mounted 09:48:14 (1713534494) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 11904 2196608 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 23552 3179520 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 18432 3187712 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 41984 6367232 1% /mnt/lustre test_70c fail mds1 2 times Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:50:36 (1713534636) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:50:50 (1713534650) targets are mounted 09:50:50 (1713534650) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 70c (372s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 70d: mkdir/rmdir striped dir 1mdts recovery ========================================================== 09:52:00 (1713534720) SKIP: replay-single test_70d needs >= 2 MDTs SKIP 70d (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 70e: rename cross-MDT with random fails ========================================================== 09:52:04 (1713534724) SKIP: replay-single test_70e needs >= 2 MDTs SKIP 70e (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 70f: OSS O_DIRECT recovery with 1 clients ========================================================== 09:52:07 (1713534727) mount clients oleg119-client.virtnet ... Starting client oleg119-client.virtnet: -o user_xattr,flock oleg119-server@tcp:/lustre /mnt/lustre Started clients oleg119-client.virtnet: 192.168.201.119@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 4096 2204416 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 10240 3734528 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 8192 3744768 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 18432 7479296 1% /mnt/lustre Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... test_70f failing OST 1 times Failing ost1 on oleg119-server Stopping /mnt/lustre-ost1 (opts:) on oleg119-server 09:52:15 (1713534735) shut down Failover ost1 to oleg119-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... Started lustre-OST0000 09:52:30 (1713534750) targets are mounted 09:52:30 (1713534750) facet_failover done ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 4096 2204416 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 8192 3749888 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 8192 3744768 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 16384 7494656 1% /mnt/lustre ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... test_70f failing OST 2 times Failing ost1 on oleg119-server Stopping /mnt/lustre-ost1 (opts:) on oleg119-server 09:52:43 (1713534763) shut down Failover ost1 to oleg119-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... Started lustre-OST0000 09:52:57 (1713534777) targets are mounted 09:52:57 (1713534777) facet_failover done ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 4096 2204416 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 10240 3727360 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 8192 3737600 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 18432 7464960 1% /mnt/lustre ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... test_70f failing OST 3 times Failing ost1 on oleg119-server Stopping /mnt/lustre-ost1 (opts:) on oleg119-server 09:53:11 (1713534791) shut down Failover ost1 to oleg119-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... Started lustre-OST0000 09:53:25 (1713534805) targets are mounted 09:53:25 (1713534805) facet_failover done ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 4096 2204416 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 10240 3751936 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 8192 3753984 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 18432 7505920 1% /mnt/lustre ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... test_70f failing OST 4 times Failing ost1 on oleg119-server Stopping /mnt/lustre-ost1 (opts:) on oleg119-server 09:53:38 (1713534818) shut down Failover ost1 to oleg119-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... Started lustre-OST0000 09:53:52 (1713534832) targets are mounted 09:53:52 (1713534832) facet_failover done ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 4096 2204416 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 10240 3741696 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 8192 3744768 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 18432 7486464 1% /mnt/lustre ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... test_70f failing OST 5 times Failing ost1 on oleg119-server Stopping /mnt/lustre-ost1 (opts:) on oleg119-server 09:54:05 (1713534845) shut down Failover ost1 to oleg119-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Started lustre-OST0000 09:54:19 (1713534859) targets are mounted 09:54:19 (1713534859) facet_failover done Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... ldlm.namespaces.MGC192.168.201.119@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800a9887800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800a9887800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg119-client.virtnet' ... PASS 70f (140s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 71a: mkdir/rmdir striped dir with 2 mdts recovery ========================================================== 09:54:30 (1713534870) SKIP: replay-single test_71a needs >= 2 MDTs SKIP 71a (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 73a: open(O_CREAT), unlink, replay, reconnect before open replay, close ========================================================== 09:54:33 (1713534873) multiop /mnt/lustre/f73a.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6753 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 4096 2204416 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre fail_loc=0x80000302 Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:54:37 (1713534877) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:54:51 (1713534891) targets are mounted 09:54:51 (1713534891) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 73a (38s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 73b: open(O_CREAT), unlink, replay, reconnect at open_replay reply, close ========================================================== 09:55:13 (1713534913) multiop /mnt/lustre/f73b.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6753 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3840 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre fail_loc=0x80000157 Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:55:17 (1713534917) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:55:31 (1713534931) targets are mounted 09:55:31 (1713534931) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 73b (38s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 74: Ensure applications don't fail waiting for OST recovery ========================================================== 09:55:53 (1713534953) Stopping clients: oleg119-client.virtnet /mnt/lustre (opts:) Stopping client oleg119-client.virtnet /mnt/lustre opts: Stopping /mnt/lustre-ost1 (opts:) on oleg119-server Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:55:57 (1713534957) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:56:11 (1713534971) targets are mounted 09:56:11 (1713534971) facet_failover done Starting client oleg119-client.virtnet: -o user_xattr,flock oleg119-server@tcp:/lustre /mnt/lustre Started clients oleg119-client.virtnet: 192.168.201.119@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-OST0000 PASS 74 (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80a: DNE: create remote dir, drop update rep from MDT0, fail MDT0 ========================================================== 09:56:21 (1713534981) SKIP: replay-single test_80a needs >= 2 MDTs SKIP 80a (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80b: DNE: create remote dir, drop update rep from MDT0, fail MDT1 ========================================================== 09:56:24 (1713534984) SKIP: replay-single test_80b needs >= 2 MDTs SKIP 80b (2s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80c: DNE: create remote dir, drop update rep from MDT1, fail MDT[0,1] ========================================================== 09:56:28 (1713534988) SKIP: replay-single test_80c needs >= 2 MDTs SKIP 80c (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80d: DNE: create remote dir, drop update rep from MDT1, fail 2 MDTs ========================================================== 09:56:31 (1713534991) SKIP: replay-single test_80d needs >= 2 MDTs SKIP 80d (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80e: DNE: create remote dir, drop MDT1 rep, fail MDT0 ========================================================== 09:56:35 (1713534995) SKIP: replay-single test_80e needs >= 2 MDTs SKIP 80e (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80f: DNE: create remote dir, drop MDT1 rep, fail MDT1 ========================================================== 09:56:38 (1713534998) SKIP: replay-single test_80f needs >= 2 MDTs SKIP 80f (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80g: DNE: create remote dir, drop MDT1 rep, fail MDT0, then MDT1 ========================================================== 09:56:41 (1713535001) SKIP: replay-single test_80g needs >= 2 MDTs SKIP 80g (2s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80h: DNE: create remote dir, drop MDT1 rep, fail 2 MDTs ========================================================== 09:56:45 (1713535005) SKIP: replay-single test_80h needs >= 2 MDTs SKIP 80h (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81a: DNE: unlink remote dir, drop MDT0 update rep, fail MDT1 ========================================================== 09:56:48 (1713535008) SKIP: replay-single test_81a needs >= 2 MDTs SKIP 81a (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81b: DNE: unlink remote dir, drop MDT0 update reply, fail MDT0 ========================================================== 09:56:52 (1713535012) SKIP: replay-single test_81b needs >= 2 MDTs SKIP 81b (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81c: DNE: unlink remote dir, drop MDT0 update reply, fail MDT0,MDT1 ========================================================== 09:56:55 (1713535015) SKIP: replay-single test_81c needs >= 2 MDTs SKIP 81c (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81d: DNE: unlink remote dir, drop MDT0 update reply, fail 2 MDTs ========================================================== 09:56:58 (1713535018) SKIP: replay-single test_81d needs >= 2 MDTs SKIP 81d (2s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81e: DNE: unlink remote dir, drop MDT1 req reply, fail MDT0 ========================================================== 09:57:02 (1713535022) SKIP: replay-single test_81e needs >= 2 MDTs SKIP 81e (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81f: DNE: unlink remote dir, drop MDT1 req reply, fail MDT1 ========================================================== 09:57:05 (1713535025) SKIP: replay-single test_81f needs >= 2 MDTs SKIP 81f (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81g: DNE: unlink remote dir, drop req reply, fail M0, then M1 ========================================================== 09:57:09 (1713535029) SKIP: replay-single test_81g needs >= 2 MDTs SKIP 81g (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81h: DNE: unlink remote dir, drop request reply, fail 2 MDTs ========================================================== 09:57:12 (1713535032) SKIP: replay-single test_81h needs >= 2 MDTs SKIP 81h (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 84a: stale open during export disconnect ========================================================== 09:57:16 (1713535036) fail_loc=0x80000144 total: 1 open/close in 0.01 seconds: 117.22 ops/second pdsh@oleg119-client: oleg119-client: ssh exited with exit code 5 PASS 84a (6s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 85a: check the cancellation of unused locks during recovery(IBITS) ========================================================== 09:57:23 (1713535043) before recovery: unused locks count = 201 Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 09:57:27 (1713535047) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 09:57:41 (1713535061) targets are mounted 09:57:41 (1713535061) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec after recovery: unused locks count = 101 PASS 85a (24s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 85b: check the cancellation of unused locks during recovery(EXTENT) ========================================================== 09:57:49 (1713535069) before recovery: unused locks count = 100 Failing ost1 on oleg119-server Stopping /mnt/lustre-ost1 (opts:) on oleg119-server 09:57:56 (1713535076) shut down Failover ost1 to oleg119-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-OST0000 09:58:10 (1713535090) targets are mounted 09:58:10 (1713535090) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec after recovery: unused locks count = 0 PASS 85b (27s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 86: umount server after clear nid_stats should not hit LBUG ========================================================== 09:58:18 (1713535098) Stopping clients: oleg119-client.virtnet /mnt/lustre (opts:) Stopping client oleg119-client.virtnet /mnt/lustre opts: mdt.lustre-MDT0000.exports.clear=0 Stopping /mnt/lustre-mds1 (opts:) on oleg119-server Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 Starting client oleg119-client.virtnet: -o user_xattr,flock oleg119-server@tcp:/lustre /mnt/lustre Started clients oleg119-client.virtnet: 192.168.201.119@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) PASS 86 (8s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 87a: write replay ================== 09:58:29 (1713535109) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3840 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre 8+0 records in 8+0 records out 8388608 bytes (8.4 MB) copied, 0.172243 s, 48.7 MB/s Failing ost1 on oleg119-server Stopping /mnt/lustre-ost1 (opts:) on oleg119-server 09:58:33 (1713535113) shut down Failover ost1 to oleg119-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-OST0000 09:58:47 (1713535127) targets are mounted 09:58:47 (1713535127) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec 8+0 records in 8+0 records out 8388608 bytes (8.4 MB) copied, 0.0605186 s, 139 MB/s PASS 87a (24s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 87b: write replay with changed data (checksum resend) ========================================================== 09:58:55 (1713535135) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3840 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 12288 3757056 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 16384 7522304 1% /mnt/lustre 8+0 records in 8+0 records out 8388608 bytes (8.4 MB) copied, 0.133272 s, 62.9 MB/s 8+0 records in 8+0 records out 8 bytes (8 B) copied, 0.0027165 s, 2.9 kB/s Failing ost1 on oleg119-server Stopping /mnt/lustre-ost1 (opts:) on oleg119-server 09:59:00 (1713535140) shut down Failover ost1 to oleg119-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-OST0000 09:59:14 (1713535154) targets are mounted 09:59:14 (1713535154) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec 0+1 records in 0+1 records out 72 bytes (72 B) copied, 0.00278722 s, 25.8 kB/s PASS 87b (24s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 88: MDS should not assign same objid to different files ========================================================== 09:59:20 (1713535160) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3840 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 13312 3756032 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 17408 7521280 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3840 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 13312 3756032 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 17408 7521280 1% /mnt/lustre before test: last_id = 11297, next_id = 11268 Creating to objid 11297 on ost lustre-OST0000... total: 31 open/close in 0.08 seconds: 391.71 ops/second total: 8 open/close in 0.03 seconds: 281.66 ops/second before recovery: last_id = 11393, next_id = 11306 Stopping /mnt/lustre-mds1 (opts:) on oleg119-server Stopping /mnt/lustre-ost1 (opts:) on oleg119-server Failover mds1 to oleg119-server Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 Failover ost1 to oleg119-server Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-OST0000 after recovery: last_id = 11337, next_id = 11306 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0245576 s, 21.3 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0274795 s, 19.1 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0222469 s, 23.6 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0238985 s, 21.9 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0266147 s, 19.7 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0219156 s, 23.9 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0253466 s, 20.7 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0307232 s, 17.1 MB/s -rw-r--r-- 1 root root 0 Apr 19 09:59 /mnt/lustre/d88.replay-single/f-11268 -rw-r--r-- 1 root root 0 Apr 19 09:59 /mnt/lustre/d88.replay-single/f-11269 -rw-r--r-- 1 root root 0 Apr 19 09:59 /mnt/lustre/d88.replay-single/f-11270 -rw-r--r-- 1 root root 0 Apr 19 09:59 /mnt/lustre/d88.replay-single/f-11271 -rw-r--r-- 1 root root 0 Apr 19 09:59 /mnt/lustre/d88.replay-single/f-11272 -rw-r--r-- 1 root root 0 Apr 19 09:59 /mnt/lustre/d88.replay-single/f-11273 -rw-r--r-- 1 root root 0 Apr 19 09:59 /mnt/lustre/d88.replay-single/f-11274 -rw-r--r-- 1 root root 0 Apr 19 09:59 /mnt/lustre/d88.replay-single/f-11275 -rw-r--r-- 1 root root 0 Apr 19 09:59 /mnt/lustre/d88.replay-single/f-11276 -rw-r--r-- 1 root root 0 Apr 19 09:59 /mnt/lustre/d88.replay-single/f-11277 -rw-r--r-- 1 root root 0 Apr 19 09:59 /mnt/lustre/d88.replay-single/f-11278 -rw-r--r-- 1 root root 0 Apr 19 09:59 /mnt/lustre/d88.replay-single/f-11279 -rw-r--r-- 1 root root 0 Apr 19 09:59 /mnt/lustre/d88.replay-single/f-11280 -rw-r--r-- 1 root root 0 Apr 19 09:59 /mnt/lustre/d88.replay-single/f-11281 -rw-r--r-- 1 root root 0 Apr 19 09:59 /mnt/lustre/d88.replay-single/f-11282 -rw-r--r-- 1 root root 0 Apr 19 09:59 /mnt/lustre/d88.replay-single/f-11283 -rw-r--r-- 1 root root 0 Apr 19 09:59 /mnt/lustre/d88.replay-single/f-11284 -rw-r--r-- 1 root root 0 Apr 19 09:59 /mnt/lustre/d88.replay-single/f-11285 -rw-r--r-- 1 root root 0 Apr 19 09:59 /mnt/lustre/d88.replay-single/f-11286 -rw-r--r-- 1 root root 0 Apr 19 09:59 /mnt/lustre/d88.replay-single/f-11287 -rw-r--r-- 1 root root 0 Apr 19 09:59 /mnt/lustre/d88.replay-single/f-11288 -rw-r--r-- 1 root root 0 Apr 19 09:59 /mnt/lustre/d88.replay-single/f-11289 -rw-r--r-- 1 root root 0 Apr 19 09:59 /mnt/lustre/d88.replay-single/f-11290 -rw-r--r-- 1 root root 0 Apr 19 09:59 /mnt/lustre/d88.replay-single/f-11291 -rw-r--r-- 1 root root 0 Apr 19 09:59 /mnt/lustre/d88.replay-single/f-11292 -rw-r--r-- 1 root root 0 Apr 19 09:59 /mnt/lustre/d88.replay-single/f-11293 -rw-r--r-- 1 root root 0 Apr 19 09:59 /mnt/lustre/d88.replay-single/f-11294 -rw-r--r-- 1 root root 0 Apr 19 09:59 /mnt/lustre/d88.replay-single/f-11295 -rw-r--r-- 1 root root 0 Apr 19 09:59 /mnt/lustre/d88.replay-single/f-11296 -rw-r--r-- 1 root root 0 Apr 19 09:59 /mnt/lustre/d88.replay-single/f-11297 -rw-r--r-- 1 root root 0 Apr 19 09:59 /mnt/lustre/d88.replay-single/f-11298 -rw-r--r-- 1 root root 0 Apr 19 09:59 /mnt/lustre/d88.replay-single/f-11299 -rw-r--r-- 1 root root 0 Apr 19 09:59 /mnt/lustre/d88.replay-single/f-11300 -rw-r--r-- 1 root root 0 Apr 19 09:59 /mnt/lustre/d88.replay-single/f-11301 -rw-r--r-- 1 root root 0 Apr 19 09:59 /mnt/lustre/d88.replay-single/f-11302 -rw-r--r-- 1 root root 0 Apr 19 09:59 /mnt/lustre/d88.replay-single/f-11303 -rw-r--r-- 1 root root 0 Apr 19 09:59 /mnt/lustre/d88.replay-single/f-11304 -rw-r--r-- 1 root root 0 Apr 19 09:59 /mnt/lustre/d88.replay-single/f-11305 -rw-r--r-- 1 root root 524288 Apr 19 10:00 /mnt/lustre/d88.replay-single/f-11309 -rw-r--r-- 1 root root 524288 Apr 19 10:00 /mnt/lustre/d88.replay-single/f-11310 -rw-r--r-- 1 root root 524288 Apr 19 10:00 /mnt/lustre/d88.replay-single/f-11311 -rw-r--r-- 1 root root 524288 Apr 19 10:00 /mnt/lustre/d88.replay-single/f-11312 -rw-r--r-- 1 root root 524288 Apr 19 10:00 /mnt/lustre/d88.replay-single/f-11313 -rw-r--r-- 1 root root 524288 Apr 19 10:00 /mnt/lustre/d88.replay-single/f-11314 -rw-r--r-- 1 root root 524288 Apr 19 10:00 /mnt/lustre/d88.replay-single/f-11315 -rw-r--r-- 1 root root 524288 Apr 19 10:00 /mnt/lustre/d88.replay-single/f-11316 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.024524 s, 21.4 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0252546 s, 20.8 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0225846 s, 23.2 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.024677 s, 21.2 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0242144 s, 21.7 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0258535 s, 20.3 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0236434 s, 22.2 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.026932 s, 19.5 MB/s PASS 88 (49s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 89: no disk space leak on late ost connection ========================================================== 10:00:10 (1713535210) Waiting for orphan cleanup... osp.lustre-OST0000-osc-MDT0000.old_sync_processed osp.lustre-OST0001-osc-MDT0000.old_sync_processed wait 40 secs maximumly for oleg119-server mds-ost sync done. sleep 5 for ZFS zfs Waiting for MDT destroys to complete 10+0 records in 10+0 records out 41943040 bytes (42 MB) copied, 0.196492 s, 213 MB/s Stopping /mnt/lustre-ost1 (opts:) on oleg119-server Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 10:00:21 (1713535221) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 10:00:34 (1713535234) targets are mounted 10:00:34 (1713535234) facet_failover done Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-OST0000 Starting client: oleg119-client.virtnet: -o user_xattr,flock oleg119-server@tcp:/lustre /mnt/lustre osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 68 sec Waiting for orphan cleanup... osp.lustre-OST0000-osc-MDT0000.old_sync_processed osp.lustre-OST0001-osc-MDT0000.old_sync_processed wait 40 secs maximumly for oleg119-server mds-ost sync done. sleep 5 for ZFS zfs Waiting for MDT destroys to complete free_before: 7517184 free_after: 7517184 PASS 89 (108s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 90: lfs find identifies the missing striped file segments ========================================================== 10:01:59 (1713535319) Create the files Fail ost2 lustre-OST0001_UUID, display the list of affected files Stopping /mnt/lustre-ost2 (opts:) on oleg119-server General Query: lfs find /mnt/lustre/d90.replay-single /mnt/lustre/d90.replay-single /mnt/lustre/d90.replay-single/f1 /mnt/lustre/d90.replay-single/f0 /mnt/lustre/d90.replay-single/all Querying files on shutdown ost2: lfs find --obd lustre-OST0001_UUID /mnt/lustre/d90.replay-single/f1 /mnt/lustre/d90.replay-single/all Check getstripe: /home/green/git/lustre-release/lustre/utils/lfs getstripe -r --obd lustre-OST0001_UUID /mnt/lustre/d90.replay-single/f1 lmm_stripe_count: 1 lmm_stripe_size: 4194304 lmm_pattern: raid0 lmm_layout_gen: 0 lmm_stripe_offset: 1 obdidx objid objid group 1 11139 0x2b83 0x280000401 * /mnt/lustre/d90.replay-single/all lmm_stripe_count: 2 lmm_stripe_size: 4194304 lmm_pattern: raid0 lmm_layout_gen: 0 lmm_stripe_offset: 1 obdidx objid objid group 1 11138 0x2b82 0x280000401 * /mnt/lustre/d90.replay-single/all /mnt/lustre/d90.replay-single/f1 Failover ost2 to oleg119-server Starting ost2: -o localrecov lustre-ost2/ost2 /mnt/lustre-ost2 seq.cli-lustre-OST0001-super.width=65536 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-OST0001 PASS 90 (18s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 93a: replay + reconnect ============ 10:02:18 (1713535338) 1+0 records in 1+0 records out 1024 bytes (1.0 kB) copied, 0.00146247 s, 700 kB/s fail_val=40 fail_loc=0x715 Failing ost1 on oleg119-server Stopping /mnt/lustre-ost1 (opts:) on oleg119-server 10:02:20 (1713535340) shut down Failover ost1 to oleg119-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-OST0000 10:02:33 (1713535353) targets are mounted 10:02:33 (1713535353) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec PASS 93a (58s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 93b: replay + reconnect on mds ===== 10:03:18 (1713535398) total: 20 open/close in 0.06 seconds: 309.88 ops/second fail_val=80 fail_loc=0x715 Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 10:03:20 (1713535400) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 10:03:32 (1713535412) targets are mounted 10:03:32 (1713535412) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 93b (100s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 100a: DNE: create striped dir, drop update rep from MDT1, fail MDT1 ========================================================== 10:05:00 (1713535500) SKIP: replay-single test_100a needs >= 2 MDTs SKIP 100a (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 100b: DNE: create striped dir, fail MDT0 ========================================================== 10:05:03 (1713535503) SKIP: replay-single test_100b needs >= 2 MDTs SKIP 100b (2s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 100c: DNE: create striped dir, abort_recov_mdt mds2 ========================================================== 10:05:07 (1713535507) SKIP: replay-single test_100c needs >= 2 MDTs SKIP 100c (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 100d: DNE: cancel update logs upon recovery abort ========================================================== 10:05:10 (1713535510) SKIP: replay-single test_100d needs > 1 MDTs SKIP 100d (2s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 100e: DNE: create striped dir on MDT0 and MDT1, fail MDT0, MDT1 ========================================================== 10:05:14 (1713535514) SKIP: replay-single test_100e needs >= 2 MDTs SKIP 100e (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 101: Shouldn't reassign precreated objs to other files after recovery ========================================================== 10:05:17 (1713535517) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3968 2204544 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 21504 3747840 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 25600 7513088 1% /mnt/lustre Stopping /mnt/lustre-mds1 (opts:) on oleg119-server Failover mds1 to oleg119-server Starting mds1: -o localrecov -o abort_recovery lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 PASS 101 (39s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 102a: check resend (request lost) with multiple modify RPCs in flight ========================================================== 10:05:58 (1713535558) creating 7 files ... fail_loc=0x159 launch 7 chmod in parallel (10:05:59) ... fail_loc=0 done (10:06:15) /mnt/lustre/d102a.replay-single/file-1 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-2 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-3 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-4 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-5 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-6 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-7 has perms 0600 OK PASS 102a (18s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 102b: check resend (reply lost) with multiple modify RPCs in flight ========================================================== 10:06:19 (1713535579) creating 7 files ... fail_loc=0x15a launch 7 chmod in parallel (10:06:19) ... fail_loc=0 done (10:06:35) /mnt/lustre/d102b.replay-single/file-1 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-2 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-3 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-4 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-5 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-6 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-7 has perms 0600 OK PASS 102b (18s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 102c: check replay w/o reconstruction with multiple mod RPCs in flight ========================================================== 10:06:38 (1713535598) creating 7 files ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3968 2204544 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 21504 3333120 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3350528 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 25600 6683648 1% /mnt/lustre fail_loc=0x15a launch 7 chmod in parallel (10:06:40) ... fail_loc=0 Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 10:06:43 (1713535603) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 10:06:56 (1713535616) targets are mounted 10:06:56 (1713535616) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec done (10:07:00) /mnt/lustre/d102c.replay-single/file-1 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-2 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-3 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-4 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-5 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-6 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-7 has perms 0600 OK PASS 102c (24s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 102d: check replay & reconstruction with multiple mod RPCs in flight ========================================================== 10:07:04 (1713535624) creating 7 files ... fail_loc=0x15a launch 7 chmod in parallel (10:07:05) ... fail_loc=0 Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 10:07:08 (1713535628) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 10:07:20 (1713535640) targets are mounted 10:07:20 (1713535640) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec done (10:07:24) /mnt/lustre/d102d.replay-single/file-1 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-2 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-3 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-4 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-5 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-6 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-7 has perms 0600 OK PASS 102d (22s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 103: Check otr_next_id overflow ==== 10:07:28 (1713535648) fail_loc=0x80000162 total: 30 open/close in 0.14 seconds: 207.13 ops/second Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 10:07:31 (1713535651) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 10:07:45 (1713535665) targets are mounted 10:07:45 (1713535665) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 103 (22s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 110a: DNE: create striped dir, fail MDT1 ========================================================== 10:07:53 (1713535673) SKIP: replay-single test_110a needs >= 2 MDTs SKIP 110a (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 110b: DNE: create striped dir, fail MDT1 and client ========================================================== 10:07:56 (1713535676) SKIP: replay-single test_110b needs >= 2 MDTs SKIP 110b (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 110c: DNE: create striped dir, fail MDT2 ========================================================== 10:08:00 (1713535680) SKIP: replay-single test_110c needs >= 2 MDTs SKIP 110c (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 110d: DNE: create striped dir, fail MDT2 and client ========================================================== 10:08:03 (1713535683) SKIP: replay-single test_110d needs >= 2 MDTs SKIP 110d (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 110e: DNE: create striped dir, uncommit on MDT2, fail client/MDT1/MDT2 ========================================================== 10:08:07 (1713535687) SKIP: replay-single test_110e needs >= 2 MDTs SKIP 110e (1s) debug_raw_pointers=0 debug_raw_pointers=0 SKIP: replay-single test_110f skipping excluded test 110f debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 110g: DNE: create striped dir, uncommit on MDT1, fail client/MDT1/MDT2 ========================================================== 10:08:11 (1713535691) SKIP: replay-single test_110g needs >= 2 MDTs SKIP 110g (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 111a: DNE: unlink striped dir, fail MDT1 ========================================================== 10:08:14 (1713535694) SKIP: replay-single test_111a needs >= 2 MDTs SKIP 111a (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 111b: DNE: unlink striped dir, fail MDT2 ========================================================== 10:08:17 (1713535697) SKIP: replay-single test_111b needs >= 2 MDTs SKIP 111b (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 111c: DNE: unlink striped dir, uncommit on MDT1, fail client/MDT1/MDT2 ========================================================== 10:08:21 (1713535701) SKIP: replay-single test_111c needs >= 2 MDTs SKIP 111c (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 111d: DNE: unlink striped dir, uncommit on MDT2, fail client/MDT1/MDT2 ========================================================== 10:08:24 (1713535704) SKIP: replay-single test_111d needs >= 2 MDTs SKIP 111d (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 111e: DNE: unlink striped dir, uncommit on MDT2, fail MDT1/MDT2 ========================================================== 10:08:27 (1713535707) SKIP: replay-single test_111e needs >= 2 MDTs SKIP 111e (0s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 111f: DNE: unlink striped dir, uncommit on MDT1, fail MDT1/MDT2 ========================================================== 10:08:29 (1713535709) SKIP: replay-single test_111f needs >= 2 MDTs SKIP 111f (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 111g: DNE: unlink striped dir, fail MDT1/MDT2 ========================================================== 10:08:32 (1713535712) SKIP: replay-single test_111g needs >= 2 MDTs SKIP 111g (0s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112a: DNE: cross MDT rename, fail MDT1 ========================================================== 10:08:34 (1713535714) SKIP: replay-single test_112a needs >= 4 MDTs SKIP 112a (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112b: DNE: cross MDT rename, fail MDT2 ========================================================== 10:08:37 (1713535717) SKIP: replay-single test_112b needs >= 4 MDTs SKIP 112b (0s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112c: DNE: cross MDT rename, fail MDT3 ========================================================== 10:08:39 (1713535719) SKIP: replay-single test_112c needs >= 4 MDTs SKIP 112c (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112d: DNE: cross MDT rename, fail MDT4 ========================================================== 10:08:42 (1713535722) SKIP: replay-single test_112d needs >= 4 MDTs SKIP 112d (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112e: DNE: cross MDT rename, fail MDT1 and MDT2 ========================================================== 10:08:45 (1713535725) SKIP: replay-single test_112e needs >= 4 MDTs SKIP 112e (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112f: DNE: cross MDT rename, fail MDT1 and MDT3 ========================================================== 10:08:49 (1713535729) SKIP: replay-single test_112f needs >= 4 MDTs SKIP 112f (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112g: DNE: cross MDT rename, fail MDT1 and MDT4 ========================================================== 10:08:52 (1713535732) SKIP: replay-single test_112g needs >= 4 MDTs SKIP 112g (2s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112h: DNE: cross MDT rename, fail MDT2 and MDT3 ========================================================== 10:08:56 (1713535736) SKIP: replay-single test_112h needs >= 4 MDTs SKIP 112h (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112i: DNE: cross MDT rename, fail MDT2 and MDT4 ========================================================== 10:08:59 (1713535739) SKIP: replay-single test_112i needs >= 4 MDTs SKIP 112i (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112j: DNE: cross MDT rename, fail MDT3 and MDT4 ========================================================== 10:09:02 (1713535742) SKIP: replay-single test_112j needs >= 4 MDTs SKIP 112j (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112k: DNE: cross MDT rename, fail MDT1,MDT2,MDT3 ========================================================== 10:09:05 (1713535745) SKIP: replay-single test_112k needs >= 4 MDTs SKIP 112k (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112l: DNE: cross MDT rename, fail MDT1,MDT2,MDT4 ========================================================== 10:09:08 (1713535748) SKIP: replay-single test_112l needs >= 4 MDTs SKIP 112l (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112m: DNE: cross MDT rename, fail MDT1,MDT3,MDT4 ========================================================== 10:09:11 (1713535751) SKIP: replay-single test_112m needs >= 4 MDTs SKIP 112m (0s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112n: DNE: cross MDT rename, fail MDT2,MDT3,MDT4 ========================================================== 10:09:14 (1713535754) SKIP: replay-single test_112n needs >= 4 MDTs SKIP 112n (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 115: failover for create/unlink striped directory ========================================================== 10:09:17 (1713535757) SKIP: replay-single test_115 needs >= 2 MDTs SKIP 115 (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 116a: large update log master MDT recovery ========================================================== 10:09:20 (1713535760) SKIP: replay-single test_116a needs >= 2 MDTs SKIP 116a (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 116b: large update log slave MDT recovery ========================================================== 10:09:23 (1713535763) SKIP: replay-single test_116b needs >= 2 MDTs SKIP 116b (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 117: DNE: cross MDT unlink, fail MDT1 and MDT2 ========================================================== 10:09:25 (1713535765) SKIP: replay-single test_117 needs >= 4 MDTs SKIP 117 (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 118: invalidate osp update will not cause update log corruption ========================================================== 10:09:28 (1713535768) SKIP: replay-single test_118 needs >= 2 MDTs SKIP 118 (0s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 119: timeout of normal replay does not cause DNE replay fails ========================================================== 10:09:30 (1713535770) SKIP: replay-single test_119 needs >= 2 MDTs SKIP 119 (0s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 120: DNE fail abort should stop both normal and DNE replay ========================================================== 10:09:32 (1713535772) SKIP: replay-single test_120 needs >= 2 MDTs SKIP 120 (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 121: lock replay timed out and race ========================================================== 10:09:34 (1713535774) multiop /mnt/lustre/f121.replay-single vs_s TMPPIPE=/tmp/multiop_open_wait_pipe.6753 Stopping /mnt/lustre-mds1 (opts:) on oleg119-server Failover mds1 to oleg119-server fail_loc=0x721 fail_val=0 at_max=0 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 fail_loc=0x0 at_max=600 PASS 121 (25s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 130a: DoM file create (setstripe) replay ========================================================== 10:10:01 (1713535801) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3968 2204544 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 21504 3333120 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 25600 7098368 1% /mnt/lustre Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 10:10:05 (1713535805) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 10:10:19 (1713535819) targets are mounted 10:10:19 (1713535819) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 130a (24s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 130b: DoM file create (inherited) replay ========================================================== 10:10:27 (1713535827) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3968 2204544 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 21504 3333120 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 25600 7098368 1% /mnt/lustre Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 10:10:31 (1713535831) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 10:10:45 (1713535845) targets are mounted 10:10:45 (1713535845) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 130b (24s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 131a: DoM file write lock replay === 10:10:53 (1713535853) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3968 2204544 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 21504 3333120 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 25600 7098368 1% /mnt/lustre 1+0 records in 1+0 records out 8 bytes (8 B) copied, 0.0028735 s, 2.8 kB/s Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 10:10:57 (1713535857) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 10:11:11 (1713535871) targets are mounted 10:11:11 (1713535871) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 131a (24s) debug_raw_pointers=0 debug_raw_pointers=0 SKIP: replay-single test_131b skipping excluded test 131b debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 132a: PFL new component instantiate replay ========================================================== 10:11:20 (1713535880) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3968 2204544 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 21504 3333120 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 25600 7098368 1% /mnt/lustre 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0280718 s, 37.4 MB/s /mnt/lustre/f132a.replay-single lcm_layout_gen: 3 lcm_mirror_count: 1 lcm_entry_count: 2 lcme_id: 1 lcme_mirror_id: 0 lcme_flags: init lcme_extent.e_start: 0 lcme_extent.e_end: 1048576 lmm_stripe_count: 1 lmm_stripe_size: 1048576 lmm_pattern: raid0 lmm_layout_gen: 0 lmm_stripe_offset: 0 lmm_objects: - 0: { l_ost_idx: 0, l_fid: [0x240000401:0x316a:0x0] } lcme_id: 2 lcme_mirror_id: 0 lcme_flags: init lcme_extent.e_start: 1048576 lcme_extent.e_end: EOF lmm_stripe_count: 2 lmm_stripe_size: 4194304 lmm_pattern: raid0 lmm_layout_gen: 0 lmm_stripe_offset: 1 lmm_objects: - 0: { l_ost_idx: 1, l_fid: [0x280000401:0x30a2:0x0] } - 1: { l_ost_idx: 0, l_fid: [0x240000401:0x316b:0x0] } Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 10:11:24 (1713535884) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 10:11:37 (1713535897) targets are mounted 10:11:37 (1713535897) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f132a.replay-single lcm_layout_gen: 1 lcm_mirror_count: 1 lcm_entry_count: 2 lcme_id: 1 lcme_mirror_id: 0 lcme_flags: init lcme_extent.e_start: 0 lcme_extent.e_end: 1048576 lmm_stripe_count: 1 lmm_stripe_size: 1048576 lmm_pattern: raid0 lmm_layout_gen: 65535 lmm_stripe_offset: 0 lmm_objects: - 0: { l_ost_idx: 0, l_fid: [0x240000401:0x316a:0x0] } lcme_id: 2 lcme_mirror_id: 0 lcme_flags: init lcme_extent.e_start: 1048576 lcme_extent.e_end: EOF lmm_stripe_count: 2 lmm_stripe_size: 4194304 lmm_pattern: raid0 lmm_layout_gen: 65535 lmm_stripe_offset: 1 lmm_objects: - 0: { l_ost_idx: 1, l_fid: [0x280000401:0x30a2:0x0] } - 1: { l_ost_idx: 0, l_fid: [0x240000401:0x316b:0x0] } PASS 132a (22s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 133: check resend of ongoing requests for lwp during failover ========================================================== 10:11:44 (1713535904) SKIP: replay-single test_133 needs >= 2 MDTs SKIP 133 (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 134: replay creation of a file created in a pool ========================================================== 10:11:47 (1713535907) Creating new pool oleg119-server: Pool lustre.pool_134 created Adding targets to pool oleg119-server: OST lustre-OST0001_UUID added to pool lustre.pool_134 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3968 2204544 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 21504 3747840 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 5120 3764224 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 26624 7512064 1% /mnt/lustre Failing mds1 on oleg119-server Stopping /mnt/lustre-mds1 (opts:) on oleg119-server 10:11:54 (1713535914) shut down Failover mds1 to oleg119-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-MDT0000 10:12:08 (1713535928) targets are mounted 10:12:08 (1713535928) facet_failover done oleg119-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Destroy the created pools: pool_134 lustre.pool_134 oleg119-server: OST lustre-OST0001_UUID removed from pool lustre.pool_134 oleg119-server: Pool lustre.pool_134 destroyed Waiting 90s for 'foo' PASS 134 (33s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 135: Server failure in lock replay phase ========================================================== 10:12:23 (1713535943) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3968 2204544 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 21504 3747840 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 5120 3764224 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 26624 7512064 1% /mnt/lustre ldlm.cancel_unused_locks_before_replay=0 Stopping /mnt/lustre-ost1 (opts:) on oleg119-server Failover ost1 to oleg119-server oleg119-server: oleg119-server.virtnet: executing load_module ../libcfs/libcfs/libcfs fail_loc=0x32d fail_val=20 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg119-server: oleg119-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg119-client: oleg119-server: ssh exited with exit code 1 Started lustre-OST0000 End of sync oleg119-client.virtnet: executing wait_import_state_mount REPLAY_LOCKS osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in REPLAY_LOCKS state after 0 sec /home/green/git/lustre-release/lustre/tests/replay-single.sh: line 5024: kill: (6260) - No such process replay-single test_135: @@@@@@ FAIL: Unexpected sync success Trace dump: = /home/green/git/lustre-release/lustre/tests/test-framework.sh:6975:error() = /home/green/git/lustre-release/lustre/tests/replay-single.sh:5024:test_135() = /home/green/git/lustre-release/lustre/tests/test-framework.sh:7315:run_one() = /home/green/git/lustre-release/lustre/tests/test-framework.sh:7375:run_one_logged() = /home/green/git/lustre-release/lustre/tests/test-framework.sh:7201:run_test() = /home/green/git/lustre-release/lustre/tests/replay-single.sh:5044:main() Dumping lctl log to /tmp/testlogs//replay-single.test_135.*.1713535956.log ldlm.cancel_unused_locks_before_replay=1 rsync: chown "/tmp/testlogs/.replay-single.test_135.debug_log.oleg119-server.1713535956.log.3735km" failed: Operation not permitted (1) rsync: chown "/tmp/testlogs/.replay-single.test_135.dmesg.oleg119-server.1713535956.log.joldjV" failed: Operation not permitted (1) rsync error: some files/attrs were not transferred (see previous errors) (code 23) at main.c(1651) [generator=3.1.2] FAIL 135 (15s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 136: MDS to disconnect all OSPs first, then cleanup ldlm ========================================================== 10:12:42 (1713535962) SKIP: replay-single test_136 needs > 2 MDTs SKIP 136 (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 200: Dropping one OBD_PING should not cause disconnect ========================================================== 10:12:45 (1713535965) SKIP: replay-single test_200 Need remote client SKIP 200 (1s) debug_raw_pointers=0 debug_raw_pointers=0 == replay-single test complete, duration 4987 sec ======== 10:12:47 (1713535967) replay-single: FAIL: test_135 Unexpected sync success === replay-single: start cleanup 10:12:48 (1713535968) === === replay-single: finish cleanup 10:12:50 (1713535970) ===