-----============= acceptance-small: replay-single ============----- Tue Apr 16 15:40:34 EDT 2024 excepting tests: 110f 131b 59 36 === replay-single: start setup 15:40:37 (1713296437) === oleg130-client.virtnet: executing check_config_client /mnt/lustre oleg130-client.virtnet: Checking config lustre mounted on /mnt/lustre Checking servers environments Checking clients oleg130-client.virtnet environments Using TIMEOUT=20 osc.lustre-OST0000-osc-ffff8800b6cdc000.idle_timeout=debug osc.lustre-OST0001-osc-ffff8800b6cdc000.idle_timeout=debug disable quota as required oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all === replay-single: finish setup 15:40:43 (1713296443) === debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 0a: empty replay =================== 15:40:45 (1713296445) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:40:47 (1713296447) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:41:00 (1713296460) targets are mounted 15:41:00 (1713296460) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 0a (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 0b: ensure object created after recover exists. (3284) ========================================================== 15:41:06 (1713296466) Failing ost1 on oleg130-server Stopping /mnt/lustre-ost1 (opts:) on oleg130-server 15:41:07 (1713296467) shut down Failover ost1 to oleg130-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-OST0000 15:41:20 (1713296480) targets are mounted 15:41:20 (1713296480) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec total: 20 open/close in 0.07 seconds: 298.10 ops/second - unlinked 0 (time 1713296483 ; total 0 ; last 0) total: 20 unlinks in 0 seconds: inf unlinks/second PASS 0b (18s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 0c: check replay-barrier =========== 15:41:26 (1713296486) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:41:28 (1713296488) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:41:41 (1713296501) targets are mounted 15:41:41 (1713296501) facet_failover done Starting client: oleg130-client.virtnet: -o user_xattr,flock oleg130-server@tcp:/lustre /mnt/lustre rm: cannot remove '/mnt/lustre/f0c.replay-single': No such file or directory PASS 0c (77s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 0d: expired recovery with no clients ========================================================== 15:42:45 (1713296565) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:42:48 (1713296568) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:43:01 (1713296581) targets are mounted 15:43:01 (1713296581) facet_failover done Starting client: oleg130-client.virtnet: -o user_xattr,flock oleg130-server@tcp:/lustre /mnt/lustre PASS 0d (78s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 1: simple create =================== 15:44:05 (1713296645) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:44:07 (1713296647) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:44:20 (1713296660) targets are mounted 15:44:20 (1713296660) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f1.replay-single has type file OK PASS 1 (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 2a: touch ========================== 15:44:26 (1713296666) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:44:29 (1713296669) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:44:42 (1713296682) targets are mounted 15:44:42 (1713296682) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f2a.replay-single has type file OK PASS 2a (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 2b: touch ========================== 15:44:48 (1713296688) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:44:50 (1713296690) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:45:03 (1713296703) targets are mounted 15:45:03 (1713296703) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f2b.replay-single has type file OK PASS 2b (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 2c: setstripe replay =============== 15:45:09 (1713296709) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:45:12 (1713296712) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:45:25 (1713296725) targets are mounted 15:45:25 (1713296725) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f2c.replay-single has type file OK PASS 2c (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 2d: setdirstripe replay ============ 15:45:31 (1713296731) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:45:33 (1713296733) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:45:46 (1713296746) targets are mounted 15:45:46 (1713296746) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d2d.replay-single has type dir OK PASS 2d (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 2e: O_CREAT|O_EXCL create replay === 15:45:52 (1713296752) fail_loc=0x8000013b UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:45:56 (1713296756) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:46:08 (1713296768) targets are mounted 15:46:08 (1713296768) facet_failover done Succeed in opening file "/mnt/lustre/f2e.replay-single"(flags=O_CREAT) oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f2e.replay-single has type file OK PASS 2e (21s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 3a: replay failed open(O_DIRECTORY) ========================================================== 15:46:14 (1713296774) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Error in opening file "/mnt/lustre/f3a.replay-single"(flags=O_DIRECTORY) 20: Not a directory Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:46:17 (1713296777) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:46:29 (1713296789) targets are mounted 15:46:29 (1713296789) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f3a.replay-single has type file OK PASS 3a (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 3b: replay failed open -ENOMEM ===== 15:46:36 (1713296796) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre fail_loc=0x80000114 touch: cannot touch '/mnt/lustre/f3b.replay-single': Cannot allocate memory fail_loc=0 Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:46:39 (1713296799) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:46:51 (1713296811) targets are mounted 15:46:51 (1713296811) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/f3b.replay-single: No such file or directory PASS 3b (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 3c: replay failed open -ENOMEM ===== 15:46:57 (1713296817) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre fail_loc=0x80000128 touch: cannot touch '/mnt/lustre/f3c.replay-single': Cannot allocate memory fail_loc=0 Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:47:00 (1713296820) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:47:12 (1713296832) targets are mounted 15:47:12 (1713296832) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/f3c.replay-single: No such file or directory PASS 3c (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 4a: |x| 10 open(O_CREAT)s ========== 15:47:18 (1713296838) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:47:21 (1713296841) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:47:33 (1713296853) targets are mounted 15:47:33 (1713296853) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 4a (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 4b: |x| rm 10 files ================ 15:47:39 (1713296859) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3761152 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7527424 1% /mnt/lustre Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:47:42 (1713296862) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:47:54 (1713296874) targets are mounted 15:47:54 (1713296874) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/f4b.replay-single-*: No such file or directory PASS 4b (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 5: |x| 220 open(O_CREAT) =========== 15:48:00 (1713296880) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:48:04 (1713296884) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:48:16 (1713296896) targets are mounted 15:48:16 (1713296896) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 5 (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 6a: mkdir + contained create ======= 15:48:28 (1713296908) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:48:30 (1713296910) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:48:43 (1713296923) targets are mounted 15:48:43 (1713296923) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d6a.replay-single has type dir OK /mnt/lustre/d6a.replay-single/f6a.replay-single has type file OK PASS 6a (21s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 6b: |X| rmdir ====================== 15:48:50 (1713296930) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:48:53 (1713296933) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:49:05 (1713296945) targets are mounted 15:49:05 (1713296945) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/d6b.replay-single: No such file or directory PASS 6b (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 7: mkdir |X| contained create ====== 15:49:11 (1713296951) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:49:13 (1713296953) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:49:26 (1713296966) targets are mounted 15:49:26 (1713296966) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d7.replay-single has type dir OK /mnt/lustre/d7.replay-single/f7.replay-single has type file OK PASS 7 (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 8: creat open |X| close ============ 15:49:32 (1713296972) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre multiop /mnt/lustre/f8.replay-single vmo_c TMPPIPE=/tmp/multiop_open_wait_pipe.6919 Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:49:34 (1713296974) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:49:47 (1713296987) targets are mounted 15:49:47 (1713296987) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f8.replay-single /mnt/lustre/f8.replay-single has type file OK PASS 8 (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 9: |X| create (same inum/gen) ====== 15:49:53 (1713296993) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:49:55 (1713296995) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:50:07 (1713297007) targets are mounted 15:50:07 (1713297007) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec old_inum == 144115305935798546, new_inum == 144115305935798546 old_inum and new_inum match PASS 9 (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 10: create |X| rename unlink ======= 15:50:13 (1713297013) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:50:16 (1713297016) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:50:28 (1713297028) targets are mounted 15:50:28 (1713297028) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/f10.replay-single: No such file or directory PASS 10 (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 11: create open write rename |X| create-old-name read ========================================================== 15:50:34 (1713297034) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre new old Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:50:37 (1713297037) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:50:49 (1713297049) targets are mounted 15:50:49 (1713297049) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec new old PASS 11 (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 12: open, unlink |X| close ========= 15:50:55 (1713297055) multiop /mnt/lustre/f12.replay-single vo_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6919 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:50:58 (1713297058) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:51:10 (1713297070) targets are mounted 15:51:10 (1713297070) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 12 (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 13: open chmod 0 |x| write close === 15:51:16 (1713297076) multiop /mnt/lustre/f13.replay-single vO_wc TMPPIPE=/tmp/multiop_open_wait_pipe.6919 /mnt/lustre/f13.replay-single has perms 00 OK UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:51:19 (1713297079) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:51:31 (1713297091) targets are mounted 15:51:31 (1713297091) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f13.replay-single has perms 00 OK /mnt/lustre/f13.replay-single has size 1 OK PASS 13 (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 14: open(O_CREAT), unlink |X| close ========================================================== 15:51:37 (1713297097) multiop /mnt/lustre/f14.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6919 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:51:40 (1713297100) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:51:53 (1713297113) targets are mounted 15:51:53 (1713297113) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 14 (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 15: open(O_CREAT), unlink |X| touch new, close ========================================================== 15:51:59 (1713297119) multiop /mnt/lustre/f15.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6919 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:52:01 (1713297121) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:52:13 (1713297133) targets are mounted 15:52:13 (1713297133) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 15 (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 16: |X| open(O_CREAT), unlink, touch new, unlink new ========================================================== 15:52:19 (1713297139) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:52:22 (1713297142) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:52:34 (1713297154) targets are mounted 15:52:34 (1713297154) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 16 (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 17: |X| open(O_CREAT), |replay| close ========================================================== 15:52:40 (1713297160) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre multiop /mnt/lustre/f17.replay-single vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.6919 Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:52:42 (1713297162) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:52:55 (1713297175) targets are mounted 15:52:55 (1713297175) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f17.replay-single has type file OK PASS 17 (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 18: open(O_CREAT), unlink, touch new, close, touch, unlink ========================================================== 15:53:01 (1713297181) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre multiop /mnt/lustre/f18.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6919 pid: 21675 will close Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:53:03 (1713297183) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:53:16 (1713297196) targets are mounted 15:53:16 (1713297196) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 18 (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 19: mcreate, open, write, rename === 15:53:22 (1713297202) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre old Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:53:24 (1713297204) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:53:37 (1713297217) targets are mounted 15:53:37 (1713297217) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec old PASS 19 (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 20a: |X| open(O_CREAT), unlink, replay, close (test mds_cleanup_orphans) ========================================================== 15:53:42 (1713297222) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre multiop /mnt/lustre/f20a.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6919 Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:53:45 (1713297225) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:53:57 (1713297237) targets are mounted 15:53:57 (1713297237) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 20a (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 20b: write, unlink, eviction, replay (test mds_cleanup_orphans) ========================================================== 15:54:03 (1713297243) /mnt/lustre/f20b.replay-single lmm_stripe_count: 1 lmm_stripe_size: 4194304 lmm_pattern: raid0 lmm_layout_gen: 0 lmm_stripe_offset: 0 obdidx objid objid group 0 1090 0x442 0x240000400 10000+0 records in 10000+0 records out 40960000 bytes (41 MB) copied, 1.40713 s, 29.1 MB/s Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:54:07 (1713297247) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:54:20 (1713297260) targets are mounted 15:54:20 (1713297260) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec affected facets: mds1 oleg130-server: oleg130-server.virtnet: executing _wait_recovery_complete *.lustre-MDT0000.recovery_status 1475 oleg130-server: *.lustre-MDT0000.recovery_status status: COMPLETE sleep 5 for ZFS zfs Waiting for MDT destroys to complete before 6144, after 6144 PASS 20b (29s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 20c: check that client eviction does not affect file content ========================================================== 15:54:33 (1713297273) multiop /mnt/lustre/f20c.replay-single vOw_c TMPPIPE=/tmp/multiop_open_wait_pipe.6919 -rw-r--r-- 1 root root 1 Apr 16 15:54 /mnt/lustre/f20c.replay-single pdsh@oleg130-client: oleg130-client: ssh exited with exit code 5 PASS 20c (4s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 21: |X| open(O_CREAT), unlink touch new, replay, close (test mds_cleanup_orphans) ========================================================== 15:54:39 (1713297279) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3764224 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7530496 1% /mnt/lustre multiop /mnt/lustre/f21.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6919 Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:54:42 (1713297282) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:54:54 (1713297294) targets are mounted 15:54:54 (1713297294) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 21 (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 22: open(O_CREAT), |X| unlink, replay, close (test mds_cleanup_orphans) ========================================================== 15:55:01 (1713297301) multiop /mnt/lustre/f22.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6919 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3764224 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7530496 1% /mnt/lustre Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:55:04 (1713297304) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:55:17 (1713297317) targets are mounted 15:55:17 (1713297317) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 22 (21s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 23: open(O_CREAT), |X| unlink touch new, replay, close (test mds_cleanup_orphans) ========================================================== 15:55:23 (1713297323) multiop /mnt/lustre/f23.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6919 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:55:26 (1713297326) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:55:40 (1713297340) targets are mounted 15:55:40 (1713297340) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 23 (22s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 24: open(O_CREAT), replay, unlink, close (test mds_cleanup_orphans) ========================================================== 15:55:47 (1713297347) multiop /mnt/lustre/f24.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6919 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:55:49 (1713297349) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:56:02 (1713297362) targets are mounted 15:56:02 (1713297362) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 24 (21s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 25: open(O_CREAT), unlink, replay, close (test mds_cleanup_orphans) ========================================================== 15:56:09 (1713297369) multiop /mnt/lustre/f25.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6919 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:56:12 (1713297372) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:56:24 (1713297384) targets are mounted 15:56:24 (1713297384) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 25 (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 26: |X| open(O_CREAT), unlink two, close one, replay, close one (test mds_cleanup_orphans) ========================================================== 15:56:31 (1713297391) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre multiop /mnt/lustre/f26.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6919 multiop /mnt/lustre/f26.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6919 Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:56:34 (1713297394) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:56:48 (1713297408) targets are mounted 15:56:48 (1713297408) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 26 (21s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 27: |X| open(O_CREAT), unlink two, replay, close two (test mds_cleanup_orphans) ========================================================== 15:56:54 (1713297414) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre multiop /mnt/lustre/f27.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6919 multiop /mnt/lustre/f27.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6919 Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:56:57 (1713297417) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:57:10 (1713297430) targets are mounted 15:57:10 (1713297430) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 27 (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 28: open(O_CREAT), |X| unlink two, close one, replay, close one (test mds_cleanup_orphans) ========================================================== 15:57:16 (1713297436) multiop /mnt/lustre/f28.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6919 multiop /mnt/lustre/f28.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6919 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:57:19 (1713297439) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:57:33 (1713297453) targets are mounted 15:57:33 (1713297453) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 28 (22s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 29: open(O_CREAT), |X| unlink two, replay, close two (test mds_cleanup_orphans) ========================================================== 15:57:41 (1713297461) multiop /mnt/lustre/f29.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6919 multiop /mnt/lustre/f29.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6919 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:57:44 (1713297464) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:57:57 (1713297477) targets are mounted 15:57:57 (1713297477) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 29 (23s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 30: open(O_CREAT) two, unlink two, replay, close two (test mds_cleanup_orphans) ========================================================== 15:58:05 (1713297485) multiop /mnt/lustre/f30.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6919 multiop /mnt/lustre/f30.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6919 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:58:08 (1713297488) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:58:21 (1713297501) targets are mounted 15:58:21 (1713297501) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 30 (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 31: open(O_CREAT) two, unlink one, |X| unlink one, close two (test mds_cleanup_orphans) ========================================================== 15:58:27 (1713297507) multiop /mnt/lustre/f31.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6919 multiop /mnt/lustre/f31.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6919 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 15:58:30 (1713297510) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 15:58:43 (1713297523) targets are mounted 15:58:43 (1713297523) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 31 (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 32: close() notices client eviction; close() after client eviction ========================================================== 15:58:49 (1713297529) multiop /mnt/lustre/f32.replay-single vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.6919 multiop /mnt/lustre/f32.replay-single vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.6919 pdsh@oleg130-client: oleg130-client: ssh exited with exit code 5 PASS 32 (5s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 33a: fid seq shouldn't be reused after abort recovery ========================================================== 15:58:56 (1713297536) total: 10 open/close in 0.04 seconds: 230.46 ops/second Replay barrier on lustre-MDT0000 Stopping /mnt/lustre-mds1 (opts:) on oleg130-server Failover mds1 to oleg130-server Starting mds1: -o localrecov -o abort_recovery lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 total: 10 open/close in 0.05 seconds: 209.12 ops/second PASS 33a (11s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 33b: test fid seq allocation ======= 15:59:09 (1713297549) fail_loc=0x1311 total: 10 open/close in 0.04 seconds: 222.28 ops/second Replay barrier on lustre-MDT0000 Stopping /mnt/lustre-mds1 (opts:) on oleg130-server Failover mds1 to oleg130-server Starting mds1: -o localrecov -o abort_recovery lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg130-client: error: invalid path '/mnt/lustre': Input/output error pdsh@oleg130-client: oleg130-client: ssh exited with exit code 5 first stat failed: 5 total: 10 open/close in 0.05 seconds: 185.95 ops/second PASS 33b (9s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 34: abort recovery before client does replay (test mds_cleanup_orphans) ========================================================== 15:59:20 (1713297560) multiop /mnt/lustre/f34.replay-single vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.6919 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3456 2205184 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Stopping /mnt/lustre-mds1 (opts:) on oleg130-server Failover mds1 to oleg130-server Starting mds1: -o localrecov -o abort_recovery lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg130-client: error: invalid path '/mnt/lustre': Input/output error pdsh@oleg130-client: oleg130-client: ssh exited with exit code 5 first stat failed: 5 PASS 34 (9s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 35: test recovery from llog for unlink op ========================================================== 15:59:31 (1713297571) fail_loc=0x80000119 Stopping /mnt/lustre-mds1 (opts:) on oleg130-server Failover mds1 to oleg130-server Starting mds1: -o localrecov -o abort_recovery lustre-mdt1/mdt1 /mnt/lustre-mds1 rm: cannot remove '/mnt/lustre/f35.replay-single': Input/output error oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 Can't lstat /mnt/lustre/f35.replay-single: No such file or directory PASS 35 (12s) debug_raw_pointers=0 debug_raw_pointers=0 SKIP: replay-single test_36 skipping ALWAYS excluded test 36 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 37: abort recovery before client does replay (test mds_cleanup_orphans for directories) ========================================================== 15:59:45 (1713297585) multiop /mnt/lustre/d37.replay-single/f37.replay-single vdD_c TMPPIPE=/tmp/multiop_open_wait_pipe.6919 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3456 2205184 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Stopping /mnt/lustre-mds1 (opts:) on oleg130-server Failover mds1 to oleg130-server Starting mds1: -o localrecov -o abort_recovery lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 PASS 37 (11s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 38: test recovery from unlink llog (test llog_gen_rec) ========================================================== 15:59:58 (1713297598) total: 800 open/close in 3.96 seconds: 201.85 ops/second - unlinked 0 (time 1713297604 ; total 0 ; last 0) total: 400 unlinks in 1 seconds: 400.000000 unlinks/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3712 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:00:09 (1713297609) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:00:21 (1713297621) targets are mounted 16:00:21 (1713297621) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec - unlinked 0 (time 1713297626 ; total 0 ; last 0) total: 400 unlinks in 2 seconds: 200.000000 unlinks/second Can't lstat /mnt/lustre/f38.replay-single-*: No such file or directory PASS 38 (34s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 39: test recovery from unlink llog (test llog_gen_rec) ========================================================== 16:00:34 (1713297634) total: 800 open/close in 3.43 seconds: 233.15 ops/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3712 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre - unlinked 0 (time 1713297641 ; total 0 ; last 0) total: 400 unlinks in 1 seconds: 400.000000 unlinks/second Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:00:43 (1713297643) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:00:57 (1713297657) targets are mounted 16:00:57 (1713297657) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec - unlinked 0 (time 1713297662 ; total 0 ; last 0) total: 400 unlinks in 2 seconds: 200.000000 unlinks/second Can't lstat /mnt/lustre/f39.replay-single-*: No such file or directory PASS 39 (34s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 41: read from a valid osc while other oscs are invalid ========================================================== 16:01:10 (1713297670) 1+0 records in 1+0 records out 4096 bytes (4.1 kB) copied, 0.00140119 s, 2.9 MB/s 1+0 records in 1+0 records out 4096 bytes (4.1 kB) copied, 0.00320252 s, 1.3 MB/s PASS 41 (3s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 42: recovery after ost failure ===== 16:01:14 (1713297674) total: 800 open/close in 3.39 seconds: 235.95 ops/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre - unlinked 0 (time 1713297682 ; total 0 ; last 0) total: 400 unlinks in 1 seconds: 400.000000 unlinks/second debug=-1 Failing ost1 on oleg130-server Stopping /mnt/lustre-ost1 (opts:) on oleg130-server 16:01:25 (1713297685) shut down Failover ost1 to oleg130-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-OST0000 16:01:38 (1713297698) targets are mounted 16:01:38 (1713297698) facet_failover done wait for MDS to timeout and recover debug=super ioctl neterror warning dlmtrace error emerg ha rpctrace vfstrace config console lfsck debug=super ioctl neterror warning dlmtrace error emerg ha rpctrace vfstrace config console lfsck - unlinked 0 (time 1713297740 ; total 0 ; last 0) total: 400 unlinks in 1 seconds: 400.000000 unlinks/second Can't lstat /mnt/lustre/f42.replay-single-*: No such file or directory PASS 42 (68s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 43: mds osc import failure during recovery; don't LBUG ========================================================== 16:02:24 (1713297744) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3712 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre fail_loc=0x80000204 Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:02:27 (1713297747) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:02:41 (1713297761) targets are mounted 16:02:41 (1713297761) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 43 (32s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 44a: race in target handle connect ========================================================== 16:02:58 (1713297778) at_max=40 1 of 10 (1713297780) service : cur 5 worst 5 (at 1713297761, 18s ago) 4 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre 2 of 10 (1713297785) service : cur 5 worst 5 (at 1713297761, 23s ago) 5 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre 3 of 10 (1713297791) service : cur 5 worst 5 (at 1713297761, 29s ago) 5 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre 4 of 10 (1713297796) service : cur 5 worst 5 (at 1713297761, 35s ago) 5 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre 5 of 10 (1713297802) service : cur 5 worst 5 (at 1713297761, 40s ago) 5 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre 6 of 10 (1713297808) service : cur 5 worst 5 (at 1713297761, 46s ago) 5 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre 7 of 10 (1713297813) service : cur 5 worst 5 (at 1713297761, 52s ago) 5 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre 8 of 10 (1713297819) service : cur 5 worst 5 (at 1713297761, 57s ago) 5 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre 9 of 10 (1713297824) service : cur 5 worst 5 (at 1713297761, 63s ago) 5 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre 10 of 10 (1713297830) service : cur 5 worst 5 (at 1713297761, 68s ago) 5 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre fail_loc=0 at_max=600 PASS 44a (59s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 44b: race in target handle connect ========================================================== 16:03:58 (1713297838) 1 of 10 (1713297839) service : cur 5 worst 5 (at 1713297761, 77s ago) 5 0 0 0 fail_loc=0x80000704 error: recover: Connection timed out Filesystem 1K-blocks Used Available Use% Mounted on 192.168.201.130@tcp:/lustre 7542784 6144 7532544 1% /mnt/lustre 2 of 10 (1713297879) service : cur 40 worst 40 (at 1713297878, 0s ago) 40 0 0 0 fail_loc=0x80000704 error: recover: Connection timed out df: '/mnt/lustre': Operation already in progress 3 of 10 (1713297900) service : cur 40 worst 40 (at 1713297878, 21s ago) 40 0 0 0 fail_loc=0x80000704 error: recover: Connection timed out df: '/mnt/lustre': Operation already in progress 4 of 10 (1713297920) service : cur 40 worst 40 (at 1713297878, 41s ago) 40 40 0 0 fail_loc=0x80000704 error: recover: Connection timed out df: '/mnt/lustre': Operation already in progress 5 of 10 (1713297940) service : cur 40 worst 40 (at 1713297878, 62s ago) 40 40 0 0 fail_loc=0x80000704 error: recover: Connection timed out df: '/mnt/lustre': Operation already in progress 6 of 10 (1713297961) service : cur 40 worst 40 (at 1713297878, 82s ago) 40 40 0 0 fail_loc=0x80000704 error: recover: Connection timed out df: '/mnt/lustre': Operation already in progress 7 of 10 (1713297981) service : cur 40 worst 40 (at 1713297878, 103s ago) 40 40 0 0 fail_loc=0x80000704 error: recover: Connection timed out df: '/mnt/lustre': Operation already in progress 8 of 10 (1713298002) service : cur 40 worst 40 (at 1713297878, 123s ago) 40 40 0 0 fail_loc=0x80000704 error: recover: Connection timed out df: '/mnt/lustre': Operation already in progress 9 of 10 (1713298022) service : cur 40 worst 40 (at 1713297878, 144s ago) 40 40 0 0 fail_loc=0x80000704 error: recover: Connection timed out df: '/mnt/lustre': Operation already in progress 10 of 10 (1713298043) service : cur 40 worst 40 (at 1713297878, 164s ago) 40 40 0 0 fail_loc=0x80000704 error: recover: Connection timed out df: '/mnt/lustre': Operation already in progress PASS 44b (226s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 44c: race in target handle connect ========================================================== 16:07:46 (1713298066) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre total: 100 create in 0.22 seconds: 459.80 ops/second fail_loc=0x80000712 Stopping /mnt/lustre-mds1 (opts:) on oleg130-server Failover mds1 to oleg130-server Starting mds1: -o localrecov -o abort_recovery lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg130-client: error: invalid path '/mnt/lustre': Input/output error pdsh@oleg130-client: oleg130-client: ssh exited with exit code 5 first stat failed: 5 unlink(/mnt/lustre/f44c.replay-single-0) error: No such file or directory total: 0 unlinks in 0 seconds: -nan unlinks/second Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:07:56 (1713298076) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:08:08 (1713298088) targets are mounted 16:08:08 (1713298088) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec unlink(/mnt/lustre/f44c.replay-single-0) error: No such file or directory total: 0 unlinks in 0 seconds: -nan unlinks/second PASS 44c (27s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 45: Handle failed close ============ 16:08:15 (1713298095) multiop /mnt/lustre/f45.replay-single vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.6919 /mnt/lustre/f45.replay-single has type file OK PASS 45 (3s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 46: Don't leak file handle after open resend (3325) ========================================================== 16:08:20 (1713298100) fail_loc=0x122 fail_loc=0 Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:08:38 (1713298118) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:08:51 (1713298131) targets are mounted 16:08:51 (1713298131) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec lfs path2fid: cannot get fid for 'f46.replay-single': No such file or directory PASS 46 (36s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 47: MDS->OSC failure during precreate cleanup (2824) ========================================================== 16:08:58 (1713298138) total: 20 open/close in 0.07 seconds: 282.78 ops/second Failing ost1 on oleg130-server Stopping /mnt/lustre-ost1 (opts:) on oleg130-server 16:08:59 (1713298139) shut down Failover ost1 to oleg130-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-OST0000 16:09:12 (1713298152) targets are mounted 16:09:12 (1713298152) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec fail_loc=0x80000204 total: 20 open/close in 0.06 seconds: 348.78 ops/second - unlinked 0 (time 1713298217 ; total 0 ; last 0) total: 20 unlinks in 0 seconds: inf unlinks/second PASS 47 (80s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 48: MDS->OSC failure during precreate cleanup (2824) ========================================================== 16:10:20 (1713298220) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre total: 20 open/close in 0.07 seconds: 296.20 ops/second Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:10:22 (1713298222) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:10:35 (1713298235) targets are mounted 16:10:35 (1713298235) facet_failover done fail_loc=0x80000216 total: 20 open/close in 0.07 seconds: 278.54 ops/second - unlinked 0 (time 1713298296 ; total 0 ; last 0) total: 40 unlinks in 0 seconds: inf unlinks/second PASS 48 (77s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 50: Double OSC recovery, don't LASSERT (3812) ========================================================== 16:11:39 (1713298299) PASS 50 (7s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 52: time out lock replay (3764) ==== 16:11:47 (1713298307) multiop /mnt/lustre/f52.replay-single vs_s TMPPIPE=/tmp/multiop_open_wait_pipe.6919 fail_loc=0x80000157 Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:11:49 (1713298309) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:12:02 (1713298322) targets are mounted 16:12:02 (1713298322) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec fail_loc=0x0 fail_loc=0x0 PASS 52 (33s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53a: |X| close request while two MDC requests in flight ========================================================== 16:12:22 (1713298342) fail_loc=0x80000115 fail_loc=0 Replay barrier on lustre-MDT0000 Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:12:26 (1713298346) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:12:38 (1713298358) targets are mounted 16:12:38 (1713298358) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d53a.replay-single-1/f has type file OK /mnt/lustre/d53a.replay-single-2/f has type file OK PASS 53a (21s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53b: |X| open request while two MDC requests in flight ========================================================== 16:12:44 (1713298364) multiop /mnt/lustre/d53b.replay-single-1/f vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.6919 fail_loc=0x80000107 fail_loc=0 Replay barrier on lustre-MDT0000 Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:12:48 (1713298368) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:13:00 (1713298380) targets are mounted 16:13:00 (1713298380) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d53b.replay-single-1/f has type file OK /mnt/lustre/d53b.replay-single-2/f has type file OK PASS 53b (21s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53c: |X| open request and close request while two MDC requests in flight ========================================================== 16:13:06 (1713298386) fail_loc=0x80000107 fail_loc=0x80000115 Replay barrier on lustre-MDT0000 Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:13:10 (1713298390) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:13:22 (1713298402) targets are mounted 16:13:22 (1713298402) facet_failover done fail_loc=0 /mnt/lustre/d53c.replay-single-1/f has type file OK /mnt/lustre/d53c.replay-single-2/f has type file OK PASS 53c (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53d: close reply while two MDC requests in flight ========================================================== 16:13:28 (1713298408) fail_loc=0x8000013b fail_loc=0 Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:13:31 (1713298411) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:13:43 (1713298423) targets are mounted 16:13:43 (1713298423) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d53d.replay-single-1/f has type file OK /mnt/lustre/d53d.replay-single-2/f has type file OK PASS 53d (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53e: |X| open reply while two MDC requests in flight ========================================================== 16:13:49 (1713298429) fail_loc=0x119 fail_loc=0 Replay barrier on lustre-MDT0000 Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:13:53 (1713298433) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:14:05 (1713298445) targets are mounted 16:14:05 (1713298445) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d53e.replay-single-1/f has type file OK /mnt/lustre/d53e.replay-single-2/f has type file OK PASS 53e (21s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53f: |X| open reply and close reply while two MDC requests in flight ========================================================== 16:14:11 (1713298451) fail_loc=0x119 fail_loc=0x8000013b Replay barrier on lustre-MDT0000 Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:14:15 (1713298455) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:14:28 (1713298468) targets are mounted 16:14:28 (1713298468) facet_failover done fail_loc=0 /mnt/lustre/d53f.replay-single-1/f has type file OK /mnt/lustre/d53f.replay-single-2/f has type file OK PASS 53f (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53g: |X| drop open reply and close request while close and open are both in flight ========================================================== 16:14:33 (1713298473) fail_loc=0x119 fail_loc=0x80000115 fail_loc=0 Replay barrier on lustre-MDT0000 Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:14:37 (1713298477) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:14:49 (1713298489) targets are mounted 16:14:49 (1713298489) facet_failover done /mnt/lustre/d53g.replay-single-1/f has type file OK /mnt/lustre/d53g.replay-single-2/f has type file OK PASS 53g (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53h: open request and close reply while two MDC requests in flight ========================================================== 16:14:54 (1713298494) fail_loc=0x80000107 fail_loc=0x8000013b Replay barrier on lustre-MDT0000 Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:14:59 (1713298499) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:15:11 (1713298511) targets are mounted 16:15:11 (1713298511) facet_failover done fail_loc=0 /mnt/lustre/d53h.replay-single-1/f has type file OK /mnt/lustre/d53h.replay-single-2/f has type file OK PASS 53h (21s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 55: let MDS_CHECK_RESENT return the original return code instead of 0 ========================================================== 16:15:16 (1713298516) fail_loc=0x8000012b fail_loc=0x0 touch: rm: cannot touch '/mnt/lustre/f55.replay-single'cannot remove '/mnt/lustre/f55.replay-single': No such file or directory : No such file or directory PASS 55 (18s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 56: don't replay a symlink open request (3440) ========================================================== 16:15:35 (1713298535) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:15:38 (1713298538) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:15:50 (1713298550) targets are mounted 16:15:50 (1713298550) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 56 (29s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 57: test recovery from llog for setattr op ========================================================== 16:16:06 (1713298566) fail_loc=0x8000012c UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:16:09 (1713298569) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:16:21 (1713298581) targets are mounted 16:16:21 (1713298581) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec affected facets: mds1 oleg130-server: oleg130-server.virtnet: executing _wait_recovery_complete *.lustre-MDT0000.recovery_status 1475 oleg130-server: *.lustre-MDT0000.recovery_status status: COMPLETE Waiting for orphan cleanup... osp.lustre-OST0000-osc-MDT0000.old_sync_processed osp.lustre-OST0001-osc-MDT0000.old_sync_processed wait 40 secs maximumly for oleg130-server mds-ost sync done. /mnt/lustre/f57.replay-single has type file OK fail_loc=0x0 PASS 57 (23s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 58a: test recovery from llog for setattr op (test llog_gen_rec) ========================================================== 16:16:30 (1713298590) fail_loc=0x8000012c total: 2500 open/close in 4.69 seconds: 532.56 ops/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3712 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:16:39 (1713298599) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:16:51 (1713298611) targets are mounted 16:16:51 (1713298611) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec fail_loc=0x0 - unlinked 0 (time 1713298623 ; total 0 ; last 0) total: 2500 unlinks in 3 seconds: 833.333313 unlinks/second PASS 58a (37s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 58b: test replay of setxattr op ==== 16:17:09 (1713298629) Starting client: oleg130-client.virtnet: -o user_xattr,flock oleg130-server@tcp:/lustre /mnt/lustre2 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:17:11 (1713298631) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:17:24 (1713298644) targets are mounted 16:17:24 (1713298644) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Stopping client oleg130-client.virtnet /mnt/lustre2 (opts:) oleg130-client.virtnet: executing wait_import_state_mount FULL mgc.*.mgs_server_uuid mgc.*.mgs_server_uuid in FULL state after 0 sec PASS 58b (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 58c: resend/reconstruct setxattr op ========================================================== 16:17:31 (1713298651) Starting client: oleg130-client.virtnet: -o user_xattr,flock oleg130-server@tcp:/lustre /mnt/lustre2 fail_val=0 fail_loc=0x123 fail_loc=0 fail_loc=0x119 fail_loc=0 Stopping client oleg130-client.virtnet /mnt/lustre2 (opts:) PASS 58c (40s) debug_raw_pointers=0 debug_raw_pointers=0 SKIP: replay-single test_59 skipping ALWAYS excluded test 59 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 60: test llog post recovery init vs llog unlink ========================================================== 16:18:13 (1713298693) total: 200 open/close in 0.37 seconds: 537.94 ops/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3712 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre - unlinked 0 (time 1713298695 ; total 0 ; last 0) total: 100 unlinks in 0 seconds: inf unlinks/second Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:18:16 (1713298696) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:18:29 (1713298709) targets are mounted 16:18:29 (1713298709) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec - unlinked 0 (time 1713298712 ; total 0 ; last 0) total: 100 unlinks in 0 seconds: inf unlinks/second PASS 60 (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 61a: test race llog recovery vs llog cleanup ========================================================== 16:18:35 (1713298715) total: 800 open/close in 1.44 seconds: 553.68 ops/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3712 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre - unlinked 0 (time 1713298719 ; total 0 ; last 0) total: 800 unlinks in 1 seconds: 800.000000 unlinks/second fail_val=0 fail_loc=0x80000221 Failing ost1 on oleg130-server Stopping /mnt/lustre-ost1 (opts:) on oleg130-server 16:18:42 (1713298722) shut down Failover ost1 to oleg130-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-OST0000 16:18:54 (1713298734) targets are mounted 16:18:54 (1713298734) facet_failover done Failing ost1 on oleg130-server Stopping /mnt/lustre-ost1 (opts:) on oleg130-server 16:19:06 (1713298746) shut down Failover ost1 to oleg130-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-OST0000 16:19:18 (1713298758) targets are mounted 16:19:18 (1713298758) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec fail_val=0 fail_loc=0x0 Can't lstat /mnt/lustre/d61a.replay-single/f61a.replay-single-*: No such file or directory PASS 61a (78s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 61b: test race mds llog sync vs llog cleanup ========================================================== 16:19:55 (1713298795) fail_loc=0x8000013a Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:19:57 (1713298797) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:20:10 (1713298810) targets are mounted 16:20:10 (1713298810) facet_failover done Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:20:21 (1713298821) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:20:34 (1713298834) targets are mounted 16:20:34 (1713298834) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec 1+0 records in 1+0 records out 4096 bytes (4.1 kB) copied, 0.00259391 s, 1.6 MB/s PASS 61b (43s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 61c: test race mds llog sync vs llog cleanup ========================================================== 16:20:40 (1713298840) fail_val=0 fail_loc=0x80000222 Failing ost1 on oleg130-server Stopping /mnt/lustre-ost1 (opts:) on oleg130-server 16:20:52 (1713298852) shut down Failover ost1 to oleg130-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-OST0000 16:21:05 (1713298865) targets are mounted 16:21:05 (1713298865) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec fail_val=0 fail_loc=0x0 PASS 61c (30s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 61d: error in llog_setup should cleanup the llog context correctly ========================================================== 16:21:11 (1713298871) Stopping /mnt/lustre-mds1 (opts:) on oleg130-server fail_loc=0x80000605 Starting mgs: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: mount.lustre: mount lustre-mdt1/mdt1 at /mnt/lustre-mds1 failed: Operation not supported pdsh@oleg130-client: oleg130-server: ssh exited with exit code 95 Start of lustre-mdt1/mdt1 on mgs failed 95 fail_loc=0 Starting mgs: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 PASS 61d (7s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 62: don't mis-drop resent replay === 16:21:20 (1713298880) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3712 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre total: 25 open/close in 0.08 seconds: 313.94 ops/second fail_loc=0x80000707 Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:21:24 (1713298884) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:21:36 (1713298896) targets are mounted 16:21:36 (1713298896) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec fail_loc=0 - unlinked 0 (time 1713298916 ; total 0 ; last 0) total: 25 unlinks in 0 seconds: inf unlinks/second PASS 62 (38s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 65a: AT: verify early replies ====== 16:22:00 (1713298920) at_history=8 at_history=8 debug=other fail_val=11000 fail_loc=0x8000050a 00000100:00001000:2.0:1713298949.584908:0:7017:0:(client.c:537:ptlrpc_at_recv_early_reply()) @@@ Early reply #1, new deadline in 35s (25s) req@ffff88012d0fbb80 x1796521467738112/t0(0) o101->lustre-MDT0000-mdc-ffff8800ab38f800@192.168.201.130@tcp:12/10 lens 664/66320 e 1 to 0 dl 1713298984 ref 2 fl Rpc:PQr/200/ffffffff rc 0/-1 job:'createmany.0' uid:0 gid:0 portal 12 : cur 36 worst 40 (at 1713297879, 1075s ago) 36 5 5 5 portal 23 : cur 5 worst 10 (at 1713296681, 2273s ago) 10 5 5 10 portal 30 : cur 5 worst 5 (at 1713296645, 2309s ago) 5 0 0 5 portal 17 : cur 5 worst 5 (at 1713296711, 2243s ago) 5 5 5 5 portal 13 : cur 10 worst 10 (at 1713297030, 1924s ago) 10 10 0 0 portal 12 : cur 5 worst 40 (at 1713297879, 1084s ago) 36 5 5 5 portal 23 : cur 5 worst 10 (at 1713296681, 2282s ago) 10 5 5 10 portal 30 : cur 5 worst 5 (at 1713296645, 2318s ago) 5 0 0 5 portal 17 : cur 5 worst 5 (at 1713296711, 2252s ago) 5 5 5 5 portal 13 : cur 10 worst 10 (at 1713297030, 1933s ago) 10 10 0 0 PASS 65a (45s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 65b: AT: verify early replies on packed reply / bulk ========================================================== 16:22:46 (1713298966) at_history=8 at_history=8 debug=other trace fail_val=11 fail_loc=0x224 fail_loc=0 00000100:00001000:2.0:1713298995.984686:0:1920:0:(client.c:537:ptlrpc_at_recv_early_reply()) @@@ Early reply #1, new deadline in 35s (25s) req@ffff8800a8ae4a80 x1796521467747520/t0(0) o4->lustre-OST0000-osc-ffff8800ab38f800@192.168.201.130@tcp:6/4 lens 4584/448 e 1 to 0 dl 1713299030 ref 2 fl Rpc:Qr/200/ffffffff rc 0/-1 job:'multiop.0' uid:0 gid:0 debug=super ioctl neterror warning dlmtrace error emerg ha rpctrace vfstrace config console lfsck debug=super ioctl neterror warning dlmtrace error emerg ha rpctrace vfstrace config console lfsck portal 28 : cur 5 worst 5 (at 1713296641, 2360s ago) 5 5 5 5 portal 7 : cur 5 worst 5 (at 1713296642, 2359s ago) 5 5 5 5 portal 17 : cur 5 worst 5 (at 1713296686, 2315s ago) 5 5 5 5 portal 6 : cur 36 worst 36 (at 1713299000, 1s ago) 36 0 0 0 PASS 65b (36s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 66a: AT: verify MDT service time adjusts with no early replies ========================================================== 16:23:24 (1713299004) at_history=8 at_history=8 portal 12 : cur 5 worst 40 (at 1713297879, 1147s ago) 36 5 5 5 fail_val=5000 fail_loc=0x8000050a portal 12 : cur 5 worst 40 (at 1713297879, 1153s ago) 36 5 5 5 fail_val=10000 fail_loc=0x8000050a portal 12 : cur 36 worst 40 (at 1713297879, 1163s ago) 36 36 5 5 fail_loc=0 portal 12 : cur 5 worst 40 (at 1713297879, 1173s ago) 36 36 5 5 Current MDT timeout 5, worst 40 PASS 66a (49s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 66b: AT: verify net latency adjusts ========================================================== 16:24:14 (1713299054) at_history=8 at_history=8 fail_val=10 fail_loc=0x50c fail_loc=0 network timeout orig 5, cur 10, worst 10 PASS 66b (84s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 67a: AT: verify slow request processing doesn't induce reconnects ========================================================== 16:25:39 (1713299139) at_history=8 at_history=8 fail_val=400 fail_loc=0x50a fail_loc=0 0 osc reconnect attempts on gradual slow PASS 67a (73s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 67b: AT: verify instant slowdown doesn't induce reconnects ========================================================== 16:26:53 (1713299213) at_history=8 at_history=8 Creating to objid 5441 on ost lustre-OST0000... fail_val=20000 fail_loc=0x80000223 total: 17 open/close in 0.05 seconds: 324.90 ops/second Connected clients: oleg130-client.virtnet oleg130-client.virtnet service : cur 5 worst 5 (at 1713296415, 2821s ago) 1 0 1 0 phase 2 0 osc reconnect attempts on instant slow fail_loc=0x80000223 fail_loc=0 Connected clients: oleg130-client.virtnet oleg130-client.virtnet service : cur 5 worst 5 (at 1713296415, 2822s ago) 1 0 1 0 0 osc reconnect attempts on 2nd slow PASS 67b (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 68: AT: verify slowing locks ======= 16:27:22 (1713299242) at_history=8 at_history=8 /home/green/git/lustre-release/lustre/tests/replay-single.sh: line 1968: $ldlm_enqueue_min: ambiguous redirect fail_val=19 fail_loc=0x80000312 fail_val=25 fail_loc=0x80000312 fail_loc=0 /home/green/git/lustre-release/lustre/tests/replay-single.sh: line 1983: $ldlm_enqueue_min: ambiguous redirect PASS 68 (68s) debug_raw_pointers=0 debug_raw_pointers=0 Cleaning up AT ... debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 70a: check multi client t-f ======== 16:28:33 (1713299313) SKIP: replay-single test_70a Need two or more clients, have 1 SKIP 70a (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 70b: dbench 1mdts recovery; 1 clients ========================================================== 16:28:36 (1713299316) Starting client oleg130-client.virtnet: -o user_xattr,flock oleg130-server@tcp:/lustre /mnt/lustre Started clients oleg130-client.virtnet: 192.168.201.130@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) + MISSING_DBENCH_OK= + PATH=/opt/iozone/bin:/opt/iozone/bin:/home/green/git/lustre-release/lustre/tests/mpi:/home/green/git/lustre-release/lustre/tests/racer:/home/green/git/lustre-release/lustre/../lustre-iokit/sgpdd-survey:/home/green/git/lustre-release/lustre/tests:/home/green/git/lustre-release/lustre/utils/gss:/home/green/git/lustre-release/lustre/utils:/opt/iozone/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin::/home/green/git/lustre-release/lustre/utils:/home/green/git/lustre-release/lustre/tests:/sbin:/usr/sbin:/home/green/git/lustre-release/lustre/utils:/home/green/git/lustre-release/lustre/tests/: + DBENCH_LIB= + TESTSUITE=replay-single + TESTNAME=test_70b + MOUNT=/mnt/lustre ++ hostname + DIR=/mnt/lustre/d70b.replay-single/oleg130-client.virtnet + LCTL=/home/green/git/lustre-release/lustre/utils/lctl + rundbench 1 -t 300 dbench: no process found Started rundbench load pid=14070 ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3968 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 32768 3683328 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3716096 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 35840 7399424 1% /mnt/lustre test_70b fail mds1 1 times Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:28:44 (1713299324) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:28:57 (1713299337) targets are mounted 16:28:57 (1713299337) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid oleg130-client.virtnet: looking for dbench program oleg130-client.virtnet: /usr/bin/dbench oleg130-client.virtnet: creating output directory /mnt/lustre/d70b.replay-single/oleg130-client.virtnet oleg130-client.virtnet: mkdir: created directory '/mnt/lustre/d70b.replay-single' oleg130-client.virtnet: mkdir: created directory '/mnt/lustre/d70b.replay-single/oleg130-client.virtnet' oleg130-client.virtnet: found dbench client file /usr/share/dbench/client.txt oleg130-client.virtnet: '/usr/share/dbench/client.txt' -> 'client.txt' oleg130-client.virtnet: running 'dbench 1 -t 300' on /mnt/lustre/d70b.replay-single/oleg130-client.virtnet at Tue Apr 16 16:28:37 EDT 2024 oleg130-client.virtnet: waiting for dbench pid 14095 oleg130-client.virtnet: dbench version 4.00 - Copyright Andrew Tridgell 1999-2004 oleg130-client.virtnet: oleg130-client.virtnet: Running for 300 seconds with load 'client.txt' and minimum warmup 60 secs oleg130-client.virtnet: failed to create barrier semaphore oleg130-client.virtnet: 0 of 1 processes prepared for launch 0 sec oleg130-client.virtnet: 1 of 1 processes prepared for launch 0 sec oleg130-client.virtnet: releasing clients oleg130-client.virtnet: 1 323 11.74 MB/sec warmup 1 sec latency 22.507 ms oleg130-client.virtnet: 1 674 10.59 MB/sec warmup 2 sec latency 100.799 ms oleg130-client.virtnet: 1 918 7.26 MB/sec warmup 3 sec latency 39.484 ms oleg130-client.virtnet: 1 1245 5.77 MB/sec warmup 4 sec latency 42.759 ms oleg130-client.virtnet: 1 1664 5.24 MB/sec warmup 5 sec latency 41.752 ms oleg130-client.virtnet: 1 1898 4.40 MB/sec warmup 6 sec latency 356.165 ms oleg130-client.virtnet: 1 1898 3.77 MB/sec warmup 7 sec latency 1356.486 ms oleg130-client.virtnet: 1 1898 3.30 MB/sec warmup 8 sec latency 2356.814 ms oleg130-client.virtnet: 1 1898 2.93 MB/sec warmup 9 sec latency 3357.188 ms oleg130-client.virtnet: 1 1898 2.64 MB/sec warmup 10 sec latency 4357.473 ms oleg130-client.virtnet: 1 1898 2.40 MB/sec warmup 11 sec latency 5357.767 ms oleg130-client.virtnet: 1 1898 2.20 MB/sec warmup 12 sec latency 6358.079 ms oleg130-client.virtnet: 1 1898 2.03 MB/sec warmup 13 sec latency 7358.393 ms oleg130-client.virtnet: 1 1898 1.88 MB/sec warmup 14 sec latency 8358.737 ms oleg130-client.virtnet: 1 1898 1.76 MB/sec warmup 15 sec latency 9359.001 ms oleg130-client.virtnet: 1 1898 1.65 MB/sec warmup 16 sec latency 10359.242 ms oleg130-client.virtnet: 1 1898 1.55 MB/sec warmup 17 sec latency 11359.479 ms oleg130-client.virtnet: 1 1898 1.47 MB/sec warmup 18 sec latency 12359.713 ms oleg130-client.virtnet: 1 1898 1.39 MB/sec warmup 19 sec latency 13359.943 ms oleg130-client.virtnet: 1 1898 1.32 MB/sec warmup 20 sec latency 14360.205 ms oleg130-client.virtnet: 1 1898 1.26 MB/sec warmup 21 sec latency 15360.433 ms oleg130-client.virtnet: 1 1898 1.20 MB/sec warmup 22 sec latency 16360.648 ms oleg130-client.virtnet: 1 1898 1.15 MB/sec warmup 23 sec latency 17360.907 ms oleg130-client.virtnet: 1 1898 1.10 MB/sec warmup 24 sec latency 18361.114 ms oleg130-client.virtnet: 1 1898 1.06 MB/sec warmup 25 sec latency 19361.353 ms oleg130-client.virtnet: 1 1898 1.01 MB/sec warmup 26 sec latency 20361.629 ms oleg130-client.virtnet: 1 1898 0.98 MB/sec warmup 27 sec latency 21361.909 ms oleg130-client.virtnet: 1 1898 0.94 MB/sec warmup 28 sec latency 22362.166 ms oleg130-client.virtnet: 1 1898 0.91 MB/sec warmup 29 sec latency 23362.463 ms oleg130-client.virtnet: 1 1898 0.88 MB/sec warmup 30 sec latency 24362.642 ms oleg130-client.virtnet: 1 1898 0.85 MB/sec warmup 31 sec latency 25362.890 ms oleg130-client.virtnet: 1 2093 0.83 MB/sec warmup 32 secmdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3840 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 33792 3665920 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 13312 3721216 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 47104 7387136 1% /mnt/lustre test_70b fail mds1 2 times Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:29:17 (1713299357) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:29:30 (1713299370) targets are mounted 16:29:30 (1713299370) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3840 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 40960 3713024 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 14336 3737600 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 55296 7450624 1% /mnt/lustre test_70b fail mds1 3 times Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:29:41 (1713299381) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:29:53 (1713299393) targets are mounted 16:29:53 (1713299393) facet_failover done latency 25860.803 ms oleg130-client.virtnet: 1 2436 0.85 MB/sec warmup 33 sec latency 55.507 ms oleg130-client.virtnet: 1 2978 0.99 MB/sec warmup 34 sec latency 40.186 ms oleg130-client.virtnet: 1 3560 1.10 MB/sec warmup 35 sec latency 49.109 ms oleg130-client.virtnet: 1 3871 1.10 MB/sec warmup 36 sec latency 16.176 ms oleg130-client.virtnet: 1 4161 1.08 MB/sec warmup 37 sec latency 16.400 ms oleg130-client.virtnet: 1 4377 1.07 MB/sec warmup 38 sec latency 59.386 ms oleg130-client.virtnet: 1 4667 1.08 MB/sec warmup 39 sec latency 66.594 ms oleg130-client.virtnet: 1 4695 1.05 MB/sec warmup 40 sec latency 943.548 ms oleg130-client.virtnet: 1 4695 1.03 MB/sec warmup 41 sec latency 1943.785 ms oleg130-client.virtnet: 1 4695 1.00 MB/sec warmup 42 sec latency 2944.007 ms oleg130-client.virtnet: 1 4695 0.98 MB/sec warmup 43 sec latency 3944.288 ms oleg130-client.virtnet: 1 4695 0.96 MB/sec warmup 44 sec latency 4944.551 ms oleg130-client.virtnet: 1 4695 0.94 MB/sec warmup 45 sec latency 5944.815 ms oleg130-client.virtnet: 1 4695 0.92 MB/sec warmup 46 sec latency 6945.022 ms oleg130-client.virtnet: 1 4695 0.90 MB/sec warmup 47 sec latency 7945.362 ms oleg130-client.virtnet: 1 4695 0.88 MB/sec warmup 48 sec latency 8945.614 ms oleg130-client.virtnet: 1 4695 0.86 MB/sec warmup 49 sec latency 9945.839 ms oleg130-client.virtnet: 1 4695 0.84 MB/sec warmup 50 sec latency 10946.072 ms oleg130-client.virtnet: 1 4695 0.83 MB/sec warmup 51 sec latency 11946.395 ms oleg130-client.virtnet: 1 4695 0.81 MB/sec warmup 52 sec latency 12946.571 ms oleg130-client.virtnet: 1 4695 0.79 MB/sec warmup 53 sec latency 13946.847 ms oleg130-client.virtnet: 1 4695 0.78 MB/sec warmup 54 sec latency 14947.094 ms oleg130-client.virtnet: 1 4695 0.77 MB/sec warmup 55 sec latency 15947.366 ms oleg130-client.virtnet: 1 4695 0.75 MB/sec warmup 56 sec latency 16947.602 ms oleg130-client.virtnet: 1 5052 0.79 MB/sec warmup 57 sec latency 17390.080 ms oleg130-client.virtnet: 1 5453 0.78 MB/sec warmup 58 sec latency 31.418 ms oleg130-client.virtnet: 1 5933 0.80 MB/sec warmup 59 sec latency 43.667 ms oleg130-client.virtnet: 1 7234 5.27 MB/sec execute 1 sec latency 59.050 ms oleg130-client.virtnet: 1 7498 3.29 MB/sec execute 2 sec latency 18.137 ms oleg130-client.virtnet: 1 7834 2.31 MB/sec execute 3 sec latency 55.695 ms oleg130-client.virtnet: 1 7844 1.74 MB/sec execute 4 sec latency 898.999 ms oleg130-client.virtnet: 1 7844 1.39 MB/sec execute 5 sec latency 1899.200 ms oleg130-client.virtnet: 1 7844 1.16 MB/sec execute 6 sec latency 2899.408 ms oleg130-client.virtnet: 1 7844 0.99 MB/sec execute 7 sec latency 3899.617 ms oleg130-client.virtnet: 1 7844 0.87 MB/sec execute 8 sec latency 4899.831 ms oleg130-client.virtnet: 1 7844 0.77 MB/sec execute 9 sec latency 5900.078 ms oleg130-client.virtnet: 1 7844 0.69 MB/sec execute 10 sec latency 6900.251 ms oleg130-client.virtnet: 1 7844 0.63 MB/sec execute 11 sec latency 7900.500 ms oleg130-client.virtnet: 1 7844 0.58 MB/sec execute 12 sec latency 8900.689 ms oleg130-client.virtnet: 1 7844 0.53 MB/sec execute 13 sec latency 9900.962 ms oleg130-client.virtnet: 1 7844 0.50 MB/sec execute 14 sec latency 10901.121 ms oleg130-client.virtnet: 1 7844 0.46 MB/sec execute 15 sec latency 11901.281 ms oleg130-client.virtnet: 1 7844 0.43 MB/sec execute 16 sec latency 12901.529 ms oleg130-client.virtnet: 1 7844 0.41 MB/sec execute 17 sec latency 13901.745 ms oleg130-client.virtnetoleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3968 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 41984 3709952 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 13312 3745792 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 55296 7455744 1% /mnt/lustre test_70b fail mds1 4 times Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:30:05 (1713299405) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:30:18 (1713299418) targets are mounted 16:30:18 (1713299418) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3840 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 40960 3707904 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 14336 3716096 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 55296 7424000 1% /mnt/lustre test_70b fail mds1 5 times Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:30:30 (1713299430) shut down : 1 7844 0.39 MB/sec execute 18 sec latency 14901.954 ms oleg130-client.virtnet: 1 7844 0.37 MB/sec execute 19 sec latency 15902.243 ms oleg130-client.virtnet: 1 7844 0.35 MB/sec execute 20 sec latency 16902.499 ms oleg130-client.virtnet: 1 7844 0.33 MB/sec execute 21 sec latency 17902.751 ms oleg130-client.virtnet: 1 8066 0.34 MB/sec execute 22 sec latency 18374.780 ms oleg130-client.virtnet: 1 8690 0.51 MB/sec execute 23 sec latency 34.370 ms oleg130-client.virtnet: 1 9266 0.51 MB/sec execute 24 sec latency 33.871 ms oleg130-client.virtnet: 1 10076 0.77 MB/sec execute 25 sec latency 29.769 ms oleg130-client.virtnet: 1 10891 0.98 MB/sec execute 26 sec latency 25.705 ms oleg130-client.virtnet: 1 11359 0.96 MB/sec execute 27 sec latency 38.097 ms oleg130-client.virtnet: 1 11615 0.95 MB/sec execute 28 sec latency 397.866 ms oleg130-client.virtnet: 1 11615 0.92 MB/sec execute 29 sec latency 1398.023 ms oleg130-client.virtnet: 1 11615 0.89 MB/sec execute 30 sec latency 2398.173 ms oleg130-client.virtnet: 1 11615 0.86 MB/sec execute 31 sec latency 3398.399 ms oleg130-client.virtnet: 1 11615 0.83 MB/sec execute 32 sec latency 4398.608 ms oleg130-client.virtnet: 1 11615 0.80 MB/sec execute 33 sec latency 5398.791 ms oleg130-client.virtnet: 1 11615 0.78 MB/sec execute 34 sec latency 6398.986 ms oleg130-client.virtnet: 1 11615 0.76 MB/sec execute 35 sec latency 7399.204 ms oleg130-client.virtnet: 1 11615 0.74 MB/sec execute 36 sec latency 8399.488 ms oleg130-client.virtnet: 1 11615 0.72 MB/sec execute 37 sec latency 9399.722 ms oleg130-client.virtnet: 1 11615 0.70 MB/sec execute 38 sec latency 10399.936 ms oleg130-client.virtnet: 1 11615 0.68 MB/sec execute 39 sec latency 11400.169 ms oleg130-client.virtnet: 1 11615 0.66 MB/sec execute 40 sec latency 12400.332 ms oleg130-client.virtnet: 1 11615 0.65 MB/sec execute 41 sec latency 13400.437 ms oleg130-client.virtnet: 1 11615 0.63 MB/sec execute 42 sec latency 14400.562 ms oleg130-client.virtnet: 1 11615 0.62 MB/sec execute 43 sec latency 15400.830 ms oleg130-client.virtnet: 1 11615 0.60 MB/sec execute 44 sec latency 16400.966 ms oleg130-client.virtnet: 1 11615 0.59 MB/sec execute 45 sec latency 17401.115 ms oleg130-client.virtnet: 1 11615 0.58 MB/sec execute 46 sec latency 18401.244 ms oleg130-client.virtnet: 1 11929 0.59 MB/sec execute 47 sec latency 18875.436 ms oleg130-client.virtnet: 1 12544 0.65 MB/sec execute 48 sec latency 28.924 ms oleg130-client.virtnet: 1 13234 0.69 MB/sec execute 49 sec latency 29.997 ms oleg130-client.virtnet: 1 14157 0.86 MB/sec execute 50 sec latency 40.015 ms oleg130-client.virtnet: 1 14583 0.88 MB/sec execute 51 sec latency 13.636 ms oleg130-client.virtnet: 1 14996 0.87 MB/sec execute 52 sec latency 36.223 ms oleg130-client.virtnet: 1 15321 0.88 MB/sec execute 53 sec latency 376.294 ms oleg130-client.virtnet: 1 15321 0.87 MB/sec execute 54 sec latency 1376.608 ms oleg130-client.virtnet: 1 15321 0.85 MB/sec execute 55 sec latency 2376.890 ms oleg130-client.virtnet: 1 15321 0.84 MB/sec execute 56 sec latency 3377.075 ms oleg130-client.virtnet: 1 15321 0.82 MB/sec execute 57 sec latency 4377.249 ms oleg130-client.virtnet: 1 15321 0.81 MB/sec execute 58 sec latency 5377.430 ms oleg130-client.virtnet: 1 15321 0.79 MB/sec execute 59 sec latency 6377.672 ms oleg130-client.virtnet: 1 15321 0.78 MB/sec execute 60 sec latency 7377.971 ms oleg130-client.virtnet: 1 15321 0.77 MB/sec execute 61 sec latency 8378.170 ms oleg130-client.virtnet: 1 Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:30:43 (1713299443) targets are mounted 16:30:43 (1713299443) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3968 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 40960 3715072 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 14336 3737600 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 55296 7452672 1% /mnt/lustre test_70b fail mds1 6 times Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:30:55 (1713299455) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:31:08 (1713299468) targets are mounted 16:31:08 (1713299468) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3968 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 40960 3707904 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 14336 3731456 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 55296 7439360 1% /mnt/lustre test_70b fail mds1 7 times Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:31:21 (1713299481) shut down 15321 0.76 MB/sec execute 62 sec latency 9378.351 ms oleg130-client.virtnet: 1 15321 0.74 MB/sec execute 63 sec latency 10378.511 ms oleg130-client.virtnet: 1 15321 0.73 MB/sec execute 64 sec latency 11378.682 ms oleg130-client.virtnet: 1 15321 0.72 MB/sec execute 65 sec latency 12378.890 ms oleg130-client.virtnet: 1 15321 0.71 MB/sec execute 66 sec latency 13379.071 ms oleg130-client.virtnet: 1 15321 0.70 MB/sec execute 67 sec latency 14379.214 ms oleg130-client.virtnet: 1 15321 0.69 MB/sec execute 68 sec latency 15379.411 ms oleg130-client.virtnet: 1 15321 0.68 MB/sec execute 69 sec latency 16379.568 ms oleg130-client.virtnet: 1 15321 0.67 MB/sec execute 70 sec latency 17379.709 ms oleg130-client.virtnet: 1 15321 0.66 MB/sec execute 71 sec latency 18379.904 ms oleg130-client.virtnet: 1 15620 0.69 MB/sec execute 72 sec latency 19011.018 ms oleg130-client.virtnet: 1 16211 0.69 MB/sec execute 73 sec latency 26.552 ms oleg130-client.virtnet: 1 16968 0.75 MB/sec execute 74 sec latency 27.294 ms oleg130-client.virtnet: 1 17956 0.85 MB/sec execute 75 sec latency 20.982 ms oleg130-client.virtnet: 1 18466 0.84 MB/sec execute 76 sec latency 34.304 ms oleg130-client.virtnet: 1 18992 0.86 MB/sec execute 77 sec latency 21.843 ms oleg130-client.virtnet: 1 19429 0.89 MB/sec execute 78 sec latency 396.362 ms oleg130-client.virtnet: 1 19429 0.88 MB/sec execute 79 sec latency 1396.524 ms oleg130-client.virtnet: 1 19429 0.87 MB/sec execute 80 sec latency 2396.713 ms oleg130-client.virtnet: 1 19429 0.85 MB/sec execute 81 sec latency 3396.883 ms oleg130-client.virtnet: 1 19429 0.84 MB/sec execute 82 sec latency 4397.121 ms oleg130-client.virtnet: 1 19429 0.83 MB/sec execute 83 sec latency 5397.357 ms oleg130-client.virtnet: 1 19429 0.82 MB/sec execute 84 sec latency 6397.577 ms oleg130-client.virtnet: 1 19429 0.81 MB/sec execute 85 sec latency 7397.753 ms oleg130-client.virtnet: 1 19429 0.80 MB/sec execute 86 sec latency 8397.909 ms oleg130-client.virtnet: 1 19429 0.80 MB/sec execute 87 sec latency 9398.058 ms oleg130-client.virtnet: 1 19429 0.79 MB/sec execute 88 sec latency 10398.269 ms oleg130-client.virtnet: 1 19429 0.78 MB/sec execute 89 sec latency 11398.459 ms oleg130-client.virtnet: 1 19429 0.77 MB/sec execute 90 sec latency 12398.606 ms oleg130-client.virtnet: 1 19429 0.76 MB/sec execute 91 sec latency 13398.848 ms oleg130-client.virtnet: 1 19429 0.75 MB/sec execute 92 sec latency 14399.062 ms oleg130-client.virtnet: 1 19429 0.74 MB/sec execute 93 sec latency 15399.369 ms oleg130-client.virtnet: 1 19429 0.74 MB/sec execute 94 sec latency 16399.657 ms oleg130-client.virtnet: 1 19429 0.73 MB/sec execute 95 sec latency 17399.804 ms oleg130-client.virtnet: 1 19429 0.72 MB/sec execute 96 sec latency 18400.005 ms oleg130-client.virtnet: 1 19521 0.71 MB/sec execute 97 sec latency 19135.888 ms oleg130-client.virtnet: 1 20144 0.72 MB/sec execute 98 sec latency 31.397 ms oleg130-client.virtnet: 1 20748 0.77 MB/sec execute 99 sec latency 37.848 ms oleg130-client.virtnet: 1 21411 0.82 MB/sec execute 100 sec latency 45.456 ms oleg130-client.virtnet: 1 21720 0.82 MB/sec execute 101 sec latency 16.124 ms oleg130-client.virtnet: 1 22003 0.82 MB/sec execute 102 sec latency 54.373 ms oleg130-client.virtnet: 1 22319 0.82 MB/sec execute 103 sec latency 41.550 ms oleg130-client.virtnet: 1 22575 0.82 MB/sec execute 104 sec latency 472.691 ms oleg130-client.virtnet: 1 22575 0.81 MB/sec execute 105 sec latency 1472.907 ms oleg130-client.virtnet: 1 22575 Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:31:34 (1713299494) targets are mounted 16:31:34 (1713299494) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3840 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 40960 3705856 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 13312 3736576 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 54272 7442432 1% /mnt/lustre test_70b fail mds1 8 times Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:31:45 (1713299505) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:31:58 (1713299518) targets are mounted 16:31:58 (1713299518) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec 0.80 MB/sec execute 106 sec latency 2473.215 ms oleg130-client.virtnet: 1 22575 0.80 MB/sec execute 107 sec latency 3473.482 ms oleg130-client.virtnet: 1 22575 0.79 MB/sec execute 108 sec latency 4473.744 ms oleg130-client.virtnet: 1 22575 0.78 MB/sec execute 109 sec latency 5473.968 ms oleg130-client.virtnet: 1 22575 0.77 MB/sec execute 110 sec latency 6474.233 ms oleg130-client.virtnet: 1 22575 0.77 MB/sec execute 111 sec latency 7474.452 ms oleg130-client.virtnet: 1 22575 0.76 MB/sec execute 112 sec latency 8474.598 ms oleg130-client.virtnet: 1 22575 0.75 MB/sec execute 113 sec latency 9474.753 ms oleg130-client.virtnet: 1 22575 0.75 MB/sec execute 114 sec latency 10474.907 ms oleg130-client.virtnet: 1 22575 0.74 MB/sec execute 115 sec latency 11475.158 ms oleg130-client.virtnet: 1 22575 0.73 MB/sec execute 116 sec latency 12475.387 ms oleg130-client.virtnet: 1 22575 0.73 MB/sec execute 117 sec latency 13475.485 ms oleg130-client.virtnet: 1 22575 0.72 MB/sec execute 118 sec latency 14475.696 ms oleg130-client.virtnet: 1 22575 0.72 MB/sec execute 119 sec latency 15476.027 ms oleg130-client.virtnet: 1 22575 0.71 MB/sec execute 120 sec latency 16476.276 ms oleg130-client.virtnet: 1 22575 0.70 MB/sec execute 121 sec latency 17476.476 ms oleg130-client.virtnet: 1 22917 0.72 MB/sec execute 122 sec latency 17992.577 ms oleg130-client.virtnet: 1 23517 0.72 MB/sec execute 123 sec latency 33.089 ms oleg130-client.virtnet: 1 24225 0.77 MB/sec execute 124 sec latency 42.085 ms oleg130-client.virtnet: 1 25075 0.82 MB/sec execute 125 sec latency 30.937 ms oleg130-client.virtnet: 1 25535 0.81 MB/sec execute 126 sec latency 10.824 ms oleg130-client.virtnet: 1 25814 0.81 MB/sec execute 127 sec latency 42.363 ms oleg130-client.virtnet: 1 26284 0.84 MB/sec execute 128 sec latency 355.024 ms oleg130-client.virtnet: 1 26284 0.83 MB/sec execute 129 sec latency 1355.166 ms oleg130-client.virtnet: 1 26284 0.83 MB/sec execute 130 sec latency 2355.429 ms oleg130-client.virtnet: 1 26284 0.82 MB/sec execute 131 sec latency 3355.655 ms oleg130-client.virtnet: 1 26284 0.81 MB/sec execute 132 sec latency 4355.870 ms oleg130-client.virtnet: 1 26284 0.81 MB/sec execute 133 sec latency 5356.082 ms oleg130-client.virtnet: 1 26284 0.80 MB/sec execute 134 sec latency 6356.363 ms oleg130-client.virtnet: 1 26284 0.79 MB/sec execute 135 sec latency 7356.549 ms oleg130-client.virtnet: 1 26284 0.79 MB/sec execute 136 sec latency 8356.755 ms oleg130-client.virtnet: 1 26284 0.78 MB/sec execute 137 sec latency 9357.087 ms oleg130-client.virtnet: 1 26284 0.78 MB/sec execute 138 sec latency 10357.321 ms oleg130-client.virtnet: 1 26284 0.77 MB/sec execute 139 sec latency 11357.502 ms oleg130-client.virtnet: 1 26284 0.77 MB/sec execute 140 sec latency 12357.657 ms oleg130-client.virtnet: 1 26284 0.76 MB/sec execute 141 sec latency 13357.857 ms oleg130-client.virtnet: 1 26284 0.76 MB/sec execute 142 sec latency 14357.984 ms oleg130-client.virtnet: 1 26284 0.75 MB/sec execute 143 sec latency 15358.222 ms oleg130-client.virtnet: 1 26284 0.75 MB/sec execute 144 sec latency 16358.407 ms oleg130-client.virtnet: 1 26284 0.74 MB/sec execute 145 sec latency 17358.601 ms oleg130-client.virtnet: 1 26284 0.73 MB/sec execute 146 sec latency 18358.858 ms oleg130-client.virtnet: 1 26468 0.73 MB/sec execute 147 sec latency 18988.443 ms oleg130-client.virtnet: 1 26899 0.73 MB/sec execute 148 sec latency 37.473 ms oleg130-client.virtnet: 1 27354 0.74 MB/sec execute 149 sec latency 41.524 ms oleg130-client.virtnet: 1 28078 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3840 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 40960 3708928 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 14336 3738624 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 55296 7447552 1% /mnt/lustre test_70b fail mds1 9 times Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:32:17 (1713299537) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:32:29 (1713299549) targets are mounted 16:32:29 (1713299549) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3968 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 40960 3716096 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 14336 3736576 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 55296 7452672 1% /mnt/lustre test_70b fail mds1 10 times Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:32:46 (1713299566) shut down 0.77 MB/sec execute 150 sec latency 37.556 ms oleg130-client.virtnet: 1 28598 0.80 MB/sec execute 151 sec latency 26.693 ms oleg130-client.virtnet: 1 28926 0.80 MB/sec execute 152 sec latency 14.933 ms oleg130-client.virtnet: 1 29341 0.80 MB/sec execute 153 sec latency 35.753 ms oleg130-client.virtnet: 1 29393 0.79 MB/sec execute 154 sec latency 887.418 ms oleg130-client.virtnet: 1 29393 0.79 MB/sec execute 155 sec latency 1887.660 ms oleg130-client.virtnet: 1 29393 0.78 MB/sec execute 156 sec latency 2887.833 ms oleg130-client.virtnet: 1 29393 0.78 MB/sec execute 157 sec latency 3887.994 ms oleg130-client.virtnet: 1 29393 0.77 MB/sec execute 158 sec latency 4888.128 ms oleg130-client.virtnet: 1 29393 0.77 MB/sec execute 159 sec latency 5888.338 ms oleg130-client.virtnet: 1 29393 0.76 MB/sec execute 160 sec latency 6888.509 ms oleg130-client.virtnet: 1 29393 0.76 MB/sec execute 161 sec latency 7888.707 ms oleg130-client.virtnet: 1 29393 0.75 MB/sec execute 162 sec latency 8888.931 ms oleg130-client.virtnet: 1 29393 0.75 MB/sec execute 163 sec latency 9889.134 ms oleg130-client.virtnet: 1 29393 0.75 MB/sec execute 164 sec latency 10889.429 ms oleg130-client.virtnet: 1 29393 0.74 MB/sec execute 165 sec latency 11889.582 ms oleg130-client.virtnet: 1 29393 0.74 MB/sec execute 166 sec latency 12889.726 ms oleg130-client.virtnet: 1 29393 0.73 MB/sec execute 167 sec latency 13889.875 ms oleg130-client.virtnet: 1 29393 0.73 MB/sec execute 168 sec latency 14890.037 ms oleg130-client.virtnet: 1 29393 0.72 MB/sec execute 169 sec latency 15890.199 ms oleg130-client.virtnet: 1 29393 0.72 MB/sec execute 170 sec latency 16890.447 ms oleg130-client.virtnet: 1 29393 0.71 MB/sec execute 171 sec latency 17890.593 ms oleg130-client.virtnet: 1 29393 0.71 MB/sec execute 172 sec latency 18890.690 ms oleg130-client.virtnet: 1 29393 0.71 MB/sec execute 173 sec latency 19890.866 ms oleg130-client.virtnet: 1 29393 0.70 MB/sec execute 174 sec latency 20891.008 ms oleg130-client.virtnet: 1 29393 0.70 MB/sec execute 175 sec latency 21891.185 ms oleg130-client.virtnet: 1 29393 0.69 MB/sec execute 176 sec latency 22891.404 ms oleg130-client.virtnet: 1 29393 0.69 MB/sec execute 177 sec latency 23891.561 ms oleg130-client.virtnet: 1 29393 0.69 MB/sec execute 178 sec latency 24891.746 ms oleg130-client.virtnet: 1 29393 0.68 MB/sec execute 179 sec latency 25891.901 ms oleg130-client.virtnet: 1 29393 0.68 MB/sec execute 180 sec latency 26892.124 ms oleg130-client.virtnet: 1 29393 0.68 MB/sec execute 181 sec latency 27892.304 ms oleg130-client.virtnet: 1 29621 0.68 MB/sec execute 182 sec latency 28520.022 ms oleg130-client.virtnet: 1 30309 0.69 MB/sec execute 183 sec latency 26.960 ms oleg130-client.virtnet: 1 30915 0.70 MB/sec execute 184 sec latency 39.428 ms oleg130-client.virtnet: 1 31976 0.75 MB/sec execute 185 sec latency 23.997 ms oleg130-client.virtnet: 1 32290 0.75 MB/sec execute 186 sec latency 16.173 ms oleg130-client.virtnet: 1 32655 0.75 MB/sec execute 187 sec latency 31.067 ms oleg130-client.virtnet: 1 32967 0.75 MB/sec execute 188 sec latency 43.262 ms oleg130-client.virtnet: 1 33065 0.75 MB/sec execute 189 sec latency 784.234 ms oleg130-client.virtnet: 1 33065 0.75 MB/sec execute 190 sec latency 1784.492 ms oleg130-client.virtnet: 1 33065 0.75 MB/sec execute 191 sec latency 2784.817 ms oleg130-client.virtnet: 1 33065 0.74 MB/sec execute 192 sec latency 3785.095 ms oleg130-client.virtnet: 1 33065 0.74 MB/sec execute 193 sec latency 4785.468 ms oleg130-client.virtnet: 1 33065 0.73 Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:33:00 (1713299580) targets are mounted 16:33:00 (1713299580) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3968 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 40960 3705856 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 14336 3721216 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 55296 7427072 1% /mnt/lustre test_70b fail mds1 11 times Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:33:12 (1713299592) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:33:24 (1713299604) targets are mounted 16:33:24 (1713299604) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3968 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 40960 3717120 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 14336 3735552 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 55296 7452672 1% /mnt/lustre test_70b fail mds1 12 times MB/sec execute 194 sec latency 5785.746 ms oleg130-client.virtnet: 1 33065 0.73 MB/sec execute 195 sec latency 6786.019 ms oleg130-client.virtnet: 1 33065 0.73 MB/sec execute 196 sec latency 7786.418 ms oleg130-client.virtnet: 1 33065 0.72 MB/sec execute 197 sec latency 8786.745 ms oleg130-client.virtnet: 1 33065 0.72 MB/sec execute 198 sec latency 9787.060 ms oleg130-client.virtnet: 1 33065 0.72 MB/sec execute 199 sec latency 10787.410 ms oleg130-client.virtnet: 1 33065 0.71 MB/sec execute 200 sec latency 11787.784 ms oleg130-client.virtnet: 1 33065 0.71 MB/sec execute 201 sec latency 12788.182 ms oleg130-client.virtnet: 1 33065 0.71 MB/sec execute 202 sec latency 13788.480 ms oleg130-client.virtnet: 1 33065 0.70 MB/sec execute 203 sec latency 14788.640 ms oleg130-client.virtnet: 1 33065 0.70 MB/sec execute 204 sec latency 15788.800 ms oleg130-client.virtnet: 1 33065 0.69 MB/sec execute 205 sec latency 16788.953 ms oleg130-client.virtnet: 1 33065 0.69 MB/sec execute 206 sec latency 17789.174 ms oleg130-client.virtnet: 1 33107 0.69 MB/sec execute 207 sec latency 18636.887 ms oleg130-client.virtnet: 1 33632 0.70 MB/sec execute 208 sec latency 45.446 ms oleg130-client.virtnet: 1 34311 0.71 MB/sec execute 209 sec latency 21.067 ms oleg130-client.virtnet: 1 34865 0.73 MB/sec execute 210 sec latency 42.072 ms oleg130-client.virtnet: 1 35364 0.74 MB/sec execute 211 sec latency 44.150 ms oleg130-client.virtnet: 1 35783 0.75 MB/sec execute 212 sec latency 15.984 ms oleg130-client.virtnet: 1 36096 0.75 MB/sec execute 213 sec latency 13.965 ms oleg130-client.virtnet: 1 36375 0.75 MB/sec execute 214 sec latency 262.872 ms oleg130-client.virtnet: 1 36375 0.75 MB/sec execute 215 sec latency 1263.060 ms oleg130-client.virtnet: 1 36375 0.74 MB/sec execute 216 sec latency 2263.309 ms oleg130-client.virtnet: 1 36375 0.74 MB/sec execute 217 sec latency 3263.621 ms oleg130-client.virtnet: 1 36375 0.73 MB/sec execute 218 sec latency 4263.844 ms oleg130-client.virtnet: 1 36375 0.73 MB/sec execute 219 sec latency 5264.077 ms oleg130-client.virtnet: 1 36375 0.73 MB/sec execute 220 sec latency 6264.300 ms oleg130-client.virtnet: 1 36375 0.72 MB/sec execute 221 sec latency 7264.472 ms oleg130-client.virtnet: 1 36375 0.72 MB/sec execute 222 sec latency 8264.739 ms oleg130-client.virtnet: 1 36375 0.72 MB/sec execute 223 sec latency 9264.887 ms oleg130-client.virtnet: 1 36375 0.72 MB/sec execute 224 sec latency 10265.018 ms oleg130-client.virtnet: 1 36375 0.71 MB/sec execute 225 sec latency 11265.193 ms oleg130-client.virtnet: 1 36375 0.71 MB/sec execute 226 sec latency 12265.317 ms oleg130-client.virtnet: 1 36375 0.71 MB/sec execute 227 sec latency 13265.510 ms oleg130-client.virtnet: 1 36375 0.70 MB/sec execute 228 sec latency 14265.698 ms oleg130-client.virtnet: 1 36375 0.70 MB/sec execute 229 sec latency 15265.942 ms oleg130-client.virtnet: 1 36375 0.70 MB/sec execute 230 sec latency 16266.167 ms oleg130-client.virtnet: 1 36375 0.69 MB/sec execute 231 sec latency 17266.382 ms oleg130-client.virtnet: 1 36388 0.69 MB/sec execute 232 sec latency 18229.008 ms oleg130-client.virtnet: 1 36744 0.69 MB/sec execute 233 sec latency 47.594 ms oleg130-client.virtnet: 1 37167 0.70 MB/sec execute 234 sec latency 39.727 ms oleg130-client.virtnet: 1 37571 0.70 MB/sec execute 235 sec latency 40.591 ms oleg130-client.virtnet: 1 37957 0.71 MB/sec execute 236 sec latency 54.367 ms oleg130-client.virtnet: 1 38514 0.73 MB/sec execute 237 sec latency 57.409 ms oleg130-client.virtnet: 1 39161 0.75 MB/sec eFailing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:33:37 (1713299617) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:33:50 (1713299630) targets are mounted 16:33:50 (1713299630) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec xecute 238 sec latency 32.723 ms oleg130-client.virtnet: 1 39470 0.75 MB/sec execute 239 sec latency 14.207 ms oleg130-client.virtnet: 1 39489 0.74 MB/sec execute 240 sec latency 900.386 ms oleg130-client.virtnet: 1 39489 0.74 MB/sec execute 241 sec latency 1900.646 ms oleg130-client.virtnet: 1 39489 0.74 MB/sec execute 242 sec latency 2900.941 ms oleg130-client.virtnet: 1 39489 0.74 MB/sec execute 243 sec latency 3901.187 ms oleg130-client.virtnet: 1 39489 0.73 MB/sec execute 244 sec latency 4901.404 ms oleg130-client.virtnet: 1 39489 0.73 MB/sec execute 245 sec latency 5901.624 ms oleg130-client.virtnet: 1 39489 0.73 MB/sec execute 246 sec latency 6901.934 ms oleg130-client.virtnet: 1 39489 0.72 MB/sec execute 247 sec latency 7902.250 ms oleg130-client.virtnet: 1 39489 0.72 MB/sec execute 248 sec latency 8902.509 ms oleg130-client.virtnet: 1 39489 0.72 MB/sec execute 249 sec latency 9902.729 ms oleg130-client.virtnet: 1 39489 0.71 MB/sec execute 250 sec latency 10902.961 ms oleg130-client.virtnet: 1 39489 0.71 MB/sec execute 251 sec latency 11903.164 ms oleg130-client.virtnet: 1 39489 0.71 MB/sec execute 252 sec latency 12903.383 ms oleg130-client.virtnet: 1 39489 0.71 MB/sec execute 253 sec latency 13903.571 ms oleg130-client.virtnet: 1 39489 0.70 MB/sec execute 254 sec latency 14903.762 ms oleg130-client.virtnet: 1 39489 0.70 MB/sec execute 255 sec latency 15904.029 ms oleg130-client.virtnet: 1 39489 0.70 MB/sec execute 256 sec latency 16904.230 ms oleg130-client.virtnet: 1 39586 0.70 MB/sec execute 257 sec latency 17643.354 ms oleg130-client.virtnet: 1 39803 0.69 MB/sec execute 258 sec latency 72.601 ms oleg130-client.virtnet: 1 40050 0.69 MB/sec execute 259 sec latency 51.592 ms oleg130-client.virtnet: 1 40504 0.71 MB/sec execute 260 sec latency 16.883 ms oleg130-client.virtnet: 1 40888 0.71 MB/sec execute 261 sec latency 43.980 ms oleg130-client.virtnet: 1 41525 0.71 MB/sec execute 262 sec latency 36.908 ms oleg130-client.virtnet: 1 42050 0.73 MB/sec execute 263 sec latency 40.997 ms oleg130-client.virtnet: 1 42637 0.74 MB/sec execute 264 sec latency 41.153 ms oleg130-client.virtnet: 1 42981 0.75 MB/sec execute 265 sec latency 15.294 ms oleg130-client.virtnet: 1 43284 0.74 MB/sec execute 266 sec latency 65.860 ms oleg130-client.virtnet: 1 43760 0.75 MB/sec execute 267 sec latency 40.591 ms oleg130-client.virtnet: 1 44246 0.76 MB/sec execute 268 sec latency 51.225 ms oleg130-client.virtnet: 1 44567 0.76 MB/sec execute 269 sec latency 52.519 ms oleg130-client.virtnet: 1 44972 0.76 MB/sec execute 270 sec latency 45.189 ms oleg130-client.virtnet: 1 45366 0.77 MB/sec execute 271 sec latency 42.620 ms oleg130-client.virtnet: 1 45948 0.78 MB/sec execute 272 sec latency 43.095 ms oleg130-client.virtnet: 1 46407 0.79 MB/sec execute 273 sec latency 16.242 ms oleg130-client.virtnet: 1 46700 0.79 MB/sec execute 274 sec latency 15.582 ms oleg130-client.virtnet: 1 47144 0.79 MB/sec execute 275 sec latency 30.242 ms oleg130-client.virtnet: 1 47864 0.81 MB/sec execute 276 sec latency 30.366 ms oleg130-client.virtnet: 1 48575 0.81 MB/sec execute 277 sec latency 25.275 ms oleg130-client.virtnet: 1 49389 0.83 MB/sec execute 278 sec latency 32.408 ms oleg130-client.virtnet: 1 50034 0.85 MB/sec execute 279 sec latency 14.495 ms oleg130-client.virtnet: 1 50489 0.84 MB/sec execute 280 sec latency 30.975 ms oleg130-client.virtnet: 1 50933 0.85 MB/sec execute 281 sec latency 42.396 ms oleg130-client.virtnet: 1 51491 0.86 MB/sec execute 282 sec latency 33.017 ms oleg130-client.virtnet: 1 52169 0.86 MB/sec execute 283 sec latency 22.253 ms oleg130-client.virtnet: 1 53235 0.89 MB/sec execute 284 sec latency 20.122 ms oleg130-client.virtnet: 1 53692 0.90 MB/sec execute 285 sec latency 13.319 ms oleg130-client.virtnet: 1 54135 0.89 MB/sec execute 286 sec latency 41.191 ms oleg130-client.virtnet: 1 54742 0.91 MB/sec execute 287 sec latency 46.164 ms oleg130-client.virtnet: 1 55299 0.91 MB/sec execute 288 sec latency 22.227 ms oleg130-client.virtnet: 1 55997 0.92 MB/sec execute 289 sec latency 37.394 ms oleg130-client.virtnet: 1 56539 0.93 MB/sec execute 290 sec latency 41.771 ms oleg130-client.virtnet: 1 56963 0.94 MB/sec execute 291 sec latency 23.617 ms oleg130-client.virtnet: 1 57285 0.94 MB/sec execute 292 sec latency 16.111 ms oleg130-client.virtnet: 1 57600 0.94 MB/sec execute 293 sec latency 44.386 ms oleg130-client.virtnet: 1 58019 0.94 MB/sec execute 294 sec latency 29.565 ms oleg130-client.virtnet: 1 58418 0.95 MB/sec execute 295 sec latency 47.228 ms oleg130-client.virtnet: 1 58744 0.95 MB/sec execute 296 sec latency 40.971 ms oleg130-client.virtnet: 1 59187 0.95 MB/sec execute 297 sec latency 34.768 ms oleg130-client.virtnet: 1 59776 0.96 MB/sec execute 298 sec latency 39.359 ms oleg130-client.virtnet: 1 60431 0.98 MB/sec execute 299 sec latency 35.683 ms oleg130-client.virtnet: 1 cleanup 300 sec oleg130-client.virtnet: 0 cleanup 300 sec oleg130-client.virtnet: oleg130-client.virtnet: Operation Count AvgLat MaxLat oleg130-client.virtnet: ---------------------------------------- oleg130-client.virtnet: NTCreateX 9376 16.507 28520.013 oleg130-client.virtnet: Close 6872 6.694 19135.862 oleg130-client.virtnet: Rename 398 53.155 17992.559 oleg130-client.virtnet: Unlink 1900 2.413 23.786 oleg130-client.virtnet: Qpathinfo 8479 1.531 37.576 oleg130-client.virtnet: Qfileinfo 1474 0.149 2.676 oleg130-client.virtnet: Qfsinfo 1574 0.128 5.483 oleg130-client.virtnet: Sfileinfo 750 53.395 18875.426 oleg130-client.virtnet: Find 3295 0.484 13.254 oleg130-client.virtnet: WriteX 4619 0.855 6.726 oleg130-client.virtnet: ReadX 14786 0.020 6.917 oleg130-client.virtnet: LockX 30 1.288 2.014 oleg130-client.virtnet: UnlockX 30 1.305 2.090 oleg130-client.virtnet: Flush 653 21.136 72.560 oleg130-client.virtnet: oleg130-client.virtnet: Throughput 0.976737 MB/sec 1 clients 1 procs max_latency=28520.022 ms oleg130-client.virtnet: stopping dbench on /mnt/lustre/d70b.replay-single/oleg130-client.virtnet at Tue Apr 16 16:34:38 EDT 2024 with return code 0 oleg130-client.virtnet: clean dbench files on /mnt/lustre/d70b.replay-single/oleg130-client.virtnet oleg130-client.virtnet: /mnt/lustre/d70b.replay-single/oleg130-client.virtnet /mnt/lustre/d70b.replay-single/oleg130-client.virtnet oleg130-client.virtnet: removed directory: 'clients/client0/~dmtmp/PM' oleg130-client.virtnet: removed directory: 'clients/client0/~dmtmp/PWRPNT' oleg130-client.virtnet: removed directory: 'clients/client0/~dmtmp/PARADOX' oleg130-client.virtnet: removed directory: 'clients/client0/~dmtmp/EXCEL' oleg130-client.virtnet: removed directory: 'clients/client0/~dmtmp/SEED' oleg130-client.virtnet: removed directory: 'clients/client0/~dmtmp/WORDPRO' oleg130-client.virtnet: removed directory: 'clients/client0/~dmtmp/COREL' oleg130-client.virtnet: removed directory: 'clients/client0/~dmtmp/WORD' oleg130-client.virtnet: removed directory: 'clients/client0/~dmtmp/ACCESS' oleg130-client.virtnet: removed directory: 'clients/client0/~dmtmp' oleg130-client.virtnet: removed directory: 'clients/client0' oleg130-client.virtnet: removed directory: 'clients' oleg130-client.virtnet: removed 'client.txt' oleg130-client.virtnet: /mnt/lustre/d70b.replay-single/oleg130-client.virtnet oleg130-client.virtnet: dbench successfully finished PASS 70b (363s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 70c: tar 1mdts recovery ============ 16:34:42 (1713299682) Starting client oleg130-client.virtnet: -o user_xattr,flock oleg130-server@tcp:/lustre /mnt/lustre Started clients oleg130-client.virtnet: 192.168.201.130@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Started tar 27267 tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 12032 2196480 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 24576 3218432 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 14336 3233792 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 38912 6452224 1% /mnt/lustre test_70c fail mds1 1 times Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:36:56 (1713299816) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:37:09 (1713299829) targets are mounted 16:37:09 (1713299829) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 12032 2196608 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 13312 3213312 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 29696 3208192 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 43008 6421504 1% /mnt/lustre tar: Removing leading `/' from member names test_70c fail mds1 2 times Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:39:35 (1713299975) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:39:49 (1713299989) targets are mounted 16:39:49 (1713299989) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid tar: Removing leading `/' from hard link targets mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 70c (400s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 70d: mkdir/rmdir striped dir 1mdts recovery ========================================================== 16:41:23 (1713300083) SKIP: replay-single test_70d needs >= 2 MDTs SKIP 70d (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 70e: rename cross-MDT with random fails ========================================================== 16:41:26 (1713300086) SKIP: replay-single test_70e needs >= 2 MDTs SKIP 70e (2s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 70f: OSS O_DIRECT recovery with 1 clients ========================================================== 16:41:30 (1713300090) mount clients oleg130-client.virtnet ... Starting client oleg130-client.virtnet: -o user_xattr,flock oleg130-server@tcp:/lustre /mnt/lustre Started clients oleg130-client.virtnet: 192.168.201.130@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 4224 2204416 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 9216 3747840 1% /mnt/lustre[OST:0] Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... lustre-OST0001_UUID 3771392 7168 3753984 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 16384 7501824 1% /mnt/lustre ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... test_70f failing OST 1 times Failing ost1 on oleg130-server Stopping /mnt/lustre-ost1 (opts:) on oleg130-server 16:41:38 (1713300098) shut down Failover ost1 to oleg130-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Started lustre-OST0000 16:41:52 (1713300112) targets are mounted 16:41:52 (1713300112) facet_failover done Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 4224 2204416 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 6144 3750912 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 7168 3746816 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 13312 7497728 1% /mnt/lustre ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... test_70f failing OST 2 times Failing ost1 on oleg130-server Stopping /mnt/lustre-ost1 (opts:) on oleg130-server 16:42:06 (1713300126) shut down Failover ost1 to oleg130-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Started lustre-OST0000 16:42:20 (1713300140) targets are mounted 16:42:20 (1713300140) facet_failover done Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 4224 2204416 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 9216 3723264 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 7168 3744768 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 16384 7468032 1% /mnt/lustre Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... test_70f failing OST 3 times Failing ost1 on oleg130-server Stopping /mnt/lustre-ost1 (opts:) on oleg130-server 16:42:33 (1713300153) shut down Failover ost1 to oleg130-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... Started lustre-OST0000 16:42:47 (1713300167) targets are mounted 16:42:47 (1713300167) facet_failover done ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 4096 2204416 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 9216 3760128 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 7168 3762176 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 16384 7522304 1% /mnt/lustre ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... test_70f failing OST 4 times Failing ost1 on oleg130-server Stopping /mnt/lustre-ost1 (opts:) on oleg130-server 16:43:00 (1713300180) shut down Failover ost1 to oleg130-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... Started lustre-OST0000 16:43:14 (1713300194) targets are mounted 16:43:14 (1713300194) facet_failover done ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 4096 2204416 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 9216 3758080 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 7168 3762176 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 16384 7520256 1% /mnt/lustre ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... test_70f failing OST 5 times Failing ost1 on oleg130-server Stopping /mnt/lustre-ost1 (opts:) on oleg130-server 16:43:27 (1713300207) shut down Failover ost1 to oleg130-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... Started lustre-OST0000 16:43:41 (1713300221) targets are mounted 16:43:41 (1713300221) facet_failover done ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... ldlm.namespaces.MGC192.168.201.130@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800ab38f800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800ab38f800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg130-client.virtnet' ... PASS 70f (139s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 71a: mkdir/rmdir striped dir with 2 mdts recovery ========================================================== 16:43:51 (1713300231) SKIP: replay-single test_71a needs >= 2 MDTs SKIP 71a (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 73a: open(O_CREAT), unlink, replay, reconnect before open replay, close ========================================================== 16:43:54 (1713300234) multiop /mnt/lustre/f73a.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6919 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 4096 2204416 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3764224 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3764224 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7528448 1% /mnt/lustre fail_loc=0x80000302 Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:43:58 (1713300238) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:44:11 (1713300251) targets are mounted 16:44:11 (1713300251) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 73a (35s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 73b: open(O_CREAT), unlink, replay, reconnect at open_replay reply, close ========================================================== 16:44:30 (1713300270) multiop /mnt/lustre/f73b.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6919 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3840 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3764224 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7530496 1% /mnt/lustre fail_loc=0x80000157 Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:44:33 (1713300273) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:44:46 (1713300286) targets are mounted 16:44:46 (1713300286) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 73b (34s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 74: Ensure applications don't fail waiting for OST recovery ========================================================== 16:45:05 (1713300305) Stopping clients: oleg130-client.virtnet /mnt/lustre (opts:) Stopping client oleg130-client.virtnet /mnt/lustre opts: Stopping /mnt/lustre-ost1 (opts:) on oleg130-server Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:45:08 (1713300308) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:45:20 (1713300320) targets are mounted 16:45:20 (1713300320) facet_failover done Starting client oleg130-client.virtnet: -o user_xattr,flock oleg130-server@tcp:/lustre /mnt/lustre Started clients oleg130-client.virtnet: 192.168.201.130@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-OST0000 PASS 74 (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80a: DNE: create remote dir, drop update rep from MDT0, fail MDT0 ========================================================== 16:45:27 (1713300327) SKIP: replay-single test_80a needs >= 2 MDTs SKIP 80a (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80b: DNE: create remote dir, drop update rep from MDT0, fail MDT1 ========================================================== 16:45:29 (1713300329) SKIP: replay-single test_80b needs >= 2 MDTs SKIP 80b (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80c: DNE: create remote dir, drop update rep from MDT1, fail MDT[0,1] ========================================================== 16:45:31 (1713300331) SKIP: replay-single test_80c needs >= 2 MDTs SKIP 80c (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80d: DNE: create remote dir, drop update rep from MDT1, fail 2 MDTs ========================================================== 16:45:33 (1713300333) SKIP: replay-single test_80d needs >= 2 MDTs SKIP 80d (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80e: DNE: create remote dir, drop MDT1 rep, fail MDT0 ========================================================== 16:45:35 (1713300335) SKIP: replay-single test_80e needs >= 2 MDTs SKIP 80e (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80f: DNE: create remote dir, drop MDT1 rep, fail MDT1 ========================================================== 16:45:37 (1713300337) SKIP: replay-single test_80f needs >= 2 MDTs SKIP 80f (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80g: DNE: create remote dir, drop MDT1 rep, fail MDT0, then MDT1 ========================================================== 16:45:39 (1713300339) SKIP: replay-single test_80g needs >= 2 MDTs SKIP 80g (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80h: DNE: create remote dir, drop MDT1 rep, fail 2 MDTs ========================================================== 16:45:41 (1713300341) SKIP: replay-single test_80h needs >= 2 MDTs SKIP 80h (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81a: DNE: unlink remote dir, drop MDT0 update rep, fail MDT1 ========================================================== 16:45:43 (1713300343) SKIP: replay-single test_81a needs >= 2 MDTs SKIP 81a (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81b: DNE: unlink remote dir, drop MDT0 update reply, fail MDT0 ========================================================== 16:45:45 (1713300345) SKIP: replay-single test_81b needs >= 2 MDTs SKIP 81b (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81c: DNE: unlink remote dir, drop MDT0 update reply, fail MDT0,MDT1 ========================================================== 16:45:48 (1713300348) SKIP: replay-single test_81c needs >= 2 MDTs SKIP 81c (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81d: DNE: unlink remote dir, drop MDT0 update reply, fail 2 MDTs ========================================================== 16:45:51 (1713300351) SKIP: replay-single test_81d needs >= 2 MDTs SKIP 81d (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81e: DNE: unlink remote dir, drop MDT1 req reply, fail MDT0 ========================================================== 16:45:55 (1713300355) SKIP: replay-single test_81e needs >= 2 MDTs SKIP 81e (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81f: DNE: unlink remote dir, drop MDT1 req reply, fail MDT1 ========================================================== 16:45:58 (1713300358) SKIP: replay-single test_81f needs >= 2 MDTs SKIP 81f (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81g: DNE: unlink remote dir, drop req reply, fail M0, then M1 ========================================================== 16:46:01 (1713300361) SKIP: replay-single test_81g needs >= 2 MDTs SKIP 81g (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81h: DNE: unlink remote dir, drop request reply, fail 2 MDTs ========================================================== 16:46:05 (1713300365) SKIP: replay-single test_81h needs >= 2 MDTs SKIP 81h (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 84a: stale open during export disconnect ========================================================== 16:46:08 (1713300368) fail_loc=0x80000144 total: 1 open/close in 0.01 seconds: 190.73 ops/second PASS 84a (4s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 85a: check the cancellation of unused locks during recovery(IBITS) ========================================================== 16:46:14 (1713300374) before recovery: unused locks count = 201 Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:46:18 (1713300378) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:46:31 (1713300391) targets are mounted 16:46:31 (1713300391) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec after recovery: unused locks count = 101 PASS 85a (23s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 85b: check the cancellation of unused locks during recovery(EXTENT) ========================================================== 16:46:39 (1713300399) before recovery: unused locks count = 100 Failing ost1 on oleg130-server Stopping /mnt/lustre-ost1 (opts:) on oleg130-server 16:46:45 (1713300405) shut down Failover ost1 to oleg130-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-OST0000 16:46:57 (1713300417) targets are mounted 16:46:57 (1713300417) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec after recovery: unused locks count = 0 PASS 85b (23s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 86: umount server after clear nid_stats should not hit LBUG ========================================================== 16:47:05 (1713300425) Stopping clients: oleg130-client.virtnet /mnt/lustre (opts:) Stopping client oleg130-client.virtnet /mnt/lustre opts: mdt.lustre-MDT0000.exports.clear=0 Stopping /mnt/lustre-mds1 (opts:) on oleg130-server Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 Starting client oleg130-client.virtnet: -o user_xattr,flock oleg130-server@tcp:/lustre /mnt/lustre Started clients oleg130-client.virtnet: 192.168.201.130@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) PASS 86 (6s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 87a: write replay ================== 16:47:13 (1713300433) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre 8+0 records in 8+0 records out 8388608 bytes (8.4 MB) copied, 0.135651 s, 61.8 MB/s Failing ost1 on oleg130-server Stopping /mnt/lustre-ost1 (opts:) on oleg130-server 16:47:17 (1713300437) shut down Failover ost1 to oleg130-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-OST0000 16:47:30 (1713300450) targets are mounted 16:47:30 (1713300450) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec 8+0 records in 8+0 records out 8388608 bytes (8.4 MB) copied, 0.0369212 s, 227 MB/s PASS 87a (22s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 87b: write replay with changed data (checksum resend) ========================================================== 16:47:37 (1713300457) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 12288 3757056 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 16384 7522304 1% /mnt/lustre 8+0 records in 8+0 records out 8388608 bytes (8.4 MB) copied, 0.104087 s, 80.6 MB/s 8+0 records in 8+0 records out 8 bytes (8 B) copied, 0.00270759 s, 3.0 kB/s Failing ost1 on oleg130-server Stopping /mnt/lustre-ost1 (opts:) on oleg130-server 16:47:41 (1713300461) shut down Failover ost1 to oleg130-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-OST0000 16:47:55 (1713300475) targets are mounted 16:47:55 (1713300475) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec 0+1 records in 0+1 records out 72 bytes (72 B) copied, 0.00276054 s, 26.1 kB/s PASS 87b (23s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 88: MDS should not assign same objid to different files ========================================================== 16:48:03 (1713300483) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 13312 3756032 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 17408 7521280 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 13312 3756032 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 17408 7521280 1% /mnt/lustre before test: last_id = 11905, next_id = 11876 Creating to objid 11905 on ost lustre-OST0000... total: 31 open/close in 0.07 seconds: 447.49 ops/second total: 8 open/close in 0.02 seconds: 422.90 ops/second before recovery: last_id = 12001, next_id = 11914 Stopping /mnt/lustre-mds1 (opts:) on oleg130-server Stopping /mnt/lustre-ost1 (opts:) on oleg130-server Failover mds1 to oleg130-server Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 Failover ost1 to oleg130-server Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-OST0000 after recovery: last_id = 11945, next_id = 11914 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0458165 s, 11.4 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0438993 s, 11.9 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0374092 s, 14.0 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0327028 s, 16.0 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0471186 s, 11.1 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0393412 s, 13.3 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.050464 s, 10.4 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0400059 s, 13.1 MB/s -rw-r--r-- 1 root root 0 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11876 -rw-r--r-- 1 root root 0 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11877 -rw-r--r-- 1 root root 0 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11878 -rw-r--r-- 1 root root 0 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11879 -rw-r--r-- 1 root root 0 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11880 -rw-r--r-- 1 root root 0 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11881 -rw-r--r-- 1 root root 0 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11882 -rw-r--r-- 1 root root 0 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11883 -rw-r--r-- 1 root root 0 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11884 -rw-r--r-- 1 root root 0 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11885 -rw-r--r-- 1 root root 0 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11886 -rw-r--r-- 1 root root 0 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11887 -rw-r--r-- 1 root root 0 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11888 -rw-r--r-- 1 root root 0 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11889 -rw-r--r-- 1 root root 0 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11890 -rw-r--r-- 1 root root 0 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11891 -rw-r--r-- 1 root root 0 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11892 -rw-r--r-- 1 root root 0 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11893 -rw-r--r-- 1 root root 0 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11894 -rw-r--r-- 1 root root 0 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11895 -rw-r--r-- 1 root root 0 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11896 -rw-r--r-- 1 root root 0 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11897 -rw-r--r-- 1 root root 0 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11898 -rw-r--r-- 1 root root 0 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11899 -rw-r--r-- 1 root root 0 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11900 -rw-r--r-- 1 root root 0 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11901 -rw-r--r-- 1 root root 0 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11902 -rw-r--r-- 1 root root 0 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11903 -rw-r--r-- 1 root root 0 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11904 -rw-r--r-- 1 root root 0 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11905 -rw-r--r-- 1 root root 0 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11906 -rw-r--r-- 1 root root 0 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11907 -rw-r--r-- 1 root root 0 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11908 -rw-r--r-- 1 root root 0 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11909 -rw-r--r-- 1 root root 0 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11910 -rw-r--r-- 1 root root 0 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11911 -rw-r--r-- 1 root root 0 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11912 -rw-r--r-- 1 root root 0 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11913 -rw-r--r-- 1 root root 524288 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11917 -rw-r--r-- 1 root root 524288 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11918 -rw-r--r-- 1 root root 524288 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11919 -rw-r--r-- 1 root root 524288 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11920 -rw-r--r-- 1 root root 524288 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11921 -rw-r--r-- 1 root root 524288 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11922 -rw-r--r-- 1 root root 524288 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11923 -rw-r--r-- 1 root root 524288 Apr 16 16:48 /mnt/lustre/d88.replay-single/f-11924 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0376875 s, 13.9 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.044287 s, 11.8 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0301909 s, 17.4 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0379999 s, 13.8 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0336498 s, 15.6 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0443652 s, 11.8 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0330588 s, 15.9 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0344693 s, 15.2 MB/s PASS 88 (52s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 89: no disk space leak on late ost connection ========================================================== 16:48:57 (1713300537) Waiting for orphan cleanup... osp.lustre-OST0000-osc-MDT0000.old_sync_processed osp.lustre-OST0001-osc-MDT0000.old_sync_processed wait 40 secs maximumly for oleg130-server mds-ost sync done. sleep 5 for ZFS zfs Waiting for MDT destroys to complete 10+0 records in 10+0 records out 41943040 bytes (42 MB) copied, 0.267655 s, 157 MB/s Stopping /mnt/lustre-ost1 (opts:) on oleg130-server Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:49:10 (1713300550) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:49:24 (1713300564) targets are mounted 16:49:24 (1713300564) facet_failover done Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-OST0000 Starting client: oleg130-client.virtnet: -o user_xattr,flock oleg130-server@tcp:/lustre /mnt/lustre osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 67 sec Waiting for orphan cleanup... osp.lustre-OST0000-osc-MDT0000.old_sync_processed osp.lustre-OST0001-osc-MDT0000.old_sync_processed wait 40 secs maximumly for oleg130-server mds-ost sync done. sleep 5 for ZFS zfs Waiting for MDT destroys to complete free_before: 7517184 free_after: 7517184 PASS 89 (113s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 90: lfs find identifies the missing striped file segments ========================================================== 16:50:52 (1713300652) Create the files Fail ost1 lustre-OST0000_UUID, display the list of affected files Stopping /mnt/lustre-ost1 (opts:) on oleg130-server General Query: lfs find /mnt/lustre/d90.replay-single /mnt/lustre/d90.replay-single /mnt/lustre/d90.replay-single/f0 /mnt/lustre/d90.replay-single/all /mnt/lustre/d90.replay-single/f1 Querying files on shutdown ost1: lfs find --obd lustre-OST0000_UUID /mnt/lustre/d90.replay-single/f0 /mnt/lustre/d90.replay-single/all Check getstripe: /home/green/git/lustre-release/lustre/utils/lfs getstripe -r --obd lustre-OST0000_UUID /mnt/lustre/d90.replay-single/f0 lmm_stripe_count: 1 lmm_stripe_size: 4194304 lmm_pattern: raid0 lmm_layout_gen: 0 lmm_stripe_offset: 0 obdidx objid objid group 0 11947 0x2eab 0x240000400 * /mnt/lustre/d90.replay-single/all lmm_stripe_count: 2 lmm_stripe_size: 4194304 lmm_pattern: raid0 lmm_layout_gen: 0 lmm_stripe_offset: 1 obdidx objid objid group 0 11946 0x2eaa 0x240000400 * /mnt/lustre/d90.replay-single/all /mnt/lustre/d90.replay-single/f0 Failover ost1 to oleg130-server Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-OST0000 PASS 90 (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 93a: replay + reconnect ============ 16:51:14 (1713300674) 1+0 records in 1+0 records out 1024 bytes (1.0 kB) copied, 0.00195393 s, 524 kB/s fail_val=40 fail_loc=0x715 Failing ost1 on oleg130-server Stopping /mnt/lustre-ost1 (opts:) on oleg130-server 16:51:17 (1713300677) shut down Failover ost1 to oleg130-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-OST0000 16:51:31 (1713300691) targets are mounted 16:51:31 (1713300691) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec PASS 93a (59s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 93b: replay + reconnect on mds ===== 16:52:15 (1713300735) total: 20 open/close in 0.08 seconds: 266.64 ops/second fail_val=80 fail_loc=0x715 Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:52:17 (1713300737) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:52:30 (1713300750) targets are mounted 16:52:30 (1713300750) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 93b (100s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 100a: DNE: create striped dir, drop update rep from MDT1, fail MDT1 ========================================================== 16:53:58 (1713300838) SKIP: replay-single test_100a needs >= 2 MDTs SKIP 100a (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 100b: DNE: create striped dir, fail MDT0 ========================================================== 16:54:01 (1713300841) SKIP: replay-single test_100b needs >= 2 MDTs SKIP 100b (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 100c: DNE: create striped dir, abort_recov_mdt mds2 ========================================================== 16:54:04 (1713300844) SKIP: replay-single test_100c needs >= 2 MDTs SKIP 100c (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 100d: DNE: cancel update logs upon recovery abort ========================================================== 16:54:07 (1713300847) SKIP: replay-single test_100d needs > 1 MDTs SKIP 100d (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 100e: DNE: create striped dir on MDT0 and MDT1, fail MDT0, MDT1 ========================================================== 16:54:10 (1713300850) SKIP: replay-single test_100e needs >= 2 MDTs SKIP 100e (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 101: Shouldn't reassign precreated objs to other files after recovery ========================================================== 16:54:13 (1713300853) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3968 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 21504 3747840 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 25600 7513088 1% /mnt/lustre Stopping /mnt/lustre-mds1 (opts:) on oleg130-server Failover mds1 to oleg130-server Starting mds1: -o localrecov -o abort_recovery lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 PASS 101 (40s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 102a: check resend (request lost) with multiple modify RPCs in flight ========================================================== 16:54:55 (1713300895) creating 7 files ... fail_loc=0x159 launch 7 chmod in parallel (16:54:56) ... fail_loc=0 done (16:55:12) /mnt/lustre/d102a.replay-single/file-1 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-2 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-3 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-4 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-5 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-6 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-7 has perms 0600 OK PASS 102a (18s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 102b: check resend (reply lost) with multiple modify RPCs in flight ========================================================== 16:55:16 (1713300916) creating 7 files ... fail_loc=0x15a launch 7 chmod in parallel (16:55:17) ... fail_loc=0 done (16:55:33) /mnt/lustre/d102b.replay-single/file-1 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-2 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-3 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-4 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-5 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-6 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-7 has perms 0600 OK PASS 102b (18s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 102c: check replay w/o reconstruction with multiple mod RPCs in flight ========================================================== 16:55:36 (1713300936) creating 7 files ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 4096 2204544 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 21504 3333120 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3350528 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 25600 6683648 1% /mnt/lustre fail_loc=0x15a launch 7 chmod in parallel (16:55:39) ... fail_loc=0 Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:55:42 (1713300942) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:55:55 (1713300955) targets are mounted 16:55:55 (1713300955) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec done (16:55:59) /mnt/lustre/d102c.replay-single/file-1 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-2 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-3 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-4 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-5 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-6 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-7 has perms 0600 OK PASS 102c (25s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 102d: check replay & reconstruction with multiple mod RPCs in flight ========================================================== 16:56:03 (1713300963) creating 7 files ... fail_loc=0x15a launch 7 chmod in parallel (16:56:04) ... fail_loc=0 Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:56:07 (1713300967) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:56:20 (1713300980) targets are mounted 16:56:20 (1713300980) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec done (16:56:24) /mnt/lustre/d102d.replay-single/file-1 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-2 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-3 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-4 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-5 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-6 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-7 has perms 0600 OK PASS 102d (23s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 103: Check otr_next_id overflow ==== 16:56:28 (1713300988) fail_loc=0x80000162 total: 30 open/close in 0.14 seconds: 214.78 ops/second Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:56:31 (1713300991) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:56:43 (1713301003) targets are mounted 16:56:43 (1713301003) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 103 (21s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 110a: DNE: create striped dir, fail MDT1 ========================================================== 16:56:51 (1713301011) SKIP: replay-single test_110a needs >= 2 MDTs SKIP 110a (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 110b: DNE: create striped dir, fail MDT1 and client ========================================================== 16:56:54 (1713301014) SKIP: replay-single test_110b needs >= 2 MDTs SKIP 110b (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 110c: DNE: create striped dir, fail MDT2 ========================================================== 16:56:58 (1713301018) SKIP: replay-single test_110c needs >= 2 MDTs SKIP 110c (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 110d: DNE: create striped dir, fail MDT2 and client ========================================================== 16:57:01 (1713301021) SKIP: replay-single test_110d needs >= 2 MDTs SKIP 110d (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 110e: DNE: create striped dir, uncommit on MDT2, fail client/MDT1/MDT2 ========================================================== 16:57:04 (1713301024) SKIP: replay-single test_110e needs >= 2 MDTs SKIP 110e (1s) debug_raw_pointers=0 debug_raw_pointers=0 SKIP: replay-single test_110f skipping excluded test 110f debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 110g: DNE: create striped dir, uncommit on MDT1, fail client/MDT1/MDT2 ========================================================== 16:57:08 (1713301028) SKIP: replay-single test_110g needs >= 2 MDTs SKIP 110g (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 111a: DNE: unlink striped dir, fail MDT1 ========================================================== 16:57:11 (1713301031) SKIP: replay-single test_111a needs >= 2 MDTs SKIP 111a (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 111b: DNE: unlink striped dir, fail MDT2 ========================================================== 16:57:15 (1713301035) SKIP: replay-single test_111b needs >= 2 MDTs SKIP 111b (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 111c: DNE: unlink striped dir, uncommit on MDT1, fail client/MDT1/MDT2 ========================================================== 16:57:18 (1713301038) SKIP: replay-single test_111c needs >= 2 MDTs SKIP 111c (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 111d: DNE: unlink striped dir, uncommit on MDT2, fail client/MDT1/MDT2 ========================================================== 16:57:22 (1713301042) SKIP: replay-single test_111d needs >= 2 MDTs SKIP 111d (2s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 111e: DNE: unlink striped dir, uncommit on MDT2, fail MDT1/MDT2 ========================================================== 16:57:25 (1713301045) SKIP: replay-single test_111e needs >= 2 MDTs SKIP 111e (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 111f: DNE: unlink striped dir, uncommit on MDT1, fail MDT1/MDT2 ========================================================== 16:57:28 (1713301048) SKIP: replay-single test_111f needs >= 2 MDTs SKIP 111f (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 111g: DNE: unlink striped dir, fail MDT1/MDT2 ========================================================== 16:57:32 (1713301052) SKIP: replay-single test_111g needs >= 2 MDTs SKIP 111g (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112a: DNE: cross MDT rename, fail MDT1 ========================================================== 16:57:35 (1713301055) SKIP: replay-single test_112a needs >= 4 MDTs SKIP 112a (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112b: DNE: cross MDT rename, fail MDT2 ========================================================== 16:57:38 (1713301058) SKIP: replay-single test_112b needs >= 4 MDTs SKIP 112b (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112c: DNE: cross MDT rename, fail MDT3 ========================================================== 16:57:41 (1713301061) SKIP: replay-single test_112c needs >= 4 MDTs SKIP 112c (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112d: DNE: cross MDT rename, fail MDT4 ========================================================== 16:57:44 (1713301064) SKIP: replay-single test_112d needs >= 4 MDTs SKIP 112d (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112e: DNE: cross MDT rename, fail MDT1 and MDT2 ========================================================== 16:57:47 (1713301067) SKIP: replay-single test_112e needs >= 4 MDTs SKIP 112e (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112f: DNE: cross MDT rename, fail MDT1 and MDT3 ========================================================== 16:57:51 (1713301071) SKIP: replay-single test_112f needs >= 4 MDTs SKIP 112f (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112g: DNE: cross MDT rename, fail MDT1 and MDT4 ========================================================== 16:57:54 (1713301074) SKIP: replay-single test_112g needs >= 4 MDTs SKIP 112g (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112h: DNE: cross MDT rename, fail MDT2 and MDT3 ========================================================== 16:57:57 (1713301077) SKIP: replay-single test_112h needs >= 4 MDTs SKIP 112h (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112i: DNE: cross MDT rename, fail MDT2 and MDT4 ========================================================== 16:58:00 (1713301080) SKIP: replay-single test_112i needs >= 4 MDTs SKIP 112i (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112j: DNE: cross MDT rename, fail MDT3 and MDT4 ========================================================== 16:58:03 (1713301083) SKIP: replay-single test_112j needs >= 4 MDTs SKIP 112j (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112k: DNE: cross MDT rename, fail MDT1,MDT2,MDT3 ========================================================== 16:58:06 (1713301086) SKIP: replay-single test_112k needs >= 4 MDTs SKIP 112k (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112l: DNE: cross MDT rename, fail MDT1,MDT2,MDT4 ========================================================== 16:58:09 (1713301089) SKIP: replay-single test_112l needs >= 4 MDTs SKIP 112l (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112m: DNE: cross MDT rename, fail MDT1,MDT3,MDT4 ========================================================== 16:58:13 (1713301093) SKIP: replay-single test_112m needs >= 4 MDTs SKIP 112m (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112n: DNE: cross MDT rename, fail MDT2,MDT3,MDT4 ========================================================== 16:58:16 (1713301096) SKIP: replay-single test_112n needs >= 4 MDTs SKIP 112n (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 115: failover for create/unlink striped directory ========================================================== 16:58:19 (1713301099) SKIP: replay-single test_115 needs >= 2 MDTs SKIP 115 (2s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 116a: large update log master MDT recovery ========================================================== 16:58:21 (1713301101) SKIP: replay-single test_116a needs >= 2 MDTs SKIP 116a (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 116b: large update log slave MDT recovery ========================================================== 16:58:24 (1713301104) SKIP: replay-single test_116b needs >= 2 MDTs SKIP 116b (0s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 117: DNE: cross MDT unlink, fail MDT1 and MDT2 ========================================================== 16:58:27 (1713301107) SKIP: replay-single test_117 needs >= 4 MDTs SKIP 117 (2s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 118: invalidate osp update will not cause update log corruption ========================================================== 16:58:30 (1713301110) SKIP: replay-single test_118 needs >= 2 MDTs SKIP 118 (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 119: timeout of normal replay does not cause DNE replay fails ========================================================== 16:58:33 (1713301113) SKIP: replay-single test_119 needs >= 2 MDTs SKIP 119 (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 120: DNE fail abort should stop both normal and DNE replay ========================================================== 16:58:35 (1713301115) SKIP: replay-single test_120 needs >= 2 MDTs SKIP 120 (2s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 121: lock replay timed out and race ========================================================== 16:58:39 (1713301119) multiop /mnt/lustre/f121.replay-single vs_s TMPPIPE=/tmp/multiop_open_wait_pipe.6919 Stopping /mnt/lustre-mds1 (opts:) on oleg130-server Failover mds1 to oleg130-server fail_loc=0x721 fail_val=0 at_max=0 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 fail_loc=0x0 at_max=600 PASS 121 (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 130a: DoM file create (setstripe) replay ========================================================== 16:59:07 (1713301147) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3840 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 21504 3333120 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 25600 7098368 1% /mnt/lustre Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:59:11 (1713301151) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:59:24 (1713301164) targets are mounted 16:59:24 (1713301164) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 130a (22s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 130b: DoM file create (inherited) replay ========================================================== 16:59:32 (1713301172) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3968 2204544 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 21504 3333120 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 25600 7098368 1% /mnt/lustre Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:59:35 (1713301175) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 16:59:48 (1713301188) targets are mounted 16:59:48 (1713301188) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 130b (21s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 131a: DoM file write lock replay === 16:59:56 (1713301196) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3968 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 21504 3333120 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 25600 7098368 1% /mnt/lustre 1+0 records in 1+0 records out 8 bytes (8 B) copied, 0.00240565 s, 3.3 kB/s Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 16:59:59 (1713301199) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 17:00:13 (1713301213) targets are mounted 17:00:13 (1713301213) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 131a (22s) debug_raw_pointers=0 debug_raw_pointers=0 SKIP: replay-single test_131b skipping excluded test 131b debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 132a: PFL new component instantiate replay ========================================================== 17:00:21 (1713301221) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3968 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 21504 3333120 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 25600 7098368 1% /mnt/lustre 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0310991 s, 33.7 MB/s /mnt/lustre/f132a.replay-single lcm_layout_gen: 3 lcm_mirror_count: 1 lcm_entry_count: 2 lcme_id: 1 lcme_mirror_id: 0 lcme_flags: init lcme_extent.e_start: 0 lcme_extent.e_end: 1048576 lmm_stripe_count: 1 lmm_stripe_size: 1048576 lmm_pattern: raid0 lmm_layout_gen: 0 lmm_stripe_offset: 0 lmm_objects: - 0: { l_ost_idx: 0, l_fid: [0x240000400:0x33ca:0x0] } lcme_id: 2 lcme_mirror_id: 0 lcme_flags: init lcme_extent.e_start: 1048576 lcme_extent.e_end: EOF lmm_stripe_count: 2 lmm_stripe_size: 4194304 lmm_pattern: raid0 lmm_layout_gen: 0 lmm_stripe_offset: 1 lmm_objects: - 0: { l_ost_idx: 1, l_fid: [0x280000400:0x32e2:0x0] } - 1: { l_ost_idx: 0, l_fid: [0x240000400:0x33cb:0x0] } Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 17:00:25 (1713301225) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 17:00:38 (1713301238) targets are mounted 17:00:38 (1713301238) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f132a.replay-single lcm_layout_gen: 1 lcm_mirror_count: 1 lcm_entry_count: 2 lcme_id: 1 lcme_mirror_id: 0 lcme_flags: init lcme_extent.e_start: 0 lcme_extent.e_end: 1048576 lmm_stripe_count: 1 lmm_stripe_size: 1048576 lmm_pattern: raid0 lmm_layout_gen: 65535 lmm_stripe_offset: 0 lmm_objects: - 0: { l_ost_idx: 0, l_fid: [0x240000400:0x33ca:0x0] } lcme_id: 2 lcme_mirror_id: 0 lcme_flags: init lcme_extent.e_start: 1048576 lcme_extent.e_end: EOF lmm_stripe_count: 2 lmm_stripe_size: 4194304 lmm_pattern: raid0 lmm_layout_gen: 65535 lmm_stripe_offset: 1 lmm_objects: - 0: { l_ost_idx: 1, l_fid: [0x280000400:0x32e2:0x0] } - 1: { l_ost_idx: 0, l_fid: [0x240000400:0x33cb:0x0] } PASS 132a (23s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 133: check resend of ongoing requests for lwp during failover ========================================================== 17:00:47 (1713301247) SKIP: replay-single test_133 needs >= 2 MDTs SKIP 133 (2s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 134: replay creation of a file created in a pool ========================================================== 17:00:50 (1713301250) Creating new pool oleg130-server: Pool lustre.pool_134 created Adding targets to pool oleg130-server: OST lustre-OST0001_UUID added to pool lustre.pool_134 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 4096 2204544 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 21504 3747840 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 5120 3764224 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 26624 7512064 1% /mnt/lustre Failing mds1 on oleg130-server Stopping /mnt/lustre-mds1 (opts:) on oleg130-server 17:00:59 (1713301259) shut down Failover mds1 to oleg130-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-MDT0000 17:01:13 (1713301273) targets are mounted 17:01:13 (1713301273) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Destroy the created pools: pool_134 lustre.pool_134 oleg130-server: OST lustre-OST0001_UUID removed from pool lustre.pool_134 oleg130-server: Pool lustre.pool_134 destroyed PASS 134 (34s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 135: Server failure in lock replay phase ========================================================== 17:01:26 (1713301286) Failing ost1 on oleg130-server Stopping /mnt/lustre-ost1 (opts:) on oleg130-server 17:01:28 (1713301288) shut down Failover ost1 to oleg130-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-OST0000 17:01:42 (1713301302) targets are mounted 17:01:42 (1713301302) facet_failover done oleg130-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 4096 2204544 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 21504 3747840 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 5120 3764224 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 26624 7512064 1% /mnt/lustre ldlm.cancel_unused_locks_before_replay=0 Stopping /mnt/lustre-ost1 (opts:) on oleg130-server Failover ost1 to oleg130-server oleg130-server: oleg130-server.virtnet: executing load_module ../libcfs/libcfs/libcfs fail_loc=0x32d fail_val=20 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-OST0000 oleg130-client.virtnet: executing wait_import_state_mount REPLAY_LOCKS osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in REPLAY_LOCKS state after 0 sec Stopping /mnt/lustre-ost1 (opts:) on oleg130-server Failover ost1 to oleg130-server oleg130-server: oleg130-server.virtnet: executing load_module ../libcfs/libcfs/libcfs fail_loc=0 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 End of sync Started lustre-OST0000 Stopping /mnt/lustre-ost1 (opts:-f) on oleg130-server Stopping /mnt/lustre-ost2 (opts:-f) on oleg130-server Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-OST0000 Starting ost2: -o localrecov lustre-ost2/ost2 /mnt/lustre-ost2 seq.cli-lustre-OST0001-super.width=65536 oleg130-server: oleg130-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg130-client: oleg130-server: ssh exited with exit code 1 Started lustre-OST0001 ldlm.cancel_unused_locks_before_replay=1 PASS 135 (72s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 136: MDS to disconnect all OSPs first, then cleanup ldlm ========================================================== 17:02:40 (1713301360) SKIP: replay-single test_136 needs > 2 MDTs SKIP 136 (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 200: Dropping one OBD_PING should not cause disconnect ========================================================== 17:02:43 (1713301363) SKIP: replay-single test_200 Need remote client SKIP 200 (1s) debug_raw_pointers=0 debug_raw_pointers=0 == replay-single test complete, duration 4931 sec ======== 17:02:45 (1713301365) === replay-single: start cleanup 17:02:45 (1713301365) === === replay-single: finish cleanup 17:02:48 (1713301368) ===