-----============= acceptance-small: replay-single ============----- Thu Apr 18 06:48:03 EDT 2024 excepting tests: 110f 131b 59 36 === replay-single: start setup 06:48:06 (1713437286) === oleg308-client.virtnet: executing check_config_client /mnt/lustre oleg308-client.virtnet: Checking config lustre mounted on /mnt/lustre Checking servers environments Checking clients oleg308-client.virtnet environments Using TIMEOUT=20 osc.lustre-OST0000-osc-ffff88012a61b800.idle_timeout=debug osc.lustre-OST0001-osc-ffff88012a61b800.idle_timeout=debug disable quota as required oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all === replay-single: finish setup 06:48:12 (1713437292) === debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 0a: empty replay =================== 06:48:13 (1713437293) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 06:48:16 (1713437296) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 06:48:28 (1713437308) targets are mounted 06:48:28 (1713437308) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 0a (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 0b: ensure object created after recover exists. (3284) ========================================================== 06:48:33 (1713437313) Failing ost1 on oleg308-server Stopping /mnt/lustre-ost1 (opts:) on oleg308-server 06:48:35 (1713437315) shut down Failover ost1 to oleg308-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-OST0000 06:48:47 (1713437327) targets are mounted 06:48:47 (1713437327) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec total: 20 open/close in 0.07 seconds: 295.12 ops/second - unlinked 0 (time 1713437331 ; total 0 ; last 0) total: 20 unlinks in 0 seconds: inf unlinks/second PASS 0b (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 0c: check replay-barrier =========== 06:48:53 (1713437333) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 06:48:56 (1713437336) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 06:49:08 (1713437348) targets are mounted 06:49:08 (1713437348) facet_failover done Starting client: oleg308-client.virtnet: -o user_xattr,flock oleg308-server@tcp:/lustre /mnt/lustre rm: cannot remove '/mnt/lustre/f0c.replay-single': No such file or directory PASS 0c (77s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 0d: expired recovery with no clients ========================================================== 06:50:12 (1713437412) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 06:50:14 (1713437414) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 06:50:27 (1713437427) targets are mounted 06:50:27 (1713437427) facet_failover done Starting client: oleg308-client.virtnet: -o user_xattr,flock oleg308-server@tcp:/lustre /mnt/lustre PASS 0d (77s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 1: simple create =================== 06:51:31 (1713437491) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 06:51:33 (1713437493) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 06:51:46 (1713437506) targets are mounted 06:51:46 (1713437506) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f1.replay-single has type file OK PASS 1 (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 2a: touch ========================== 06:51:52 (1713437512) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 06:51:54 (1713437514) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 06:52:06 (1713437526) targets are mounted 06:52:06 (1713437526) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f2a.replay-single has type file OK PASS 2a (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 2b: touch ========================== 06:52:12 (1713437532) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 06:52:15 (1713437535) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 06:52:27 (1713437547) targets are mounted 06:52:27 (1713437547) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f2b.replay-single has type file OK PASS 2b (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 2c: setstripe replay =============== 06:52:33 (1713437553) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 06:52:36 (1713437556) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 06:52:48 (1713437568) targets are mounted 06:52:48 (1713437568) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f2c.replay-single has type file OK PASS 2c (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 2d: setdirstripe replay ============ 06:52:54 (1713437574) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 06:52:57 (1713437577) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 06:53:09 (1713437589) targets are mounted 06:53:09 (1713437589) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d2d.replay-single has type dir OK PASS 2d (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 2e: O_CREAT|O_EXCL create replay === 06:53:15 (1713437595) fail_loc=0x8000013b UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 06:53:19 (1713437599) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 06:53:31 (1713437611) targets are mounted 06:53:31 (1713437611) facet_failover done Succeed in opening file "/mnt/lustre/f2e.replay-single"(flags=O_CREAT) oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f2e.replay-single has type file OK PASS 2e (21s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 3a: replay failed open(O_DIRECTORY) ========================================================== 06:53:37 (1713437617) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Error in opening file "/mnt/lustre/f3a.replay-single"(flags=O_DIRECTORY) 20: Not a directory Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 06:53:40 (1713437620) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 06:53:52 (1713437632) targets are mounted 06:53:52 (1713437632) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f3a.replay-single has type file OK PASS 3a (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 3b: replay failed open -ENOMEM ===== 06:53:58 (1713437638) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre fail_loc=0x80000114 touch: cannot touch '/mnt/lustre/f3b.replay-single': Cannot allocate memory fail_loc=0 Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 06:54:01 (1713437641) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 06:54:13 (1713437653) targets are mounted 06:54:13 (1713437653) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/f3b.replay-single: No such file or directory PASS 3b (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 3c: replay failed open -ENOMEM ===== 06:54:19 (1713437659) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre fail_loc=0x80000128 touch: cannot touch '/mnt/lustre/f3c.replay-single': Cannot allocate memory fail_loc=0 Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 06:54:22 (1713437662) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 06:54:34 (1713437674) targets are mounted 06:54:34 (1713437674) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/f3c.replay-single: No such file or directory PASS 3c (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 4a: |x| 10 open(O_CREAT)s ========== 06:54:40 (1713437680) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 06:54:42 (1713437682) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 06:54:55 (1713437695) targets are mounted 06:54:55 (1713437695) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 4a (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 4b: |x| rm 10 files ================ 06:55:01 (1713437701) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3761152 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3761152 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7522304 1% /mnt/lustre Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 06:55:03 (1713437703) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 06:55:16 (1713437716) targets are mounted 06:55:16 (1713437716) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/f4b.replay-single-*: No such file or directory PASS 4b (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 5: |x| 220 open(O_CREAT) =========== 06:55:22 (1713437722) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 06:55:26 (1713437726) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 06:55:38 (1713437738) targets are mounted 06:55:38 (1713437738) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 5 (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 6a: mkdir + contained create ======= 06:55:50 (1713437750) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 06:55:52 (1713437752) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 06:56:05 (1713437765) targets are mounted 06:56:05 (1713437765) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d6a.replay-single has type dir OK /mnt/lustre/d6a.replay-single/f6a.replay-single has type file OK PASS 6a (21s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 6b: |X| rmdir ====================== 06:56:13 (1713437773) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 06:56:15 (1713437775) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 06:56:28 (1713437788) targets are mounted 06:56:28 (1713437788) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/d6b.replay-single: No such file or directory PASS 6b (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 7: mkdir |X| contained create ====== 06:56:34 (1713437794) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 06:56:37 (1713437797) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 06:56:50 (1713437810) targets are mounted 06:56:50 (1713437810) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d7.replay-single has type dir OK /mnt/lustre/d7.replay-single/f7.replay-single has type file OK PASS 7 (21s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 8: creat open |X| close ============ 06:56:57 (1713437817) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre multiop /mnt/lustre/f8.replay-single vmo_c TMPPIPE=/tmp/multiop_open_wait_pipe.6922 Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 06:56:59 (1713437819) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 06:57:13 (1713437833) targets are mounted 06:57:13 (1713437833) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f8.replay-single /mnt/lustre/f8.replay-single has type file OK PASS 8 (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 9: |X| create (same inum/gen) ====== 06:57:19 (1713437839) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 06:57:22 (1713437842) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 06:57:35 (1713437855) targets are mounted 06:57:36 (1713437856) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec old_inum == 144115305935798546, new_inum == 144115305935798546 old_inum and new_inum match PASS 9 (21s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 10: create |X| rename unlink ======= 06:57:42 (1713437862) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 06:57:44 (1713437864) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 06:57:57 (1713437877) targets are mounted 06:57:57 (1713437877) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/f10.replay-single: No such file or directory PASS 10 (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 11: create open write rename |X| create-old-name read ========================================================== 06:58:03 (1713437883) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre new old Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 06:58:05 (1713437885) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 06:58:18 (1713437898) targets are mounted 06:58:18 (1713437898) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec new old PASS 11 (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 12: open, unlink |X| close ========= 06:58:24 (1713437904) multiop /mnt/lustre/f12.replay-single vo_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6922 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 06:58:26 (1713437906) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 06:58:39 (1713437919) targets are mounted 06:58:39 (1713437919) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 12 (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 13: open chmod 0 |x| write close === 06:58:46 (1713437926) multiop /mnt/lustre/f13.replay-single vO_wc TMPPIPE=/tmp/multiop_open_wait_pipe.6922 /mnt/lustre/f13.replay-single has perms 00 OK UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 06:58:49 (1713437929) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 06:59:01 (1713437941) targets are mounted 06:59:01 (1713437941) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f13.replay-single has perms 00 OK /mnt/lustre/f13.replay-single has size 1 OK PASS 13 (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 14: open(O_CREAT), unlink |X| close ========================================================== 06:59:07 (1713437947) multiop /mnt/lustre/f14.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6922 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 06:59:10 (1713437950) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 06:59:23 (1713437963) targets are mounted 06:59:23 (1713437963) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 14 (21s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 15: open(O_CREAT), unlink |X| touch new, close ========================================================== 06:59:30 (1713437970) multiop /mnt/lustre/f15.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6922 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 06:59:32 (1713437972) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 06:59:45 (1713437985) targets are mounted 06:59:45 (1713437985) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 15 (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 16: |X| open(O_CREAT), unlink, touch new, unlink new ========================================================== 06:59:51 (1713437991) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 06:59:53 (1713437993) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:00:07 (1713438007) targets are mounted 07:00:07 (1713438007) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 16 (21s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 17: |X| open(O_CREAT), |replay| close ========================================================== 07:00:14 (1713438014) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre multiop /mnt/lustre/f17.replay-single vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.6922 Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:00:16 (1713438016) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:00:29 (1713438029) targets are mounted 07:00:29 (1713438029) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f17.replay-single has type file OK PASS 17 (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 18: open(O_CREAT), unlink, touch new, close, touch, unlink ========================================================== 07:00:36 (1713438036) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre multiop /mnt/lustre/f18.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6922 pid: 21678 will close Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:00:39 (1713438039) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:00:52 (1713438052) targets are mounted 07:00:52 (1713438052) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 18 (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 19: mcreate, open, write, rename === 07:00:58 (1713438058) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre old Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:01:01 (1713438061) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:01:14 (1713438074) targets are mounted 07:01:14 (1713438074) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec old PASS 19 (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 20a: |X| open(O_CREAT), unlink, replay, close (test mds_cleanup_orphans) ========================================================== 07:01:20 (1713438080) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre multiop /mnt/lustre/f20a.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6922 Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:01:23 (1713438083) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:01:36 (1713438096) targets are mounted 07:01:36 (1713438096) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 20a (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 20b: write, unlink, eviction, replay (test mds_cleanup_orphans) ========================================================== 07:01:42 (1713438102) /mnt/lustre/f20b.replay-single lmm_stripe_count: 1 lmm_stripe_size: 4194304 lmm_pattern: raid0 lmm_layout_gen: 0 lmm_stripe_offset: 0 obdidx objid objid group 0 1090 0x442 0x240000400 10000+0 records in 10000+0 records out 40960000 bytes (41 MB) copied, 1.28981 s, 31.8 MB/s Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:01:46 (1713438106) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:01:59 (1713438119) targets are mounted 07:01:59 (1713438119) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec affected facets: mds1 oleg308-server: oleg308-server.virtnet: executing _wait_recovery_complete *.lustre-MDT0000.recovery_status 1475 oleg308-server: *.lustre-MDT0000.recovery_status status: COMPLETE sleep 5 for ZFS zfs Waiting for MDT destroys to complete before 6144, after 6144 PASS 20b (29s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 20c: check that client eviction does not affect file content ========================================================== 07:02:12 (1713438132) multiop /mnt/lustre/f20c.replay-single vOw_c TMPPIPE=/tmp/multiop_open_wait_pipe.6922 -rw-r--r-- 1 root root 1 Apr 18 07:02 /mnt/lustre/f20c.replay-single PASS 20c (3s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 21: |X| open(O_CREAT), unlink touch new, replay, close (test mds_cleanup_orphans) ========================================================== 07:02:17 (1713438137) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3764224 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7530496 1% /mnt/lustre multiop /mnt/lustre/f21.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6922 Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:02:20 (1713438140) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:02:32 (1713438152) targets are mounted 07:02:32 (1713438152) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 21 (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 22: open(O_CREAT), |X| unlink, replay, close (test mds_cleanup_orphans) ========================================================== 07:02:38 (1713438158) multiop /mnt/lustre/f22.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6922 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3764224 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7530496 1% /mnt/lustre Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:02:41 (1713438161) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:02:54 (1713438174) targets are mounted 07:02:54 (1713438174) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 22 (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 23: open(O_CREAT), |X| unlink touch new, replay, close (test mds_cleanup_orphans) ========================================================== 07:03:00 (1713438180) multiop /mnt/lustre/f23.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6922 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3764224 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7530496 1% /mnt/lustre Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:03:03 (1713438183) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:03:15 (1713438195) targets are mounted 07:03:15 (1713438195) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 23 (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 24: open(O_CREAT), replay, unlink, close (test mds_cleanup_orphans) ========================================================== 07:03:22 (1713438202) multiop /mnt/lustre/f24.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6922 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3764224 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7530496 1% /mnt/lustre Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:03:24 (1713438204) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:03:37 (1713438217) targets are mounted 07:03:37 (1713438217) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 24 (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 25: open(O_CREAT), unlink, replay, close (test mds_cleanup_orphans) ========================================================== 07:03:44 (1713438224) multiop /mnt/lustre/f25.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6922 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3764224 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7530496 1% /mnt/lustre Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:03:47 (1713438227) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:04:01 (1713438241) targets are mounted 07:04:01 (1713438241) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 25 (21s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 26: |X| open(O_CREAT), unlink two, close one, replay, close one (test mds_cleanup_orphans) ========================================================== 07:04:07 (1713438247) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3764224 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7530496 1% /mnt/lustre multiop /mnt/lustre/f26.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6922 multiop /mnt/lustre/f26.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6922 Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:04:10 (1713438250) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:04:23 (1713438263) targets are mounted 07:04:23 (1713438263) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 26 (21s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 27: |X| open(O_CREAT), unlink two, replay, close two (test mds_cleanup_orphans) ========================================================== 07:04:29 (1713438269) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3764224 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7530496 1% /mnt/lustre multiop /mnt/lustre/f27.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6922 multiop /mnt/lustre/f27.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6922 Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:04:32 (1713438272) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:04:45 (1713438285) targets are mounted 07:04:45 (1713438285) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 27 (21s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 28: open(O_CREAT), |X| unlink two, close one, replay, close one (test mds_cleanup_orphans) ========================================================== 07:04:52 (1713438292) multiop /mnt/lustre/f28.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6922 multiop /mnt/lustre/f28.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6922 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3764224 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7530496 1% /mnt/lustre Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:04:55 (1713438295) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:05:07 (1713438307) targets are mounted 07:05:07 (1713438307) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 28 (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 29: open(O_CREAT), |X| unlink two, replay, close two (test mds_cleanup_orphans) ========================================================== 07:05:13 (1713438313) multiop /mnt/lustre/f29.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6922 multiop /mnt/lustre/f29.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6922 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3764224 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7530496 1% /mnt/lustre Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:05:16 (1713438316) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:05:29 (1713438329) targets are mounted 07:05:29 (1713438329) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 29 (21s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 30: open(O_CREAT) two, unlink two, replay, close two (test mds_cleanup_orphans) ========================================================== 07:05:35 (1713438335) multiop /mnt/lustre/f30.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6922 multiop /mnt/lustre/f30.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6922 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3764224 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7530496 1% /mnt/lustre Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:05:38 (1713438338) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:05:51 (1713438351) targets are mounted 07:05:51 (1713438351) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 30 (21s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 31: open(O_CREAT) two, unlink one, |X| unlink one, close two (test mds_cleanup_orphans) ========================================================== 07:05:57 (1713438357) multiop /mnt/lustre/f31.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6922 multiop /mnt/lustre/f31.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6922 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3764224 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7530496 1% /mnt/lustre Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:06:00 (1713438360) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:06:13 (1713438373) targets are mounted 07:06:13 (1713438373) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 31 (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 32: close() notices client eviction; close() after client eviction ========================================================== 07:06:19 (1713438379) multiop /mnt/lustre/f32.replay-single vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.6922 multiop /mnt/lustre/f32.replay-single vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.6922 pdsh@oleg308-client: oleg308-client: ssh exited with exit code 5 PASS 32 (4s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 33a: fid seq shouldn't be reused after abort recovery ========================================================== 07:06:25 (1713438385) total: 10 open/close in 0.06 seconds: 159.70 ops/second Replay barrier on lustre-MDT0000 Stopping /mnt/lustre-mds1 (opts:) on oleg308-server Failover mds1 to oleg308-server Starting mds1: -o localrecov -o abort_recovery lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg308-client: error: invalid path '/mnt/lustre': Input/output error pdsh@oleg308-client: oleg308-client: ssh exited with exit code 5 first stat failed: 5 total: 10 open/close in 0.04 seconds: 233.60 ops/second PASS 33a (9s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 33b: test fid seq allocation ======= 07:06:36 (1713438396) fail_loc=0x1311 total: 10 open/close in 0.04 seconds: 247.60 ops/second Replay barrier on lustre-MDT0000 Stopping /mnt/lustre-mds1 (opts:) on oleg308-server Failover mds1 to oleg308-server Starting mds1: -o localrecov -o abort_recovery lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 total: 10 open/close in 0.07 seconds: 147.21 ops/second PASS 33b (11s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 34: abort recovery before client does replay (test mds_cleanup_orphans) ========================================================== 07:06:49 (1713438409) multiop /mnt/lustre/f34.replay-single vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.6922 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3764224 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7530496 1% /mnt/lustre Stopping /mnt/lustre-mds1 (opts:) on oleg308-server Failover mds1 to oleg308-server Starting mds1: -o localrecov -o abort_recovery lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg308-client: error: invalid path '/mnt/lustre': Input/output error pdsh@oleg308-client: oleg308-client: ssh exited with exit code 5 first stat failed: 5 PASS 34 (11s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 35: test recovery from llog for unlink op ========================================================== 07:07:02 (1713438422) fail_loc=0x80000119 Stopping /mnt/lustre-mds1 (opts:) on oleg308-server Failover mds1 to oleg308-server Starting mds1: -o localrecov -o abort_recovery lustre-mdt1/mdt1 /mnt/lustre-mds1 rm: cannot remove '/mnt/lustre/f35.replay-single': Input/output error oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 Can't lstat /mnt/lustre/f35.replay-single: No such file or directory PASS 35 (12s) debug_raw_pointers=0 debug_raw_pointers=0 SKIP: replay-single test_36 skipping ALWAYS excluded test 36 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 37: abort recovery before client does replay (test mds_cleanup_orphans for directories) ========================================================== 07:07:16 (1713438436) multiop /mnt/lustre/d37.replay-single/f37.replay-single vdD_c TMPPIPE=/tmp/multiop_open_wait_pipe.6922 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3456 2205184 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3764224 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7530496 1% /mnt/lustre Stopping /mnt/lustre-mds1 (opts:) on oleg308-server Failover mds1 to oleg308-server Starting mds1: -o localrecov -o abort_recovery lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 PASS 37 (11s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 38: test recovery from unlink llog (test llog_gen_rec) ========================================================== 07:07:30 (1713438450) total: 800 open/close in 2.99 seconds: 267.90 ops/second - unlinked 0 (time 1713438455 ; total 0 ; last 0) total: 400 unlinks in 2 seconds: 200.000000 unlinks/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3712 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3764224 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7530496 1% /mnt/lustre Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:07:40 (1713438460) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:07:52 (1713438472) targets are mounted 07:07:52 (1713438472) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec - unlinked 0 (time 1713438477 ; total 0 ; last 0) total: 400 unlinks in 1 seconds: 400.000000 unlinks/second Can't lstat /mnt/lustre/f38.replay-single-*: No such file or directory PASS 38 (32s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 39: test recovery from unlink llog (test llog_gen_rec) ========================================================== 07:08:04 (1713438484) total: 800 open/close in 3.16 seconds: 253.27 ops/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3764224 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7530496 1% /mnt/lustre - unlinked 0 (time 1713438491 ; total 0 ; last 0) total: 400 unlinks in 1 seconds: 400.000000 unlinks/second Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:08:14 (1713438494) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:08:27 (1713438507) targets are mounted 07:08:27 (1713438507) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec - unlinked 0 (time 1713438513 ; total 0 ; last 0) total: 400 unlinks in 2 seconds: 200.000000 unlinks/second Can't lstat /mnt/lustre/f39.replay-single-*: No such file or directory PASS 39 (35s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 41: read from a valid osc while other oscs are invalid ========================================================== 07:08:42 (1713438522) 1+0 records in 1+0 records out 4096 bytes (4.1 kB) copied, 0.00304001 s, 1.3 MB/s 1+0 records in 1+0 records out 4096 bytes (4.1 kB) copied, 0.00671167 s, 610 kB/s PASS 41 (3s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 42: recovery after ost failure ===== 07:08:47 (1713438527) total: 800 open/close in 3.96 seconds: 201.99 ops/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3712 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre - unlinked 0 (time 1713438534 ; total 0 ; last 0) total: 400 unlinks in 1 seconds: 400.000000 unlinks/second debug=-1 Failing ost1 on oleg308-server Stopping /mnt/lustre-ost1 (opts:) on oleg308-server 07:08:57 (1713438537) shut down Failover ost1 to oleg308-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-OST0000 07:09:10 (1713438550) targets are mounted 07:09:10 (1713438550) facet_failover done wait for MDS to timeout and recover debug=super ioctl neterror warning dlmtrace error emerg ha rpctrace vfstrace config console lfsck debug=super ioctl neterror warning dlmtrace error emerg ha rpctrace vfstrace config console lfsck - unlinked 0 (time 1713438592 ; total 0 ; last 0) total: 400 unlinks in 1 seconds: 400.000000 unlinks/second Can't lstat /mnt/lustre/f42.replay-single-*: No such file or directory PASS 42 (67s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 43: mds osc import failure during recovery; don't LBUG ========================================================== 07:09:56 (1713438596) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3712 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre fail_loc=0x80000204 Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:09:59 (1713438599) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:10:11 (1713438611) targets are mounted 07:10:11 (1713438611) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 43 (30s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 44a: race in target handle connect ========================================================== 07:10:27 (1713438627) at_max=40 1 of 10 (1713438628) service : cur 5 worst 5 (at 1713438612, 16s ago) 4 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre 2 of 10 (1713438634) service : cur 5 worst 5 (at 1713438612, 22s ago) 5 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre 3 of 10 (1713438639) service : cur 5 worst 5 (at 1713438612, 27s ago) 5 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre 4 of 10 (1713438645) service : cur 5 worst 5 (at 1713438612, 33s ago) 5 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre 5 of 10 (1713438650) service : cur 5 worst 5 (at 1713438612, 38s ago) 5 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre 6 of 10 (1713438656) service : cur 5 worst 5 (at 1713438612, 44s ago) 5 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre 7 of 10 (1713438661) service : cur 5 worst 5 (at 1713438612, 50s ago) 5 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre 8 of 10 (1713438667) service : cur 5 worst 5 (at 1713438612, 55s ago) 5 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre 9 of 10 (1713438672) service : cur 5 worst 5 (at 1713438612, 60s ago) 5 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre 10 of 10 (1713438678) service : cur 5 worst 5 (at 1713438612, 66s ago) 5 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre fail_loc=0 at_max=600 PASS 44a (58s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 44b: race in target handle connect ========================================================== 07:11:26 (1713438686) 1 of 10 (1713438687) service : cur 5 worst 5 (at 1713438612, 75s ago) 5 0 0 0 fail_loc=0x80000704 error: recover: Connection timed out Filesystem 1K-blocks Used Available Use% Mounted on 192.168.203.108@tcp:/lustre 7542784 6144 7532544 1% /mnt/lustre 2 of 10 (1713438727) service : cur 40 worst 40 (at 1713438727, 0s ago) 40 0 0 0 fail_loc=0x80000704 error: recover: Connection timed out df: '/mnt/lustre': Operation already in progress 3 of 10 (1713438748) service : cur 40 worst 40 (at 1713438727, 21s ago) 40 0 0 0 fail_loc=0x80000704 error: recover: Connection timed out df: '/mnt/lustre': Operation already in progress 4 of 10 (1713438768) service : cur 40 worst 40 (at 1713438727, 41s ago) 40 40 0 0 fail_loc=0x80000704 error: recover: Connection timed out df: '/mnt/lustre': Operation already in progress 5 of 10 (1713438789) service : cur 40 worst 40 (at 1713438727, 62s ago) 40 40 0 0 fail_loc=0x80000704 error: recover: Connection timed out df: '/mnt/lustre': Operation already in progress 6 of 10 (1713438809) service : cur 40 worst 40 (at 1713438727, 82s ago) 40 40 0 0 fail_loc=0x80000704 error: recover: Connection timed out df: '/mnt/lustre': Operation already in progress 7 of 10 (1713438830) service : cur 40 worst 40 (at 1713438727, 103s ago) 40 40 0 0 fail_loc=0x80000704 error: recover: Connection timed out df: '/mnt/lustre': Operation already in progress 8 of 10 (1713438850) service : cur 40 worst 40 (at 1713438727, 123s ago) 40 40 0 0 fail_loc=0x80000704 error: recover: Connection timed out df: '/mnt/lustre': Operation already in progress 9 of 10 (1713438871) service : cur 40 worst 40 (at 1713438727, 144s ago) 40 40 0 0 fail_loc=0x80000704 error: recover: Connection timed out df: '/mnt/lustre': Operation already in progress 10 of 10 (1713438891) service : cur 40 worst 40 (at 1713438727, 164s ago) 40 40 0 0 fail_loc=0x80000704 error: recover: Connection timed out df: '/mnt/lustre': Operation already in progress PASS 44b (227s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 44c: race in target handle connect ========================================================== 07:15:14 (1713438914) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre total: 100 create in 0.21 seconds: 485.08 ops/second fail_loc=0x80000712 Stopping /mnt/lustre-mds1 (opts:) on oleg308-server Failover mds1 to oleg308-server Starting mds1: -o localrecov -o abort_recovery lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg308-client: error: invalid path '/mnt/lustre': Input/output error pdsh@oleg308-client: oleg308-client: ssh exited with exit code 5 first stat failed: 5 unlink(/mnt/lustre/f44c.replay-single-0) error: No such file or directory total: 0 unlinks in 0 seconds: -nan unlinks/second Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:15:23 (1713438923) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:15:36 (1713438936) targets are mounted 07:15:36 (1713438936) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec unlink(/mnt/lustre/f44c.replay-single-0) error: No such file or directory total: 0 unlinks in 0 seconds: -nan unlinks/second PASS 44c (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 45: Handle failed close ============ 07:15:42 (1713438942) multiop /mnt/lustre/f45.replay-single vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.6922 /mnt/lustre/f45.replay-single has type file OK PASS 45 (2s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 46: Don't leak file handle after open resend (3325) ========================================================== 07:15:46 (1713438946) fail_loc=0x122 fail_loc=0 Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:16:04 (1713438964) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:16:16 (1713438976) targets are mounted 07:16:16 (1713438976) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec lfs path2fid: cannot get fid for 'f46.replay-single': No such file or directory PASS 46 (35s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 47: MDS->OSC failure during precreate cleanup (2824) ========================================================== 07:16:23 (1713438983) total: 20 open/close in 0.07 seconds: 296.26 ops/second Failing ost1 on oleg308-server Stopping /mnt/lustre-ost1 (opts:) on oleg308-server 07:16:24 (1713438984) shut down Failover ost1 to oleg308-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-OST0000 07:16:37 (1713438997) targets are mounted 07:16:37 (1713438997) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec fail_loc=0x80000204 total: 20 open/close in 0.06 seconds: 339.02 ops/second - unlinked 0 (time 1713439062 ; total 0 ; last 0) total: 20 unlinks in 0 seconds: inf unlinks/second PASS 47 (80s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 48: MDS->OSC failure during precreate cleanup (2824) ========================================================== 07:17:44 (1713439064) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre total: 20 open/close in 0.07 seconds: 291.83 ops/second Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:17:47 (1713439067) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:18:00 (1713439080) targets are mounted 07:18:00 (1713439080) facet_failover done fail_loc=0x80000216 total: 20 open/close in 0.07 seconds: 294.84 ops/second - unlinked 0 (time 1713439141 ; total 0 ; last 0) total: 40 unlinks in 0 seconds: inf unlinks/second PASS 48 (79s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 50: Double OSC recovery, don't LASSERT (3812) ========================================================== 07:19:04 (1713439144) PASS 50 (8s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 52: time out lock replay (3764) ==== 07:19:14 (1713439154) multiop /mnt/lustre/f52.replay-single vs_s TMPPIPE=/tmp/multiop_open_wait_pipe.6922 fail_loc=0x80000157 Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:19:16 (1713439156) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:19:28 (1713439168) targets are mounted 07:19:28 (1713439168) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec fail_loc=0x0 fail_loc=0x0 PASS 52 (35s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53a: |X| close request while two MDC requests in flight ========================================================== 07:19:51 (1713439191) fail_loc=0x80000115 fail_loc=0 Replay barrier on lustre-MDT0000 Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:19:56 (1713439196) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:20:09 (1713439209) targets are mounted 07:20:09 (1713439209) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d53a.replay-single-1/f has type file OK /mnt/lustre/d53a.replay-single-2/f has type file OK PASS 53a (22s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53b: |X| open request while two MDC requests in flight ========================================================== 07:20:15 (1713439215) multiop /mnt/lustre/d53b.replay-single-1/f vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.6922 fail_loc=0x80000107 fail_loc=0 Replay barrier on lustre-MDT0000 Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:20:19 (1713439219) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:20:33 (1713439233) targets are mounted 07:20:33 (1713439233) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d53b.replay-single-1/f has type file OK /mnt/lustre/d53b.replay-single-2/f has type file OK PASS 53b (23s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53c: |X| open request and close request while two MDC requests in flight ========================================================== 07:20:39 (1713439239) fail_loc=0x80000107 fail_loc=0x80000115 Replay barrier on lustre-MDT0000 Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:20:43 (1713439243) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:20:56 (1713439256) targets are mounted 07:20:56 (1713439256) facet_failover done fail_loc=0 /mnt/lustre/d53c.replay-single-1/f has type file OK /mnt/lustre/d53c.replay-single-2/f has type file OK PASS 53c (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53d: close reply while two MDC requests in flight ========================================================== 07:21:01 (1713439261) fail_loc=0x8000013b fail_loc=0 Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:21:04 (1713439264) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:21:17 (1713439277) targets are mounted 07:21:17 (1713439277) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d53d.replay-single-1/f has type file OK /mnt/lustre/d53d.replay-single-2/f has type file OK PASS 53d (21s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53e: |X| open reply while two MDC requests in flight ========================================================== 07:21:24 (1713439284) fail_loc=0x119 fail_loc=0 Replay barrier on lustre-MDT0000 Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:21:28 (1713439288) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:21:41 (1713439301) targets are mounted 07:21:41 (1713439301) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d53e.replay-single-1/f has type file OK /mnt/lustre/d53e.replay-single-2/f has type file OK PASS 53e (21s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53f: |X| open reply and close reply while two MDC requests in flight ========================================================== 07:21:46 (1713439306) fail_loc=0x119 fail_loc=0x8000013b Replay barrier on lustre-MDT0000 Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:21:50 (1713439310) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:22:03 (1713439323) targets are mounted 07:22:03 (1713439323) facet_failover done fail_loc=0 /mnt/lustre/d53f.replay-single-1/f has type file OK /mnt/lustre/d53f.replay-single-2/f has type file OK PASS 53f (21s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53g: |X| drop open reply and close request while close and open are both in flight ========================================================== 07:22:08 (1713439328) fail_loc=0x119 fail_loc=0x80000115 fail_loc=0 Replay barrier on lustre-MDT0000 Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:22:12 (1713439332) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:22:24 (1713439344) targets are mounted 07:22:24 (1713439344) facet_failover done /mnt/lustre/d53g.replay-single-1/f has type file OK /mnt/lustre/d53g.replay-single-2/f has type file OK PASS 53g (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53h: open request and close reply while two MDC requests in flight ========================================================== 07:22:29 (1713439349) fail_loc=0x80000107 fail_loc=0x8000013b Replay barrier on lustre-MDT0000 Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:22:34 (1713439354) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:22:46 (1713439366) targets are mounted 07:22:46 (1713439366) facet_failover done fail_loc=0 /mnt/lustre/d53h.replay-single-1/f has type file OK /mnt/lustre/d53h.replay-single-2/f has type file OK PASS 53h (21s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 55: let MDS_CHECK_RESENT return the original return code instead of 0 ========================================================== 07:22:51 (1713439371) fail_loc=0x8000012b fail_loc=0x0 touch: cannot touch '/mnt/lustre/f55.replay-single'rm: cannot remove '/mnt/lustre/f55.replay-single': No such file or directory: No such file or directory PASS 55 (18s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 56: don't replay a symlink open request (3440) ========================================================== 07:23:10 (1713439390) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:23:13 (1713439393) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:23:25 (1713439405) targets are mounted 07:23:25 (1713439405) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 56 (29s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 57: test recovery from llog for setattr op ========================================================== 07:23:41 (1713439421) fail_loc=0x8000012c UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:23:43 (1713439423) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:23:56 (1713439436) targets are mounted 07:23:56 (1713439436) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec affected facets: mds1 oleg308-server: oleg308-server.virtnet: executing _wait_recovery_complete *.lustre-MDT0000.recovery_status 1475 oleg308-server: *.lustre-MDT0000.recovery_status status: COMPLETE Waiting for orphan cleanup... osp.lustre-OST0000-osc-MDT0000.old_sync_processed osp.lustre-OST0001-osc-MDT0000.old_sync_processed wait 40 secs maximumly for oleg308-server mds-ost sync done. /mnt/lustre/f57.replay-single has type file OK fail_loc=0x0 PASS 57 (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 58a: test recovery from llog for setattr op (test llog_gen_rec) ========================================================== 07:24:11 (1713439451) fail_loc=0x8000012c total: 2500 open/close in 4.64 seconds: 538.40 ops/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3712 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:24:19 (1713439459) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:24:32 (1713439472) targets are mounted 07:24:32 (1713439472) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec fail_loc=0x0 - unlinked 0 (time 1713439483 ; total 0 ; last 0) total: 2500 unlinks in 3 seconds: 833.333313 unlinks/second PASS 58a (36s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 58b: test replay of setxattr op ==== 07:24:49 (1713439489) Starting client: oleg308-client.virtnet: -o user_xattr,flock oleg308-server@tcp:/lustre /mnt/lustre2 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3840 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:24:51 (1713439491) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:25:04 (1713439504) targets are mounted 07:25:04 (1713439504) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Stopping client oleg308-client.virtnet /mnt/lustre2 (opts:) oleg308-client.virtnet: executing wait_import_state_mount FULL mgc.*.mgs_server_uuid mgc.*.mgs_server_uuid in FULL state after 0 sec PASS 58b (20s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 58c: resend/reconstruct setxattr op ========================================================== 07:25:11 (1713439511) Starting client: oleg308-client.virtnet: -o user_xattr,flock oleg308-server@tcp:/lustre /mnt/lustre2 fail_val=0 fail_loc=0x123 fail_loc=0 fail_loc=0x119 fail_loc=0 Stopping client oleg308-client.virtnet /mnt/lustre2 (opts:) PASS 58c (40s) debug_raw_pointers=0 debug_raw_pointers=0 SKIP: replay-single test_59 skipping ALWAYS excluded test 59 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 60: test llog post recovery init vs llog unlink ========================================================== 07:25:53 (1713439553) total: 200 open/close in 0.38 seconds: 531.21 ops/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre - unlinked 0 (time 1713439555 ; total 0 ; last 0) total: 100 unlinks in 0 seconds: inf unlinks/second Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:25:56 (1713439556) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:26:09 (1713439569) targets are mounted 07:26:09 (1713439569) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec - unlinked 0 (time 1713439572 ; total 0 ; last 0) total: 100 unlinks in 0 seconds: inf unlinks/second PASS 60 (21s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 61a: test race llog recovery vs llog cleanup ========================================================== 07:26:15 (1713439575) total: 800 open/close in 1.45 seconds: 552.74 ops/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3712 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre - unlinked 0 (time 1713439579 ; total 0 ; last 0) total: 800 unlinks in 1 seconds: 800.000000 unlinks/second fail_val=0 fail_loc=0x80000221 Failing ost1 on oleg308-server Stopping /mnt/lustre-ost1 (opts:) on oleg308-server 07:26:22 (1713439582) shut down Failover ost1 to oleg308-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-OST0000 07:26:34 (1713439594) targets are mounted 07:26:34 (1713439594) facet_failover done Failing ost1 on oleg308-server Stopping /mnt/lustre-ost1 (opts:) on oleg308-server 07:26:46 (1713439606) shut down Failover ost1 to oleg308-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-OST0000 07:26:59 (1713439619) targets are mounted 07:26:59 (1713439619) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec fail_val=0 fail_loc=0x0 Can't lstat /mnt/lustre/d61a.replay-single/f61a.replay-single-*: No such file or directory PASS 61a (79s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 61b: test race mds llog sync vs llog cleanup ========================================================== 07:27:35 (1713439655) fail_loc=0x8000013a Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:27:37 (1713439657) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:27:50 (1713439670) targets are mounted 07:27:50 (1713439670) facet_failover done Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:28:01 (1713439681) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:28:13 (1713439693) targets are mounted 07:28:13 (1713439693) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec 1+0 records in 1+0 records out 4096 bytes (4.1 kB) copied, 0.00224742 s, 1.8 MB/s PASS 61b (43s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 61c: test race mds llog sync vs llog cleanup ========================================================== 07:28:19 (1713439699) fail_val=0 fail_loc=0x80000222 Failing ost1 on oleg308-server Stopping /mnt/lustre-ost1 (opts:) on oleg308-server 07:28:31 (1713439711) shut down Failover ost1 to oleg308-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-OST0000 07:28:43 (1713439723) targets are mounted 07:28:43 (1713439723) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec fail_val=0 fail_loc=0x0 PASS 61c (29s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 61d: error in llog_setup should cleanup the llog context correctly ========================================================== 07:28:50 (1713439730) Stopping /mnt/lustre-mds1 (opts:) on oleg308-server fail_loc=0x80000605 Starting mgs: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: mount.lustre: mount lustre-mdt1/mdt1 at /mnt/lustre-mds1 failed: Operation not supported pdsh@oleg308-client: oleg308-server: ssh exited with exit code 95 Start of lustre-mdt1/mdt1 on mgs failed 95 fail_loc=0 Starting mgs: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 PASS 61d (6s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 62: don't mis-drop resent replay === 07:28:58 (1713439738) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3712 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre total: 25 open/close in 0.07 seconds: 378.48 ops/second fail_loc=0x80000707 Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:29:01 (1713439741) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:29:13 (1713439753) targets are mounted 07:29:13 (1713439753) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec fail_loc=0 - unlinked 0 (time 1713439772 ; total 0 ; last 0) total: 25 unlinks in 0 seconds: inf unlinks/second PASS 62 (35s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 65a: AT: verify early replies ====== 07:29:35 (1713439775) at_history=8 at_history=8 debug=other fail_val=11000 fail_loc=0x8000050a 00000100:00001000:0.0:1713439804.137716:0:6949:0:(client.c:537:ptlrpc_at_recv_early_reply()) @@@ Early reply #1, new deadline in 36s (26s) req@ffff8800a601e300 x1796669160716608/t0(0) o101->lustre-MDT0000-mdc-ffff8800aab7c000@192.168.203.108@tcp:12/10 lens 664/66320 e 1 to 0 dl 1713439840 ref 2 fl Rpc:PQr/200/ffffffff rc 0/-1 job:'createmany.0' uid:0 gid:0 portal 12 : cur 37 worst 40 (at 1713438727, 1082s ago) 37 5 5 5 portal 23 : cur 5 worst 10 (at 1713437527, 2282s ago) 10 5 5 10 portal 30 : cur 5 worst 5 (at 1713437491, 2318s ago) 5 0 0 5 portal 17 : cur 5 worst 5 (at 1713437555, 2254s ago) 5 5 5 5 portal 13 : cur 10 worst 10 (at 1713437878, 1931s ago) 10 10 0 0 portal 12 : cur 12 worst 40 (at 1713438727, 1091s ago) 37 5 5 5 portal 23 : cur 5 worst 10 (at 1713437527, 2291s ago) 10 5 5 10 portal 30 : cur 5 worst 5 (at 1713437491, 2327s ago) 5 0 0 5 portal 17 : cur 5 worst 5 (at 1713437555, 2263s ago) 5 5 5 5 portal 13 : cur 10 worst 10 (at 1713437878, 1940s ago) 10 10 0 0 PASS 65a (44s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 65b: AT: verify early replies on packed reply / bulk ========================================================== 07:30:21 (1713439821) at_history=8 at_history=8 debug=other trace fail_val=11 fail_loc=0x224 fail_loc=0 00000100:00001000:1.0:1713439849.849788:0:1916:0:(client.c:537:ptlrpc_at_recv_early_reply()) @@@ Early reply #1, new deadline in 35s (25s) req@ffff8800a8ab4e00 x1796669160726464/t0(0) o4->lustre-OST0000-osc-ffff8800aab7c000@192.168.203.108@tcp:6/4 lens 4584/448 e 1 to 0 dl 1713439884 ref 2 fl Rpc:Qr/200/ffffffff rc 0/-1 job:'multiop.0' uid:0 gid:0 debug=super ioctl neterror warning dlmtrace error emerg ha rpctrace vfstrace config console lfsck debug=super ioctl neterror warning dlmtrace error emerg ha rpctrace vfstrace config console lfsck portal 28 : cur 5 worst 5 (at 1713437487, 2368s ago) 5 5 5 5 portal 7 : cur 5 worst 5 (at 1713437488, 2367s ago) 5 5 5 5 portal 6 : cur 36 worst 36 (at 1713439854, 1s ago) 36 5 0 0 portal 17 : cur 5 worst 5 (at 1713437703, 2152s ago) 5 0 5 5 PASS 65b (35s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 66a: AT: verify MDT service time adjusts with no early replies ========================================================== 07:30:58 (1713439858) at_history=8 at_history=8 portal 12 : cur 5 worst 40 (at 1713438727, 1153s ago) 37 5 5 5 fail_val=5000 fail_loc=0x8000050a portal 12 : cur 5 worst 40 (at 1713438727, 1158s ago) 37 5 5 5 fail_val=10000 fail_loc=0x8000050a portal 12 : cur 36 worst 40 (at 1713438727, 1169s ago) 36 37 5 5 fail_loc=0 portal 12 : cur 5 worst 40 (at 1713438727, 1178s ago) 36 37 5 5 Current MDT timeout 5, worst 40 PASS 66a (49s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 66b: AT: verify net latency adjusts ========================================================== 07:31:49 (1713439909) at_history=8 at_history=8 fail_val=10 fail_loc=0x50c fail_loc=0 network timeout orig 5, cur 10, worst 10 PASS 66b (84s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 67a: AT: verify slow request processing doesn't induce reconnects ========================================================== 07:33:15 (1713439995) at_history=8 at_history=8 fail_val=400 fail_loc=0x50a fail_loc=0 0 osc reconnect attempts on gradual slow PASS 67a (71s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 67b: AT: verify instant slowdown doesn't induce reconnects ========================================================== 07:34:28 (1713440068) at_history=8 at_history=8 Creating to objid 5441 on ost lustre-OST0000... fail_val=20000 fail_loc=0x80000223 total: 17 open/close in 0.04 seconds: 426.57 ops/second Connected clients: oleg308-client.virtnet oleg308-client.virtnet service : cur 5 worst 5 (at 1713437261, 2830s ago) 1 1 0 1 phase 2 0 osc reconnect attempts on instant slow fail_loc=0x80000223 fail_loc=0 Connected clients: oleg308-client.virtnet oleg308-client.virtnet service : cur 5 worst 5 (at 1713437261, 2831s ago) 1 1 0 1 0 osc reconnect attempts on 2nd slow PASS 67b (25s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 68: AT: verify slowing locks ======= 07:34:55 (1713440095) at_history=8 at_history=8 /home/green/git/lustre-release/lustre/tests/replay-single.sh: line 1968: $ldlm_enqueue_min: ambiguous redirect fail_val=19 fail_loc=0x80000312 fail_val=25 fail_loc=0x80000312 fail_loc=0 /home/green/git/lustre-release/lustre/tests/replay-single.sh: line 1983: $ldlm_enqueue_min: ambiguous redirect PASS 68 (68s) debug_raw_pointers=0 debug_raw_pointers=0 Cleaning up AT ... debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 70a: check multi client t-f ======== 07:36:05 (1713440165) SKIP: replay-single test_70a Need two or more clients, have 1 SKIP 70a (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 70b: dbench 1mdts recovery; 1 clients ========================================================== 07:36:07 (1713440167) Starting client oleg308-client.virtnet: -o user_xattr,flock oleg308-server@tcp:/lustre /mnt/lustre Started clients oleg308-client.virtnet: 192.168.203.108@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) + MISSING_DBENCH_OK= + PATH=/opt/iozone/bin:/opt/iozone/bin:/home/green/git/lustre-release/lustre/tests/mpi:/home/green/git/lustre-release/lustre/tests/racer:/home/green/git/lustre-release/lustre/../lustre-iokit/sgpdd-survey:/home/green/git/lustre-release/lustre/tests:/home/green/git/lustre-release/lustre/utils/gss:/home/green/git/lustre-release/lustre/utils:/opt/iozone/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin::/home/green/git/lustre-release/lustre/utils:/home/green/git/lustre-release/lustre/tests:/sbin:/usr/sbin:/home/green/git/lustre-release/lustre/utils:/home/green/git/lustre-release/lustre/tests/: + DBENCH_LIB= + TESTSUITE=replay-single + TESTNAME=test_70b + MOUNT=/mnt/lustre ++ hostname + DIR=/mnt/lustre/d70b.replay-single/oleg308-client.virtnet + LCTL=/home/green/git/lustre-release/lustre/utils/lctl + rundbench 1 -t 300 dbench: no process found Started rundbench load pid=13998 ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3968 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 32768 3683328 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3716096 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 35840 7399424 1% /mnt/lustre test_70b fail mds1 1 times Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:36:14 (1713440174) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:36:26 (1713440186) targets are mounted 07:36:26 (1713440186) facet_failover done oleg308-client.virtnet: looking for dbench program oleg308-client.virtnet: /usr/bin/dbench oleg308-client.virtnet: creating output directory /mnt/lustre/d70b.replay-single/oleg308-client.virtnet oleg308-client.virtnet: mkdir: created directory '/mnt/lustre/d70b.replay-single' oleg308-client.virtnet: mkdir: created directory '/mnt/lustre/d70b.replay-single/oleg308-client.virtnet' oleg308-client.virtnet: found dbench client file /usr/share/dbench/client.txt oleg308-client.virtnet: '/usr/share/dbench/client.txt' -> 'client.txt' oleg308-client.virtnet: running 'dbench 1 -t 300' on /mnt/lustre/d70b.replay-single/oleg308-client.virtnet at Thu Apr 18 07:36:08 EDT 2024 oleg308-client.virtnet: waiting for dbench pid 14022 oleg308-client.virtnet: dbench version 4.00 - Copyright Andrew Tridgell 1999-2004 oleg308-client.virtnet: oleg308-client.virtnet: Running for 300 seconds with load 'client.txt' and minimum warmup 60 secs oleg308-client.virtnet: failed to create barrier semaphore oleg308-client.virtnet: 0 of 1 processes prepared for launch 0 sec oleg308-client.virtnet: 1 of 1 processes prepared for launch 0 sec oleg308-client.virtnet: releasing clients oleg308-client.virtnet: 1 645 21.06 MB/sec warmup 1 sec latency 13.267 ms oleg308-client.virtnet: 1 1071 11.45 MB/sec warmup 2 sec latency 57.541 ms oleg308-client.virtnet: 1 1736 8.76 MB/sec warmup 3 sec latency 52.472 ms oleg308-client.virtnet: 1 2416 7.02 MB/sec warmup 4 sec latency 26.706 ms oleg308-client.virtnet: 1 3271 7.07 MB/sec warmup 5 sec latency 143.977 ms oleg308-client.virtnet: 1 3271 5.89 MB/sec warmup 6 sec latency 1144.167 ms oleg308-client.virtnet: 1 3271 5.05 MB/sec warmup 7 sec latency 2144.351 ms oleg308-client.virtnet: 1 3271 4.42 MB/sec warmup 8 sec latency 3144.539 ms oleg308-client.virtnet: 1 3271 3.93 MB/sec warmup 9 sec latency 4144.750 ms oleg308-client.virtnet: 1 3271 3.53 MB/sec warmup 10 sec latency 5144.910 ms oleg308-client.virtnet: 1 3271 3.21 MB/sec warmup 11 sec latency 6145.118 ms oleg308-client.virtnet: 1 3271 2.94 MB/sec warmup 12 sec latency 7145.305 ms oleg308-client.virtnet: 1 3271 2.72 MB/sec warmup 13 sec latency 8145.457 ms oleg308-client.virtnet: 1 3271 2.52 MB/sec warmup 14 sec latency 9145.603 ms oleg308-client.virtnet: 1 3271 2.35 MB/sec warmup 15 sec latency 10145.787 ms oleg308-client.virtnet: 1 3271 2.21 MB/sec warmup 16 sec latency 11145.901 ms oleg308-client.virtnet: 1 3271 2.08 MB/sec warmup 17 sec latency 12146.062 ms oleg308-client.virtnet: 1 3271 1.96 MB/sec warmup 18 sec latency 13146.214 ms oleg308-client.virtnet: 1 3271 1.86 MB/sec warmup 19 sec latency 14146.320 ms oleg308-client.virtnet: 1 3271 1.77 MB/sec warmup 20 sec latency 15146.493 ms oleg308-client.virtnet: 1 3271 1.68 MB/sec warmup 21 sec latency 16146.713 ms oleg308-client.virtnet: 1 3271 1.61 MB/sec warmup 22 sec latency 17146.908 ms oleg308-client.virtnet: 1 3271 1.54 MB/sec warmup 23 sec latency 18147.133 ms oleg308-client.virtnet: 1 3271 1.47 MB/sec warmup 24 sec latency 19147.305 ms oleg308-client.virtnet: 1 3271 1.41 MB/sec warmup 25 sec latency 20147.476 ms oleg308-client.virtnet: 1 3271 1.36 MB/sec warmup 26 sec latency 21147.647 ms oleg308-client.virtnet: 1 3271 1.31 MB/sec warmup 27 sec latency 22147.822 ms oleg308-client.virtnet: 1 3271 1.26 MB/sec warmup 28 sec latency 23148.006 ms oleg308-client.virtnet: 1 3271 1.22 MB/sec warmup 29 sec latency 24148.199 ms oleg308-client.virtnet: 1 3271 1.18 MB/sec warmup 30 sec latency 25148.417 ms oleg308-client.virtnet: 1 3271 1.14 MB/sec warmup 31 sec latency 26148.618 ms oleg308-client.virtnet: 1 3271 1.10 MB/sec warmup 32 soleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3968 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 33792 3693568 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 13312 3735552 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 47104 7429120 1% /mnt/lustre test_70b fail mds1 2 times Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:36:47 (1713440207) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:36:59 (1713440219) targets are mounted 07:36:59 (1713440219) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3840 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 40960 3704832 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 13312 3739648 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 54272 7444480 1% /mnt/lustre test_70b fail mds1 3 times Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:37:12 (1713440232) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:37:25 (1713440245) targets are mounted 07:37:25 (1713440245) facet_failover done ec latency 27148.819 ms oleg308-client.virtnet: 1 3776 1.20 MB/sec warmup 33 sec latency 27548.122 ms oleg308-client.virtnet: 1 4260 1.19 MB/sec warmup 34 sec latency 41.825 ms oleg308-client.virtnet: 1 4764 1.20 MB/sec warmup 35 sec latency 30.840 ms oleg308-client.virtnet: 1 5410 1.26 MB/sec warmup 36 sec latency 31.851 ms oleg308-client.virtnet: 1 6066 1.29 MB/sec warmup 37 sec latency 32.761 ms oleg308-client.virtnet: 1 7180 1.52 MB/sec warmup 38 sec latency 20.037 ms oleg308-client.virtnet: 1 7372 1.51 MB/sec warmup 39 sec latency 670.875 ms oleg308-client.virtnet: 1 7372 1.47 MB/sec warmup 40 sec latency 1671.020 ms oleg308-client.virtnet: 1 7372 1.44 MB/sec warmup 41 sec latency 2671.207 ms oleg308-client.virtnet: 1 7372 1.40 MB/sec warmup 42 sec latency 3671.372 ms oleg308-client.virtnet: 1 7372 1.37 MB/sec warmup 43 sec latency 4671.548 ms oleg308-client.virtnet: 1 7372 1.34 MB/sec warmup 44 sec latency 5671.710 ms oleg308-client.virtnet: 1 7372 1.31 MB/sec warmup 45 sec latency 6671.943 ms oleg308-client.virtnet: 1 7372 1.28 MB/sec warmup 46 sec latency 7672.152 ms oleg308-client.virtnet: 1 7372 1.25 MB/sec warmup 47 sec latency 8672.370 ms oleg308-client.virtnet: 1 7372 1.23 MB/sec warmup 48 sec latency 9672.554 ms oleg308-client.virtnet: 1 7372 1.20 MB/sec warmup 49 sec latency 10672.734 ms oleg308-client.virtnet: 1 7372 1.18 MB/sec warmup 50 sec latency 11672.939 ms oleg308-client.virtnet: 1 7372 1.15 MB/sec warmup 51 sec latency 12673.079 ms oleg308-client.virtnet: 1 7372 1.13 MB/sec warmup 52 sec latency 13673.239 ms oleg308-client.virtnet: 1 7372 1.11 MB/sec warmup 53 sec latency 14673.427 ms oleg308-client.virtnet: 1 7372 1.09 MB/sec warmup 54 sec latency 15673.616 ms oleg308-client.virtnet: 1 7372 1.07 MB/sec warmup 55 sec latency 16673.777 ms oleg308-client.virtnet: 1 7372 1.05 MB/sec warmup 56 sec latency 17673.974 ms oleg308-client.virtnet: 1 7372 1.03 MB/sec warmup 57 sec latency 18674.171 ms oleg308-client.virtnet: 1 7614 1.02 MB/sec warmup 58 sec latency 19142.558 ms oleg308-client.virtnet: 1 8000 1.02 MB/sec warmup 59 sec latency 58.954 ms oleg308-client.virtnet: 1 9399 1.66 MB/sec execute 1 sec latency 28.305 ms oleg308-client.virtnet: 1 10306 4.09 MB/sec execute 2 sec latency 22.902 ms oleg308-client.virtnet: 1 11040 4.65 MB/sec execute 3 sec latency 13.926 ms oleg308-client.virtnet: 1 11335 3.55 MB/sec execute 4 sec latency 431.169 ms oleg308-client.virtnet: 1 11335 2.84 MB/sec execute 5 sec latency 1431.357 ms oleg308-client.virtnet: 1 11335 2.36 MB/sec execute 6 sec latency 2431.577 ms oleg308-client.virtnet: 1 11335 2.03 MB/sec execute 7 sec latency 3431.746 ms oleg308-client.virtnet: 1 11335 1.77 MB/sec execute 8 sec latency 4431.928 ms oleg308-client.virtnet: 1 11335 1.58 MB/sec execute 9 sec latency 5432.128 ms oleg308-client.virtnet: 1 11335 1.42 MB/sec execute 10 sec latency 6432.386 ms oleg308-client.virtnet: 1 11335 1.29 MB/sec execute 11 sec latency 7432.562 ms oleg308-client.virtnet: 1 11335 1.18 MB/sec execute 12 sec latency 8432.734 ms oleg308-client.virtnet: 1 11335 1.09 MB/sec execute 13 sec latency 9432.938 ms oleg308-client.virtnet: 1 11335 1.01 MB/sec execute 14 sec latency 10433.197 ms oleg308-client.virtnet: 1 11335 0.95 MB/sec execute 15 sec latency 11433.337 ms oleg308-client.virtnet: 1 11335 0.89 MB/sec execute 16 sec latency 12433.454 ms oleg308-client.virtnet: 1 11335 0.83 MB/sec execute 17 sec latency 13433.614 ms oleg308-clioleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3968 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 40960 3698688 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 13312 3739648 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 54272 7438336 1% /mnt/lustre test_70b fail mds1 4 times Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:37:37 (1713440257) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:37:50 (1713440270) targets are mounted 07:37:50 (1713440270) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3968 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 40960 3696640 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 13312 3732480 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 54272 7429120 1% /mnt/lustre test_70b fail mds1 5 times Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:38:02 (1713440282) shut down ent.virtnet: 1 11335 0.79 MB/sec execute 18 sec latency 14433.811 ms oleg308-client.virtnet: 1 11335 0.75 MB/sec execute 19 sec latency 15434.001 ms oleg308-client.virtnet: 1 11335 0.71 MB/sec execute 20 sec latency 16434.220 ms oleg308-client.virtnet: 1 11335 0.68 MB/sec execute 21 sec latency 17434.415 ms oleg308-client.virtnet: 1 11335 0.64 MB/sec execute 22 sec latency 18434.603 ms oleg308-client.virtnet: 1 11475 0.63 MB/sec execute 23 sec latency 19033.584 ms oleg308-client.virtnet: 1 12137 0.79 MB/sec execute 24 sec latency 24.926 ms oleg308-client.virtnet: 1 12743 0.79 MB/sec execute 25 sec latency 32.545 ms oleg308-client.virtnet: 1 13572 1.02 MB/sec execute 26 sec latency 29.446 ms oleg308-client.virtnet: 1 14424 1.21 MB/sec execute 27 sec latency 27.706 ms oleg308-client.virtnet: 1 14893 1.19 MB/sec execute 28 sec latency 45.453 ms oleg308-client.virtnet: 1 15098 1.17 MB/sec execute 29 sec latency 503.328 ms oleg308-client.virtnet: 1 15098 1.13 MB/sec execute 30 sec latency 1503.496 ms oleg308-client.virtnet: 1 15098 1.09 MB/sec execute 31 sec latency 2503.661 ms oleg308-client.virtnet: 1 15098 1.06 MB/sec execute 32 sec latency 3503.840 ms oleg308-client.virtnet: 1 15098 1.02 MB/sec execute 33 sec latency 4504.062 ms oleg308-client.virtnet: 1 15098 0.99 MB/sec execute 34 sec latency 5504.251 ms oleg308-client.virtnet: 1 15098 0.97 MB/sec execute 35 sec latency 6504.461 ms oleg308-client.virtnet: 1 15098 0.94 MB/sec execute 36 sec latency 7504.660 ms oleg308-client.virtnet: 1 15098 0.91 MB/sec execute 37 sec latency 8504.908 ms oleg308-client.virtnet: 1 15098 0.89 MB/sec execute 38 sec latency 9505.091 ms oleg308-client.virtnet: 1 15098 0.87 MB/sec execute 39 sec latency 10505.307 ms oleg308-client.virtnet: 1 15098 0.85 MB/sec execute 40 sec latency 11505.467 ms oleg308-client.virtnet: 1 15098 0.82 MB/sec execute 41 sec latency 12505.583 ms oleg308-client.virtnet: 1 15098 0.81 MB/sec execute 42 sec latency 13505.769 ms oleg308-client.virtnet: 1 15098 0.79 MB/sec execute 43 sec latency 14505.974 ms oleg308-client.virtnet: 1 15098 0.77 MB/sec execute 44 sec latency 15506.204 ms oleg308-client.virtnet: 1 15098 0.75 MB/sec execute 45 sec latency 16506.397 ms oleg308-client.virtnet: 1 15098 0.74 MB/sec execute 46 sec latency 17506.644 ms oleg308-client.virtnet: 1 15098 0.72 MB/sec execute 47 sec latency 18506.900 ms oleg308-client.virtnet: 1 15321 0.73 MB/sec execute 48 sec latency 19104.461 ms oleg308-client.virtnet: 1 15976 0.78 MB/sec execute 49 sec latency 41.701 ms oleg308-client.virtnet: 1 16624 0.80 MB/sec execute 50 sec latency 33.145 ms oleg308-client.virtnet: 1 17566 0.93 MB/sec execute 51 sec latency 26.249 ms oleg308-client.virtnet: 1 18177 1.00 MB/sec execute 52 sec latency 8.481 ms oleg308-client.virtnet: 1 18657 1.00 MB/sec execute 53 sec latency 31.903 ms oleg308-client.virtnet: 1 18965 1.01 MB/sec execute 54 sec latency 484.932 ms oleg308-client.virtnet: 1 18965 0.99 MB/sec execute 55 sec latency 1485.111 ms oleg308-client.virtnet: 1 18965 0.97 MB/sec execute 56 sec latency 2485.311 ms oleg308-client.virtnet: 1 18965 0.95 MB/sec execute 57 sec latency 3485.510 ms oleg308-client.virtnet: 1 18965 0.94 MB/sec execute 58 sec latency 4485.692 ms oleg308-client.virtnet: 1 18965 0.92 MB/sec execute 59 sec latency 5485.878 ms oleg308-client.virtnet: 1 18965 0.91 MB/sec execute 60 sec latency 6486.133 ms oleg308-client.virtnet: 1 18965 0.89 MB/sec execute 61 sec latency 7486.350 ms oleg308-client.virtnetFailover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:38:15 (1713440295) targets are mounted 07:38:15 (1713440295) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3968 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 40960 3714048 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 14336 3739648 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 55296 7453696 1% /mnt/lustre test_70b fail mds1 6 times Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:38:27 (1713440307) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:38:39 (1713440319) targets are mounted 07:38:39 (1713440319) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3968 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 40960 3713024 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 14336 3713024 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 55296 7426048 1% /mnt/lustre test_70b fail mds1 7 times Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:38:52 (1713440332) shut down : 1 18965 0.88 MB/sec execute 62 sec latency 8486.513 ms oleg308-client.virtnet: 1 18965 0.86 MB/sec execute 63 sec latency 9486.653 ms oleg308-client.virtnet: 1 18965 0.85 MB/sec execute 64 sec latency 10486.845 ms oleg308-client.virtnet: 1 18965 0.84 MB/sec execute 65 sec latency 11486.993 ms oleg308-client.virtnet: 1 18965 0.82 MB/sec execute 66 sec latency 12487.192 ms oleg308-client.virtnet: 1 18965 0.81 MB/sec execute 67 sec latency 13487.400 ms oleg308-client.virtnet: 1 18965 0.80 MB/sec execute 68 sec latency 14487.552 ms oleg308-client.virtnet: 1 18965 0.79 MB/sec execute 69 sec latency 15487.743 ms oleg308-client.virtnet: 1 18965 0.78 MB/sec execute 70 sec latency 16487.968 ms oleg308-client.virtnet: 1 18965 0.76 MB/sec execute 71 sec latency 17488.146 ms oleg308-client.virtnet: 1 18965 0.75 MB/sec execute 72 sec latency 18488.358 ms oleg308-client.virtnet: 1 19244 0.78 MB/sec execute 73 sec latency 19126.361 ms oleg308-client.virtnet: 1 19876 0.78 MB/sec execute 74 sec latency 25.583 ms oleg308-client.virtnet: 1 20697 0.86 MB/sec execute 75 sec latency 26.346 ms oleg308-client.virtnet: 1 21561 0.94 MB/sec execute 76 sec latency 24.328 ms oleg308-client.virtnet: 1 22057 0.93 MB/sec execute 77 sec latency 30.514 ms oleg308-client.virtnet: 1 22557 0.94 MB/sec execute 78 sec latency 24.039 ms oleg308-client.virtnet: 1 22964 0.97 MB/sec execute 79 sec latency 381.112 ms oleg308-client.virtnet: 1 22964 0.96 MB/sec execute 80 sec latency 1381.310 ms oleg308-client.virtnet: 1 22964 0.95 MB/sec execute 81 sec latency 2381.494 ms oleg308-client.virtnet: 1 22964 0.93 MB/sec execute 82 sec latency 3381.676 ms oleg308-client.virtnet: 1 22964 0.92 MB/sec execute 83 sec latency 4381.870 ms oleg308-client.virtnet: 1 22964 0.91 MB/sec execute 84 sec latency 5382.082 ms oleg308-client.virtnet: 1 22964 0.90 MB/sec execute 85 sec latency 6382.271 ms oleg308-client.virtnet: 1 22964 0.89 MB/sec execute 86 sec latency 7382.456 ms oleg308-client.virtnet: 1 22964 0.88 MB/sec execute 87 sec latency 8382.653 ms oleg308-client.virtnet: 1 22964 0.87 MB/sec execute 88 sec latency 9382.866 ms oleg308-client.virtnet: 1 22964 0.86 MB/sec execute 89 sec latency 10383.016 ms oleg308-client.virtnet: 1 22964 0.85 MB/sec execute 90 sec latency 11383.155 ms oleg308-client.virtnet: 1 22964 0.84 MB/sec execute 91 sec latency 12383.307 ms oleg308-client.virtnet: 1 22964 0.83 MB/sec execute 92 sec latency 13383.456 ms oleg308-client.virtnet: 1 22964 0.82 MB/sec execute 93 sec latency 14383.664 ms oleg308-client.virtnet: 1 22964 0.81 MB/sec execute 94 sec latency 15383.877 ms oleg308-client.virtnet: 1 22964 0.81 MB/sec execute 95 sec latency 16384.043 ms oleg308-client.virtnet: 1 22964 0.80 MB/sec execute 96 sec latency 17384.235 ms oleg308-client.virtnet: 1 22964 0.79 MB/sec execute 97 sec latency 18384.440 ms oleg308-client.virtnet: 1 23214 0.78 MB/sec execute 98 sec latency 18963.821 ms oleg308-client.virtnet: 1 23932 0.80 MB/sec execute 99 sec latency 27.875 ms oleg308-client.virtnet: 1 25004 0.89 MB/sec execute 100 sec latency 22.076 ms oleg308-client.virtnet: 1 25556 0.90 MB/sec execute 101 sec latency 39.320 ms oleg308-client.virtnet: 1 26031 0.91 MB/sec execute 102 sec latency 23.092 ms oleg308-client.virtnet: 1 26711 0.93 MB/sec execute 103 sec latency 26.247 ms oleg308-client.virtnet: 1 26976 0.92 MB/sec execute 104 sec latency 645.125 ms oleg308-client.virtnet: 1 26976 0.92 MB/sec execute 105 sec latency 1645.288 ms oleg308-client.virtnet: 1Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:39:05 (1713440345) targets are mounted 07:39:05 (1713440345) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3968 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 40960 3693568 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 14336 3697664 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 55296 7391232 1% /mnt/lustre test_70b fail mds1 8 times Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:39:18 (1713440358) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:39:30 (1713440370) targets are mounted 07:39:30 (1713440370) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec 26976 0.91 MB/sec execute 106 sec latency 2645.486 ms oleg308-client.virtnet: 1 26976 0.90 MB/sec execute 107 sec latency 3645.634 ms oleg308-client.virtnet: 1 26976 0.89 MB/sec execute 108 sec latency 4645.815 ms oleg308-client.virtnet: 1 26976 0.88 MB/sec execute 109 sec latency 5646.021 ms oleg308-client.virtnet: 1 26976 0.87 MB/sec execute 110 sec latency 6646.237 ms oleg308-client.virtnet: 1 26976 0.87 MB/sec execute 111 sec latency 7646.418 ms oleg308-client.virtnet: 1 26976 0.86 MB/sec execute 112 sec latency 8646.652 ms oleg308-client.virtnet: 1 26976 0.85 MB/sec execute 113 sec latency 9646.882 ms oleg308-client.virtnet: 1 26976 0.84 MB/sec execute 114 sec latency 10647.017 ms oleg308-client.virtnet: 1 26976 0.84 MB/sec execute 115 sec latency 11647.210 ms oleg308-client.virtnet: 1 26976 0.83 MB/sec execute 116 sec latency 12647.446 ms oleg308-client.virtnet: 1 26976 0.82 MB/sec execute 117 sec latency 13647.768 ms oleg308-client.virtnet: 1 26976 0.82 MB/sec execute 118 sec latency 14648.038 ms oleg308-client.virtnet: 1 26976 0.81 MB/sec execute 119 sec latency 15648.235 ms oleg308-client.virtnet: 1 26976 0.80 MB/sec execute 120 sec latency 16648.464 ms oleg308-client.virtnet: 1 26976 0.80 MB/sec execute 121 sec latency 17648.664 ms oleg308-client.virtnet: 1 26976 0.79 MB/sec execute 122 sec latency 18648.838 ms oleg308-client.virtnet: 1 27168 0.79 MB/sec execute 123 sec latency 19313.398 ms oleg308-client.virtnet: 1 27775 0.83 MB/sec execute 124 sec latency 39.112 ms oleg308-client.virtnet: 1 28581 0.87 MB/sec execute 125 sec latency 25.987 ms oleg308-client.virtnet: 1 29098 0.87 MB/sec execute 126 sec latency 39.194 ms oleg308-client.virtnet: 1 29502 0.88 MB/sec execute 127 sec latency 28.515 ms oleg308-client.virtnet: 1 29976 0.90 MB/sec execute 128 sec latency 43.553 ms oleg308-client.virtnet: 1 30525 0.89 MB/sec execute 129 sec latency 89.682 ms oleg308-client.virtnet: 1 30525 0.89 MB/sec execute 130 sec latency 1089.872 ms oleg308-client.virtnet: 1 30525 0.88 MB/sec execute 131 sec latency 2090.008 ms oleg308-client.virtnet: 1 30525 0.87 MB/sec execute 132 sec latency 3090.184 ms oleg308-client.virtnet: 1 30525 0.87 MB/sec execute 133 sec latency 4090.464 ms oleg308-client.virtnet: 1 30525 0.86 MB/sec execute 134 sec latency 5090.719 ms oleg308-client.virtnet: 1 30525 0.85 MB/sec execute 135 sec latency 6090.921 ms oleg308-client.virtnet: 1 30525 0.85 MB/sec execute 136 sec latency 7091.161 ms oleg308-client.virtnet: 1 30525 0.84 MB/sec execute 137 sec latency 8091.377 ms oleg308-client.virtnet: 1 30525 0.84 MB/sec execute 138 sec latency 9091.659 ms oleg308-client.virtnet: 1 30525 0.83 MB/sec execute 139 sec latency 10091.871 ms oleg308-client.virtnet: 1 30525 0.82 MB/sec execute 140 sec latency 11092.010 ms oleg308-client.virtnet: 1 30525 0.82 MB/sec execute 141 sec latency 12092.160 ms oleg308-client.virtnet: 1 30525 0.81 MB/sec execute 142 sec latency 13092.354 ms oleg308-client.virtnet: 1 30525 0.81 MB/sec execute 143 sec latency 14092.515 ms oleg308-client.virtnet: 1 30525 0.80 MB/sec execute 144 sec latency 15092.677 ms oleg308-client.virtnet: 1 30525 0.80 MB/sec execute 145 sec latency 16092.823 ms oleg308-client.virtnet: 1 30525 0.79 MB/sec execute 146 sec latency 17092.928 ms oleg308-client.virtnet: 1 30525 0.78 MB/sec execute 147 sec latency 18093.119 ms oleg308-client.virtnet: 1 30753 0.79 MB/sec execute 148 sec latency 18786.142 ms oleg308-client.virtnet: 1 31021 0.79 MB/sec execute 149 sec latency 58.982 ms oleg308-client.virtnet: UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3968 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 40960 3702784 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 14336 3726336 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 55296 7429120 1% /mnt/lustre test_70b fail mds1 9 times Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:39:43 (1713440383) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:39:56 (1713440396) targets are mounted 07:39:56 (1713440396) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3968 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 40960 3706880 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 13312 3729408 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 54272 7436288 1% /mnt/lustre test_70b fail mds1 10 times Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:40:08 (1713440408) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:40:21 (1713440421) targets are mounted 07:40:21 (1713440421) facet_failover done 1 31998 0.85 MB/sec execute 150 sec latency 34.290 ms oleg308-client.virtnet: 1 32496 0.85 MB/sec execute 151 sec latency 8.901 ms oleg308-client.virtnet: 1 32845 0.85 MB/sec execute 152 sec latency 49.247 ms oleg308-client.virtnet: 1 33127 0.85 MB/sec execute 153 sec latency 47.113 ms oleg308-client.virtnet: 1 33509 0.87 MB/sec execute 154 sec latency 56.559 ms oleg308-client.virtnet: 1 33763 0.87 MB/sec execute 155 sec latency 358.411 ms oleg308-client.virtnet: 1 33763 0.86 MB/sec execute 156 sec latency 1358.615 ms oleg308-client.virtnet: 1 33763 0.85 MB/sec execute 157 sec latency 2358.896 ms oleg308-client.virtnet: 1 33763 0.85 MB/sec execute 158 sec latency 3359.085 ms oleg308-client.virtnet: 1 33763 0.84 MB/sec execute 159 sec latency 4359.245 ms oleg308-client.virtnet: 1 33763 0.84 MB/sec execute 160 sec latency 5359.451 ms oleg308-client.virtnet: 1 33763 0.83 MB/sec execute 161 sec latency 6359.684 ms oleg308-client.virtnet: 1 33763 0.83 MB/sec execute 162 sec latency 7359.860 ms oleg308-client.virtnet: 1 33763 0.82 MB/sec execute 163 sec latency 8360.036 ms oleg308-client.virtnet: 1 33763 0.82 MB/sec execute 164 sec latency 9360.265 ms oleg308-client.virtnet: 1 33763 0.81 MB/sec execute 165 sec latency 10360.424 ms oleg308-client.virtnet: 1 33763 0.81 MB/sec execute 166 sec latency 11360.603 ms oleg308-client.virtnet: 1 33763 0.80 MB/sec execute 167 sec latency 12360.837 ms oleg308-client.virtnet: 1 33763 0.80 MB/sec execute 168 sec latency 13361.066 ms oleg308-client.virtnet: 1 33763 0.79 MB/sec execute 169 sec latency 14361.243 ms oleg308-client.virtnet: 1 33763 0.79 MB/sec execute 170 sec latency 15361.378 ms oleg308-client.virtnet: 1 33763 0.78 MB/sec execute 171 sec latency 16361.512 ms oleg308-client.virtnet: 1 33763 0.78 MB/sec execute 172 sec latency 17361.726 ms oleg308-client.virtnet: 1 33946 0.78 MB/sec execute 173 sec latency 17976.470 ms oleg308-client.virtnet: 1 34371 0.78 MB/sec execute 174 sec latency 25.931 ms oleg308-client.virtnet: 1 35174 0.81 MB/sec execute 175 sec latency 29.123 ms oleg308-client.virtnet: 1 35886 0.84 MB/sec execute 176 sec latency 16.917 ms oleg308-client.virtnet: 1 36189 0.84 MB/sec execute 177 sec latency 53.076 ms oleg308-client.virtnet: 1 36663 0.84 MB/sec execute 178 sec latency 24.321 ms oleg308-client.virtnet: 1 37237 0.86 MB/sec execute 179 sec latency 25.613 ms oleg308-client.virtnet: 1 37381 0.85 MB/sec execute 180 sec latency 529.786 ms oleg308-client.virtnet: 1 37381 0.85 MB/sec execute 181 sec latency 1529.977 ms oleg308-client.virtnet: 1 37381 0.84 MB/sec execute 182 sec latency 2530.245 ms oleg308-client.virtnet: 1 37381 0.84 MB/sec execute 183 sec latency 3530.540 ms oleg308-client.virtnet: 1 37381 0.83 MB/sec execute 184 sec latency 4530.759 ms oleg308-client.virtnet: 1 37381 0.83 MB/sec execute 185 sec latency 5530.946 ms oleg308-client.virtnet: 1 37381 0.82 MB/sec execute 186 sec latency 6531.202 ms oleg308-client.virtnet: 1 37381 0.82 MB/sec execute 187 sec latency 7531.489 ms oleg308-client.virtnet: 1 37381 0.82 MB/sec execute 188 sec latency 8531.776 ms oleg308-client.virtnet: 1 37381 0.81 MB/sec execute 189 sec latency 9531.952 ms oleg308-client.virtnet: 1 37381 0.81 MB/sec execute 190 sec latency 10532.173 ms oleg308-client.virtnet: 1 37381 0.80 MB/sec execute 191 sec latency 11532.365 ms oleg308-client.virtnet: 1 37381 0.80 MB/sec execute 192 sec latency 12532.499 ms oleg308-client.virtnet: 1 37381 0.79 MB/sec execute 193 sec latency 13532.706 ms oleg308-client.virtnet: 1 37381 oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3968 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 40960 3712000 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 14336 3720192 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 55296 7432192 1% /mnt/lustre test_70b fail mds1 11 times Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:40:38 (1713440438) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:40:52 (1713440452) targets are mounted 07:40:52 (1713440452) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3968 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 40960 3712000 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 14336 3719168 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 55296 7431168 1% /mnt/lustre 0.79 MB/sec execute 194 sec latency 14532.944 ms oleg308-client.virtnet: 1 37381 0.79 MB/sec execute 195 sec latency 15533.191 ms oleg308-client.virtnet: 1 37381 0.78 MB/sec execute 196 sec latency 16533.491 ms oleg308-client.virtnet: 1 37381 0.78 MB/sec execute 197 sec latency 17533.869 ms oleg308-client.virtnet: 1 37553 0.78 MB/sec execute 198 sec latency 18233.028 ms oleg308-client.virtnet: 1 38384 0.80 MB/sec execute 199 sec latency 26.413 ms oleg308-client.virtnet: 1 39191 0.83 MB/sec execute 200 sec latency 37.730 ms oleg308-client.virtnet: 1 39732 0.83 MB/sec execute 201 sec latency 8.458 ms oleg308-client.virtnet: 1 40155 0.84 MB/sec execute 202 sec latency 38.055 ms oleg308-client.virtnet: 1 40696 0.85 MB/sec execute 203 sec latency 45.057 ms oleg308-client.virtnet: 1 41087 0.85 MB/sec execute 204 sec latency 366.030 ms oleg308-client.virtnet: 1 41087 0.84 MB/sec execute 205 sec latency 1366.212 ms oleg308-client.virtnet: 1 41087 0.84 MB/sec execute 206 sec latency 2366.420 ms oleg308-client.virtnet: 1 41087 0.83 MB/sec execute 207 sec latency 3366.610 ms oleg308-client.virtnet: 1 41087 0.83 MB/sec execute 208 sec latency 4366.798 ms oleg308-client.virtnet: 1 41087 0.83 MB/sec execute 209 sec latency 5366.966 ms oleg308-client.virtnet: 1 41087 0.82 MB/sec execute 210 sec latency 6367.105 ms oleg308-client.virtnet: 1 41087 0.82 MB/sec execute 211 sec latency 7367.264 ms oleg308-client.virtnet: 1 41087 0.81 MB/sec execute 212 sec latency 8367.462 ms oleg308-client.virtnet: 1 41087 0.81 MB/sec execute 213 sec latency 9367.702 ms oleg308-client.virtnet: 1 41087 0.81 MB/sec execute 214 sec latency 10367.857 ms oleg308-client.virtnet: 1 41087 0.80 MB/sec execute 215 sec latency 11368.019 ms oleg308-client.virtnet: 1 41087 0.80 MB/sec execute 216 sec latency 12368.129 ms oleg308-client.virtnet: 1 41087 0.80 MB/sec execute 217 sec latency 13368.284 ms oleg308-client.virtnet: 1 41087 0.79 MB/sec execute 218 sec latency 14368.415 ms oleg308-client.virtnet: 1 41087 0.79 MB/sec execute 219 sec latency 15368.542 ms oleg308-client.virtnet: 1 41087 0.78 MB/sec execute 220 sec latency 16368.700 ms oleg308-client.virtnet: 1 41087 0.78 MB/sec execute 221 sec latency 17368.845 ms oleg308-client.virtnet: 1 41087 0.78 MB/sec execute 222 sec latency 18368.970 ms oleg308-client.virtnet: 1 41087 0.77 MB/sec execute 223 sec latency 19369.089 ms oleg308-client.virtnet: 1 41087 0.77 MB/sec execute 224 sec latency 20369.201 ms oleg308-client.virtnet: 1 41087 0.77 MB/sec execute 225 sec latency 21369.358 ms oleg308-client.virtnet: 1 41087 0.76 MB/sec execute 226 sec latency 22369.522 ms oleg308-client.virtnet: 1 41087 0.76 MB/sec execute 227 sec latency 23369.672 ms oleg308-client.virtnet: 1 41087 0.76 MB/sec execute 228 sec latency 24369.808 ms oleg308-client.virtnet: 1 41087 0.75 MB/sec execute 229 sec latency 25369.969 ms oleg308-client.virtnet: 1 41087 0.75 MB/sec execute 230 sec latency 26370.125 ms oleg308-client.virtnet: 1 41087 0.75 MB/sec execute 231 sec latency 27370.264 ms oleg308-client.virtnet: 1 41087 0.74 MB/sec execute 232 sec latency 28370.386 ms oleg308-client.virtnet: 1 41221 0.74 MB/sec execute 233 sec latency 29124.491 ms oleg308-client.virtnet: 1 41654 0.75 MB/sec execute 234 sec latency 48.739 ms oleg308-client.virtnet: 1 42423 0.77 MB/sec execute 235 sec latency 40.025 ms oleg308-client.virtnet: 1 42939 0.79 MB/sec execute 236 sec latency 12.325 ms oleg308-client.virtnet: 1 43330 0.79 MB/sec execute 237 sec latency 51.891 ms oleg308-client.virtnet: 1 43796 test_70b fail mds1 12 times Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:41:08 (1713440468) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:41:21 (1713440481) targets are mounted 07:41:21 (1713440481) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec 0.79 MB/sec execute 238 sec latency 25.488 ms oleg308-client.virtnet: 1 44459 0.80 MB/sec execute 239 sec latency 26.923 ms oleg308-client.virtnet: 1 44459 0.80 MB/sec execute 240 sec latency 1019.432 ms oleg308-client.virtnet: 1 44459 0.79 MB/sec execute 241 sec latency 2019.580 ms oleg308-client.virtnet: 1 44459 0.79 MB/sec execute 242 sec latency 3019.749 ms oleg308-client.virtnet: 1 44459 0.79 MB/sec execute 243 sec latency 4019.904 ms oleg308-client.virtnet: 1 44459 0.78 MB/sec execute 244 sec latency 5020.049 ms oleg308-client.virtnet: 1 44459 0.78 MB/sec execute 245 sec latency 6020.210 ms oleg308-client.virtnet: 1 44459 0.78 MB/sec execute 246 sec latency 7020.342 ms oleg308-client.virtnet: 1 44459 0.78 MB/sec execute 247 sec latency 8020.466 ms oleg308-client.virtnet: 1 44459 0.77 MB/sec execute 248 sec latency 9020.693 ms oleg308-client.virtnet: 1 44459 0.77 MB/sec execute 249 sec latency 10020.850 ms oleg308-client.virtnet: 1 44459 0.77 MB/sec execute 250 sec latency 11021.046 ms oleg308-client.virtnet: 1 44459 0.76 MB/sec execute 251 sec latency 12021.234 ms oleg308-client.virtnet: 1 44459 0.76 MB/sec execute 252 sec latency 13021.470 ms oleg308-client.virtnet: 1 44459 0.76 MB/sec execute 253 sec latency 14021.710 ms oleg308-client.virtnet: 1 44459 0.75 MB/sec execute 254 sec latency 15021.841 ms oleg308-client.virtnet: 1 44459 0.75 MB/sec execute 255 sec latency 16022.010 ms oleg308-client.virtnet: 1 44459 0.75 MB/sec execute 256 sec latency 17022.221 ms oleg308-client.virtnet: 1 44459 0.75 MB/sec execute 257 sec latency 18022.424 ms oleg308-client.virtnet: 1 44489 0.74 MB/sec execute 258 sec latency 18887.397 ms oleg308-client.virtnet: 1 45006 0.75 MB/sec execute 259 sec latency 25.676 ms oleg308-client.virtnet: 1 45771 0.77 MB/sec execute 260 sec latency 35.592 ms oleg308-client.virtnet: 1 46326 0.78 MB/sec execute 261 sec latency 47.142 ms oleg308-client.virtnet: 1 46664 0.78 MB/sec execute 262 sec latency 11.852 ms oleg308-client.virtnet: 1 47031 0.78 MB/sec execute 263 sec latency 54.389 ms oleg308-client.virtnet: 1 47545 0.80 MB/sec execute 264 sec latency 41.280 ms oleg308-client.virtnet: 1 48020 0.79 MB/sec execute 265 sec latency 33.706 ms oleg308-client.virtnet: 1 48552 0.80 MB/sec execute 266 sec latency 43.124 ms oleg308-client.virtnet: 1 49155 0.82 MB/sec execute 267 sec latency 38.613 ms oleg308-client.virtnet: 1 49879 0.83 MB/sec execute 268 sec latency 30.415 ms oleg308-client.virtnet: 1 50258 0.83 MB/sec execute 269 sec latency 16.005 ms oleg308-client.virtnet: 1 50612 0.83 MB/sec execute 270 sec latency 38.921 ms oleg308-client.virtnet: 1 51116 0.85 MB/sec execute 271 sec latency 57.651 ms oleg308-client.virtnet: 1 51495 0.84 MB/sec execute 272 sec latency 52.803 ms oleg308-client.virtnet: 1 52028 0.85 MB/sec execute 273 sec latency 36.403 ms oleg308-client.virtnet: 1 52585 0.86 MB/sec execute 274 sec latency 44.120 ms oleg308-client.virtnet: 1 53248 0.88 MB/sec execute 275 sec latency 36.129 ms oleg308-client.virtnet: 1 53542 0.88 MB/sec execute 276 sec latency 15.381 ms oleg308-client.virtnet: 1 53804 0.88 MB/sec execute 277 sec latency 15.577 ms oleg308-client.virtnet: 1 54049 0.88 MB/sec execute 278 sec latency 65.399 ms oleg308-client.virtnet: 1 54419 0.88 MB/sec execute 279 sec latency 41.953 ms oleg308-client.virtnet: 1 54893 0.89 MB/sec execute 280 sec latency 47.978 ms oleg308-client.virtnet: 1 55238 0.89 MB/sec execute 281 sec latency 39.593 ms oleg308-client.virtnet: 1 55624 0.89 MB/sec execute 282 sec latency 49.482 ms oleg308-client.virtnet: 1 56116 0.90 MB/sec execute 283 sec latency 42.028 ms oleg308-client.virtnet: 1 56789 0.92 MB/sec execute 284 sec latency 42.032 ms oleg308-client.virtnet: 1 57207 0.92 MB/sec execute 285 sec latency 15.538 ms oleg308-client.virtnet: 1 57497 0.92 MB/sec execute 286 sec latency 68.086 ms oleg308-client.virtnet: 1 57704 0.92 MB/sec execute 287 sec latency 50.723 ms oleg308-client.virtnet: 1 58018 0.92 MB/sec execute 288 sec latency 42.227 ms oleg308-client.virtnet: 1 58492 0.93 MB/sec execute 289 sec latency 52.456 ms oleg308-client.virtnet: 1 59086 0.93 MB/sec execute 290 sec latency 19.306 ms oleg308-client.virtnet: 1 59743 0.95 MB/sec execute 291 sec latency 30.549 ms oleg308-client.virtnet: 1 60356 0.96 MB/sec execute 292 sec latency 28.371 ms oleg308-client.virtnet: 1 60735 0.96 MB/sec execute 293 sec latency 14.309 ms oleg308-client.virtnet: 1 61120 0.96 MB/sec execute 294 sec latency 51.070 ms oleg308-client.virtnet: 1 61549 0.96 MB/sec execute 295 sec latency 28.232 ms oleg308-client.virtnet: 1 62093 0.97 MB/sec execute 296 sec latency 29.046 ms oleg308-client.virtnet: 1 62751 0.97 MB/sec execute 297 sec latency 33.712 ms oleg308-client.virtnet: 1 63602 0.99 MB/sec execute 298 sec latency 34.410 ms oleg308-client.virtnet: 1 64126 1.00 MB/sec execute 299 sec latency 20.926 ms oleg308-client.virtnet: 1 cleanup 300 sec oleg308-client.virtnet: 0 cleanup 300 sec oleg308-client.virtnet: oleg308-client.virtnet: Operation Count AvgLat MaxLat oleg308-client.virtnet: ---------------------------------------- oleg308-client.virtnet: NTCreateX 9688 17.851 29124.446 oleg308-client.virtnet: Close 7096 3.815 18887.367 oleg308-client.virtnet: Rename 407 7.719 17.080 oleg308-client.virtnet: Unlink 1965 2.154 6.006 oleg308-client.virtnet: Qpathinfo 8782 3.567 18786.129 oleg308-client.virtnet: Qfileinfo 1531 0.133 2.219 oleg308-client.virtnet: Qfsinfo 1608 0.103 3.267 oleg308-client.virtnet: Sfileinfo 770 28.603 19033.573 oleg308-client.virtnet: Find 3393 0.430 12.308 oleg308-client.virtnet: WriteX 4768 0.786 5.351 oleg308-client.virtnet: ReadX 15197 0.018 2.900 oleg308-client.virtnet: LockX 32 1.116 1.985 oleg308-client.virtnet: UnlockX 32 1.179 2.185 oleg308-client.virtnet: Flush 673 49.129 18963.803 oleg308-client.virtnet: oleg308-client.virtnet: Throughput 1.00492 MB/sec 1 clients 1 procs max_latency=29124.491 ms oleg308-client.virtnet: stopping dbench on /mnt/lustre/d70b.replay-single/oleg308-client.virtnet at Thu Apr 18 07:42:08 EDT 2024 with return code 0 oleg308-client.virtnet: clean dbench files on /mnt/lustre/d70b.replay-single/oleg308-client.virtnet oleg308-client.virtnet: /mnt/lustre/d70b.replay-single/oleg308-client.virtnet /mnt/lustre/d70b.replay-single/oleg308-client.virtnet oleg308-client.virtnet: removed directory: 'clients/client0/~dmtmp/WORDPRO' oleg308-client.virtnet: removed directory: 'clients/client0/~dmtmp/PWRPNT' oleg308-client.virtnet: removed directory: 'clients/client0/~dmtmp/SEED' oleg308-client.virtnet: removed directory: 'clients/client0/~dmtmp/WORD' oleg308-client.virtnet: removed directory: 'clients/client0/~dmtmp/COREL' oleg308-client.virtnet: removed directory: 'clients/client0/~dmtmp/PM' oleg308-client.virtnet: removed directory: 'clients/client0/~dmtmp/ACCESS' oleg308-client.virtnet: removed directory: 'clients/client0/~dmtmp/PARADOX' oleg308-client.virtnet: removed directory: 'clients/client0/~dmtmp/EXCEL' oleg308-client.virtnet: removed directory: 'clients/client0/~dmtmp' oleg308-client.virtnet: removed directory: 'clients/client0' oleg308-client.virtnet: removed directory: 'clients' oleg308-client.virtnet: removed 'client.txt' oleg308-client.virtnet: /mnt/lustre/d70b.replay-single/oleg308-client.virtnet oleg308-client.virtnet: dbench successfully finished PASS 70b (363s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 70c: tar 1mdts recovery ============ 07:42:11 (1713440531) Starting client oleg308-client.virtnet: -o user_xattr,flock oleg308-server@tcp:/lustre /mnt/lustre Started clients oleg308-client.virtnet: 192.168.203.108@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Started tar 27191 tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 12032 2196608 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 24576 3203072 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 14336 3202048 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 38912 6405120 1% /mnt/lustre test_70c fail mds1 1 times Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:44:25 (1713440665) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:44:39 (1713440679) targets are mounted 07:44:39 (1713440679) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 12032 2196480 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 18432 3234816 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 18432 3226624 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 36864 6461440 1% /mnt/lustre test_70c fail mds1 2 times Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:47:01 (1713440821) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:47:15 (1713440835) targets are mounted 07:47:15 (1713440835) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 70c (370s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 70d: mkdir/rmdir striped dir 1mdts recovery ========================================================== 07:48:24 (1713440904) SKIP: replay-single test_70d needs >= 2 MDTs SKIP 70d (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 70e: rename cross-MDT with random fails ========================================================== 07:48:27 (1713440907) SKIP: replay-single test_70e needs >= 2 MDTs SKIP 70e (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 70f: OSS O_DIRECT recovery with 1 clients ========================================================== 07:48:30 (1713440910) mount clients oleg308-client.virtnet ... Starting client oleg308-client.virtnet: -o user_xattr,flock oleg308-server@tcp:/lustre /mnt/lustre Started clients oleg308-client.virtnet: 192.168.203.108@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3968 2204544 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 9216 3735552 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 7168 3745792 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 16384 7481344 1% /mnt/lustre ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... test_70f failing OST 1 times Failing ost1 on oleg308-server Stopping /mnt/lustre-ost1 (opts:) on oleg308-server 07:48:38 (1713440918) shut down Failover ost1 to oleg308-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... Started lustre-OST0000 07:48:52 (1713440932) targets are mounted 07:48:52 (1713440932) facet_failover done ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3968 2204544 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 9216 3737600 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 7168 3753984 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 16384 7491584 1% /mnt/lustre ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... test_70f failing OST 2 times Failing ost1 on oleg308-server Stopping /mnt/lustre-ost1 (opts:) on oleg308-server 07:49:05 (1713440945) shut down Failover ost1 to oleg308-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Started lustre-OST0000 07:49:19 (1713440959) targets are mounted 07:49:19 (1713440959) facet_failover done Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3968 2204544 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 9216 3749888 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 7168 3752960 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 16384 7502848 1% /mnt/lustre ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... test_70f failing OST 3 times Failing ost1 on oleg308-server Stopping /mnt/lustre-ost1 (opts:) on oleg308-server 07:49:32 (1713440972) shut down Failover ost1 to oleg308-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... Started lustre-OST0000 07:49:46 (1713440986) targets are mounted 07:49:46 (1713440986) facet_failover done ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3968 2204544 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 9216 3749888 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 7168 3752960 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 16384 7502848 1% /mnt/lustre ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... test_70f failing OST 4 times Failing ost1 on oleg308-server Stopping /mnt/lustre-ost1 (opts:) on oleg308-server 07:50:00 (1713441000) shut down Failover ost1 to oleg308-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... Started lustre-OST0000 07:50:14 (1713441014) targets are mounted 07:50:14 (1713441014) facet_failover done ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 4096 2204544 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 9216 3749888 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 7168 3755008 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 16384 7504896 1% /mnt/lustre ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... test_70f failing OST 5 times Failing ost1 on oleg308-server Stopping /mnt/lustre-ost1 (opts:) on oleg308-server 07:50:27 (1713441027) shut down Failover ost1 to oleg308-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... Started lustre-OST0000 ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear 07:50:41 (1713441041) targets are mounted 07:50:41 (1713441041) facet_failover done Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... ldlm.namespaces.MGC192.168.203.108@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aab7c000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aab7c000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg308-client.virtnet' ... PASS 70f (139s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 71a: mkdir/rmdir striped dir with 2 mdts recovery ========================================================== 07:50:51 (1713441051) SKIP: replay-single test_71a needs >= 2 MDTs SKIP 71a (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 73a: open(O_CREAT), unlink, replay, reconnect before open replay, close ========================================================== 07:50:55 (1713441055) multiop /mnt/lustre/f73a.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6922 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 4096 2204544 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre fail_loc=0x80000302 Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:50:58 (1713441058) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:51:12 (1713441072) targets are mounted 07:51:12 (1713441072) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 73a (38s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 73b: open(O_CREAT), unlink, replay, reconnect at open_replay reply, close ========================================================== 07:51:34 (1713441094) multiop /mnt/lustre/f73b.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6922 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre fail_loc=0x80000157 Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:51:38 (1713441098) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:51:52 (1713441112) targets are mounted 07:51:52 (1713441112) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 73b (38s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 74: Ensure applications don't fail waiting for OST recovery ========================================================== 07:52:14 (1713441134) Stopping clients: oleg308-client.virtnet /mnt/lustre (opts:) Stopping client oleg308-client.virtnet /mnt/lustre opts: Stopping /mnt/lustre-ost1 (opts:) on oleg308-server Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:52:18 (1713441138) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:52:31 (1713441151) targets are mounted 07:52:31 (1713441151) facet_failover done Starting client oleg308-client.virtnet: -o user_xattr,flock oleg308-server@tcp:/lustre /mnt/lustre Started clients oleg308-client.virtnet: 192.168.203.108@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-OST0000 PASS 74 (22s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80a: DNE: create remote dir, drop update rep from MDT0, fail MDT0 ========================================================== 07:52:38 (1713441158) SKIP: replay-single test_80a needs >= 2 MDTs SKIP 80a (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80b: DNE: create remote dir, drop update rep from MDT0, fail MDT1 ========================================================== 07:52:41 (1713441161) SKIP: replay-single test_80b needs >= 2 MDTs SKIP 80b (2s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80c: DNE: create remote dir, drop update rep from MDT1, fail MDT[0,1] ========================================================== 07:52:46 (1713441166) SKIP: replay-single test_80c needs >= 2 MDTs SKIP 80c (2s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80d: DNE: create remote dir, drop update rep from MDT1, fail 2 MDTs ========================================================== 07:52:51 (1713441171) SKIP: replay-single test_80d needs >= 2 MDTs SKIP 80d (2s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80e: DNE: create remote dir, drop MDT1 rep, fail MDT0 ========================================================== 07:52:55 (1713441175) SKIP: replay-single test_80e needs >= 2 MDTs SKIP 80e (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80f: DNE: create remote dir, drop MDT1 rep, fail MDT1 ========================================================== 07:52:59 (1713441179) SKIP: replay-single test_80f needs >= 2 MDTs SKIP 80f (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80g: DNE: create remote dir, drop MDT1 rep, fail MDT0, then MDT1 ========================================================== 07:53:02 (1713441182) SKIP: replay-single test_80g needs >= 2 MDTs SKIP 80g (2s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80h: DNE: create remote dir, drop MDT1 rep, fail 2 MDTs ========================================================== 07:53:06 (1713441186) SKIP: replay-single test_80h needs >= 2 MDTs SKIP 80h (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81a: DNE: unlink remote dir, drop MDT0 update rep, fail MDT1 ========================================================== 07:53:09 (1713441189) SKIP: replay-single test_81a needs >= 2 MDTs SKIP 81a (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81b: DNE: unlink remote dir, drop MDT0 update reply, fail MDT0 ========================================================== 07:53:13 (1713441193) SKIP: replay-single test_81b needs >= 2 MDTs SKIP 81b (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81c: DNE: unlink remote dir, drop MDT0 update reply, fail MDT0,MDT1 ========================================================== 07:53:16 (1713441196) SKIP: replay-single test_81c needs >= 2 MDTs SKIP 81c (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81d: DNE: unlink remote dir, drop MDT0 update reply, fail 2 MDTs ========================================================== 07:53:19 (1713441199) SKIP: replay-single test_81d needs >= 2 MDTs SKIP 81d (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81e: DNE: unlink remote dir, drop MDT1 req reply, fail MDT0 ========================================================== 07:53:22 (1713441202) SKIP: replay-single test_81e needs >= 2 MDTs SKIP 81e (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81f: DNE: unlink remote dir, drop MDT1 req reply, fail MDT1 ========================================================== 07:53:25 (1713441205) SKIP: replay-single test_81f needs >= 2 MDTs SKIP 81f (2s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81g: DNE: unlink remote dir, drop req reply, fail M0, then M1 ========================================================== 07:53:29 (1713441209) SKIP: replay-single test_81g needs >= 2 MDTs SKIP 81g (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81h: DNE: unlink remote dir, drop request reply, fail 2 MDTs ========================================================== 07:53:32 (1713441212) SKIP: replay-single test_81h needs >= 2 MDTs SKIP 81h (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 84a: stale open during export disconnect ========================================================== 07:53:35 (1713441215) fail_loc=0x80000144 total: 1 open/close in 0.01 seconds: 102.72 ops/second pdsh@oleg308-client: oleg308-client: ssh exited with exit code 5 PASS 84a (6s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 85a: check the cancellation of unused locks during recovery(IBITS) ========================================================== 07:53:45 (1713441225) before recovery: unused locks count = 201 Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:53:51 (1713441231) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:54:05 (1713441245) targets are mounted 07:54:05 (1713441245) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec after recovery: unused locks count = 101 PASS 85a (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 85b: check the cancellation of unused locks during recovery(EXTENT) ========================================================== 07:54:17 (1713441257) before recovery: unused locks count = 100 Failing ost1 on oleg308-server Stopping /mnt/lustre-ost1 (opts:) on oleg308-server 07:54:28 (1713441268) shut down Failover ost1 to oleg308-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-OST0000 07:54:45 (1713441285) targets are mounted 07:54:45 (1713441285) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec after recovery: unused locks count = 0 PASS 85b (35s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 86: umount server after clear nid_stats should not hit LBUG ========================================================== 07:54:56 (1713441296) Stopping clients: oleg308-client.virtnet /mnt/lustre (opts:) Stopping client oleg308-client.virtnet /mnt/lustre opts: mdt.lustre-MDT0000.exports.clear=0 Stopping /mnt/lustre-mds1 (opts:) on oleg308-server Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 Starting client oleg308-client.virtnet: -o user_xattr,flock oleg308-server@tcp:/lustre /mnt/lustre Started clients oleg308-client.virtnet: 192.168.203.108@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) PASS 86 (10s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 87a: write replay ================== 07:55:08 (1713441308) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7530496 1% /mnt/lustre 8+0 records in 8+0 records out 8388608 bytes (8.4 MB) copied, 0.138613 s, 60.5 MB/s Failing ost1 on oleg308-server Stopping /mnt/lustre-ost1 (opts:) on oleg308-server 07:55:11 (1713441311) shut down Failover ost1 to oleg308-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-OST0000 07:55:27 (1713441327) targets are mounted 07:55:27 (1713441327) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec 8+0 records in 8+0 records out 8388608 bytes (8.4 MB) copied, 0.147675 s, 56.8 MB/s PASS 87a (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 87b: write replay with changed data (checksum resend) ========================================================== 07:55:41 (1713441341) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3968 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 12288 3757056 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 16384 7522304 1% /mnt/lustre 8+0 records in 8+0 records out 8388608 bytes (8.4 MB) copied, 0.252005 s, 33.3 MB/s 8+0 records in 8+0 records out 8 bytes (8 B) copied, 0.00646988 s, 1.2 kB/s Failing ost1 on oleg308-server Stopping /mnt/lustre-ost1 (opts:) on oleg308-server 07:55:48 (1713441348) shut down Failover ost1 to oleg308-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-OST0000 07:56:03 (1713441363) targets are mounted 07:56:03 (1713441363) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec 0+1 records in 0+1 records out 72 bytes (72 B) copied, 0.00637844 s, 11.3 kB/s PASS 87b (29s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 88: MDS should not assign same objid to different files ========================================================== 07:56:10 (1713441370) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 13312 3756032 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 17408 7521280 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 13312 3756032 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 17408 7521280 1% /mnt/lustre before test: last_id = 11969, next_id = 11940 Creating to objid 11969 on ost lustre-OST0000... total: 31 open/close in 0.14 seconds: 229.56 ops/second total: 8 open/close in 0.02 seconds: 337.24 ops/second before recovery: last_id = 12065, next_id = 11978 Stopping /mnt/lustre-mds1 (opts:) on oleg308-server Stopping /mnt/lustre-ost1 (opts:) on oleg308-server Failover mds1 to oleg308-server Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 Failover ost1 to oleg308-server Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-OST0000 after recovery: last_id = 12009, next_id = 11978 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0751495 s, 7.0 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0916817 s, 5.7 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0624084 s, 8.4 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0959738 s, 5.5 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0969803 s, 5.4 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0486255 s, 10.8 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0866528 s, 6.1 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0456945 s, 11.5 MB/s -rw-r--r-- 1 root root 0 Apr 18 07:56 /mnt/lustre/d88.replay-single/f-11940 -rw-r--r-- 1 root root 0 Apr 18 07:56 /mnt/lustre/d88.replay-single/f-11941 -rw-r--r-- 1 root root 0 Apr 18 07:56 /mnt/lustre/d88.replay-single/f-11942 -rw-r--r-- 1 root root 0 Apr 18 07:56 /mnt/lustre/d88.replay-single/f-11943 -rw-r--r-- 1 root root 0 Apr 18 07:56 /mnt/lustre/d88.replay-single/f-11944 -rw-r--r-- 1 root root 0 Apr 18 07:56 /mnt/lustre/d88.replay-single/f-11945 -rw-r--r-- 1 root root 0 Apr 18 07:56 /mnt/lustre/d88.replay-single/f-11946 -rw-r--r-- 1 root root 0 Apr 18 07:56 /mnt/lustre/d88.replay-single/f-11947 -rw-r--r-- 1 root root 0 Apr 18 07:56 /mnt/lustre/d88.replay-single/f-11948 -rw-r--r-- 1 root root 0 Apr 18 07:56 /mnt/lustre/d88.replay-single/f-11949 -rw-r--r-- 1 root root 0 Apr 18 07:56 /mnt/lustre/d88.replay-single/f-11950 -rw-r--r-- 1 root root 0 Apr 18 07:56 /mnt/lustre/d88.replay-single/f-11951 -rw-r--r-- 1 root root 0 Apr 18 07:56 /mnt/lustre/d88.replay-single/f-11952 -rw-r--r-- 1 root root 0 Apr 18 07:56 /mnt/lustre/d88.replay-single/f-11953 -rw-r--r-- 1 root root 0 Apr 18 07:56 /mnt/lustre/d88.replay-single/f-11954 -rw-r--r-- 1 root root 0 Apr 18 07:56 /mnt/lustre/d88.replay-single/f-11955 -rw-r--r-- 1 root root 0 Apr 18 07:56 /mnt/lustre/d88.replay-single/f-11956 -rw-r--r-- 1 root root 0 Apr 18 07:56 /mnt/lustre/d88.replay-single/f-11957 -rw-r--r-- 1 root root 0 Apr 18 07:56 /mnt/lustre/d88.replay-single/f-11958 -rw-r--r-- 1 root root 0 Apr 18 07:56 /mnt/lustre/d88.replay-single/f-11959 -rw-r--r-- 1 root root 0 Apr 18 07:56 /mnt/lustre/d88.replay-single/f-11960 -rw-r--r-- 1 root root 0 Apr 18 07:56 /mnt/lustre/d88.replay-single/f-11961 -rw-r--r-- 1 root root 0 Apr 18 07:56 /mnt/lustre/d88.replay-single/f-11962 -rw-r--r-- 1 root root 0 Apr 18 07:56 /mnt/lustre/d88.replay-single/f-11963 -rw-r--r-- 1 root root 0 Apr 18 07:56 /mnt/lustre/d88.replay-single/f-11964 -rw-r--r-- 1 root root 0 Apr 18 07:56 /mnt/lustre/d88.replay-single/f-11965 -rw-r--r-- 1 root root 0 Apr 18 07:56 /mnt/lustre/d88.replay-single/f-11966 -rw-r--r-- 1 root root 0 Apr 18 07:56 /mnt/lustre/d88.replay-single/f-11967 -rw-r--r-- 1 root root 0 Apr 18 07:56 /mnt/lustre/d88.replay-single/f-11968 -rw-r--r-- 1 root root 0 Apr 18 07:56 /mnt/lustre/d88.replay-single/f-11969 -rw-r--r-- 1 root root 0 Apr 18 07:56 /mnt/lustre/d88.replay-single/f-11970 -rw-r--r-- 1 root root 0 Apr 18 07:56 /mnt/lustre/d88.replay-single/f-11971 -rw-r--r-- 1 root root 0 Apr 18 07:56 /mnt/lustre/d88.replay-single/f-11972 -rw-r--r-- 1 root root 0 Apr 18 07:56 /mnt/lustre/d88.replay-single/f-11973 -rw-r--r-- 1 root root 0 Apr 18 07:56 /mnt/lustre/d88.replay-single/f-11974 -rw-r--r-- 1 root root 0 Apr 18 07:56 /mnt/lustre/d88.replay-single/f-11975 -rw-r--r-- 1 root root 0 Apr 18 07:56 /mnt/lustre/d88.replay-single/f-11976 -rw-r--r-- 1 root root 0 Apr 18 07:56 /mnt/lustre/d88.replay-single/f-11977 -rw-r--r-- 1 root root 524288 Apr 18 07:57 /mnt/lustre/d88.replay-single/f-11981 -rw-r--r-- 1 root root 524288 Apr 18 07:57 /mnt/lustre/d88.replay-single/f-11982 -rw-r--r-- 1 root root 524288 Apr 18 07:57 /mnt/lustre/d88.replay-single/f-11983 -rw-r--r-- 1 root root 524288 Apr 18 07:57 /mnt/lustre/d88.replay-single/f-11984 -rw-r--r-- 1 root root 524288 Apr 18 07:57 /mnt/lustre/d88.replay-single/f-11985 -rw-r--r-- 1 root root 524288 Apr 18 07:57 /mnt/lustre/d88.replay-single/f-11986 -rw-r--r-- 1 root root 524288 Apr 18 07:57 /mnt/lustre/d88.replay-single/f-11987 -rw-r--r-- 1 root root 524288 Apr 18 07:57 /mnt/lustre/d88.replay-single/f-11988 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0299008 s, 17.5 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0348008 s, 15.1 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0388094 s, 13.5 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0609941 s, 8.6 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.104825 s, 5.0 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0576712 s, 9.1 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0494115 s, 10.6 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0472102 s, 11.1 MB/s PASS 88 (61s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 89: no disk space leak on late ost connection ========================================================== 07:57:15 (1713441435) Waiting for orphan cleanup... osp.lustre-OST0000-osc-MDT0000.old_sync_processed osp.lustre-OST0001-osc-MDT0000.old_sync_processed wait 40 secs maximumly for oleg308-server mds-ost sync done. sleep 5 for ZFS zfs Waiting for MDT destroys to complete 10+0 records in 10+0 records out 41943040 bytes (42 MB) copied, 0.528592 s, 79.3 MB/s Stopping /mnt/lustre-ost1 (opts:) on oleg308-server Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 07:57:32 (1713441452) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 07:57:48 (1713441468) targets are mounted 07:57:48 (1713441468) facet_failover done Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-OST0000 Starting client: oleg308-client.virtnet: -o user_xattr,flock oleg308-server@tcp:/lustre /mnt/lustre osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 66 sec Waiting for orphan cleanup... osp.lustre-OST0000-osc-MDT0000.old_sync_processed osp.lustre-OST0001-osc-MDT0000.old_sync_processed wait 40 secs maximumly for oleg308-server mds-ost sync done. sleep 5 for ZFS zfs sleep 5 for ZFS zfs Waiting for MDT destroys to complete free_before: 7517184 free_after: 7517184 PASS 89 (132s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 90: lfs find identifies the missing striped file segments ========================================================== 07:59:30 (1713441570) Create the files Fail ost1 lustre-OST0000_UUID, display the list of affected files Stopping /mnt/lustre-ost1 (opts:) on oleg308-server General Query: lfs find /mnt/lustre/d90.replay-single /mnt/lustre/d90.replay-single /mnt/lustre/d90.replay-single/f1 /mnt/lustre/d90.replay-single/all /mnt/lustre/d90.replay-single/f0 Querying files on shutdown ost1: lfs find --obd lustre-OST0000_UUID /mnt/lustre/d90.replay-single/all /mnt/lustre/d90.replay-single/f0 Check getstripe: /home/green/git/lustre-release/lustre/utils/lfs getstripe -r --obd lustre-OST0000_UUID /mnt/lustre/d90.replay-single/all lmm_stripe_count: 2 lmm_stripe_size: 4194304 lmm_pattern: raid0 lmm_layout_gen: 0 lmm_stripe_offset: 1 obdidx objid objid group 0 12010 0x2eea 0x240000400 * /mnt/lustre/d90.replay-single/f0 lmm_stripe_count: 1 lmm_stripe_size: 4194304 lmm_pattern: raid0 lmm_layout_gen: 0 lmm_stripe_offset: 0 obdidx objid objid group 0 12011 0x2eeb 0x240000400 * /mnt/lustre/d90.replay-single/all /mnt/lustre/d90.replay-single/f0 Failover ost1 to oleg308-server Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-OST0000 PASS 90 (21s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 93a: replay + reconnect ============ 07:59:53 (1713441593) 1+0 records in 1+0 records out 1024 bytes (1.0 kB) copied, 0.00164171 s, 624 kB/s fail_val=40 fail_loc=0x715 Failing ost1 on oleg308-server Stopping /mnt/lustre-ost1 (opts:) on oleg308-server 07:59:57 (1713441597) shut down Failover ost1 to oleg308-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-OST0000 08:00:10 (1713441610) targets are mounted 08:00:10 (1713441610) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec PASS 93a (64s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 93b: replay + reconnect on mds ===== 08:01:01 (1713441661) total: 20 open/close in 0.30 seconds: 67.61 ops/second fail_val=80 fail_loc=0x715 Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 08:01:07 (1713441667) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 08:01:24 (1713441684) targets are mounted 08:01:24 (1713441684) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 93b (107s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 100a: DNE: create striped dir, drop update rep from MDT1, fail MDT1 ========================================================== 08:02:50 (1713441770) SKIP: replay-single test_100a needs >= 2 MDTs SKIP 100a (2s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 100b: DNE: create striped dir, fail MDT0 ========================================================== 08:02:54 (1713441774) SKIP: replay-single test_100b needs >= 2 MDTs SKIP 100b (2s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 100c: DNE: create striped dir, abort_recov_mdt mds2 ========================================================== 08:02:59 (1713441779) SKIP: replay-single test_100c needs >= 2 MDTs SKIP 100c (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 100d: DNE: cancel update logs upon recovery abort ========================================================== 08:03:02 (1713441782) SKIP: replay-single test_100d needs > 1 MDTs SKIP 100d (2s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 100e: DNE: create striped dir on MDT0 and MDT1, fail MDT0, MDT1 ========================================================== 08:03:06 (1713441786) SKIP: replay-single test_100e needs >= 2 MDTs SKIP 100e (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 101: Shouldn't reassign precreated objs to other files after recovery ========================================================== 08:03:10 (1713441790) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3968 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 21504 3747840 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 25600 7513088 1% /mnt/lustre Stopping /mnt/lustre-mds1 (opts:) on oleg308-server Failover mds1 to oleg308-server Starting mds1: -o localrecov -o abort_recovery lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 pdsh@oleg308-client: oleg308-client: ssh exited with exit code 5 first stat failed: 5 PASS 101 (32s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 102a: check resend (request lost) with multiple modify RPCs in flight ========================================================== 08:03:44 (1713441824) creating 7 files ... fail_loc=0x159 launch 7 chmod in parallel (08:03:44) ... fail_loc=0 done (08:04:00) /mnt/lustre/d102a.replay-single/file-1 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-2 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-3 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-4 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-5 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-6 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-7 has perms 0600 OK PASS 102a (18s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 102b: check resend (reply lost) with multiple modify RPCs in flight ========================================================== 08:04:03 (1713441843) creating 7 files ... fail_loc=0x15a launch 7 chmod in parallel (08:04:04) ... fail_loc=0 done (08:04:20) /mnt/lustre/d102b.replay-single/file-1 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-2 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-3 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-4 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-5 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-6 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-7 has perms 0600 OK PASS 102b (18s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 102c: check replay w/o reconstruction with multiple mod RPCs in flight ========================================================== 08:04:23 (1713441863) creating 7 files ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3968 2204544 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 21504 3333120 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3350528 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 25600 6683648 1% /mnt/lustre fail_loc=0x15a launch 7 chmod in parallel (08:04:24) ... fail_loc=0 Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 08:04:27 (1713441867) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 08:04:40 (1713441880) targets are mounted 08:04:40 (1713441880) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec done (08:04:43) /mnt/lustre/d102c.replay-single/file-1 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-2 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-3 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-4 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-5 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-6 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-7 has perms 0600 OK PASS 102c (22s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 102d: check replay & reconstruction with multiple mod RPCs in flight ========================================================== 08:04:47 (1713441887) creating 7 files ... fail_loc=0x15a launch 7 chmod in parallel (08:04:48) ... fail_loc=0 Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 08:04:51 (1713441891) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 08:05:04 (1713441904) targets are mounted 08:05:04 (1713441904) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec done (08:05:07) /mnt/lustre/d102d.replay-single/file-1 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-2 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-3 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-4 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-5 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-6 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-7 has perms 0600 OK PASS 102d (22s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 103: Check otr_next_id overflow ==== 08:05:11 (1713441911) fail_loc=0x80000162 total: 30 open/close in 0.17 seconds: 180.33 ops/second Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 08:05:14 (1713441914) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 08:05:28 (1713441928) targets are mounted 08:05:28 (1713441928) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 103 (22s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 110a: DNE: create striped dir, fail MDT1 ========================================================== 08:05:36 (1713441936) SKIP: replay-single test_110a needs >= 2 MDTs SKIP 110a (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 110b: DNE: create striped dir, fail MDT1 and client ========================================================== 08:05:39 (1713441939) SKIP: replay-single test_110b needs >= 2 MDTs SKIP 110b (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 110c: DNE: create striped dir, fail MDT2 ========================================================== 08:05:42 (1713441942) SKIP: replay-single test_110c needs >= 2 MDTs SKIP 110c (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 110d: DNE: create striped dir, fail MDT2 and client ========================================================== 08:05:46 (1713441946) SKIP: replay-single test_110d needs >= 2 MDTs SKIP 110d (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 110e: DNE: create striped dir, uncommit on MDT2, fail client/MDT1/MDT2 ========================================================== 08:05:49 (1713441949) SKIP: replay-single test_110e needs >= 2 MDTs SKIP 110e (1s) debug_raw_pointers=0 debug_raw_pointers=0 SKIP: replay-single test_110f skipping excluded test 110f debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 110g: DNE: create striped dir, uncommit on MDT1, fail client/MDT1/MDT2 ========================================================== 08:05:53 (1713441953) SKIP: replay-single test_110g needs >= 2 MDTs SKIP 110g (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 111a: DNE: unlink striped dir, fail MDT1 ========================================================== 08:05:57 (1713441957) SKIP: replay-single test_111a needs >= 2 MDTs SKIP 111a (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 111b: DNE: unlink striped dir, fail MDT2 ========================================================== 08:06:00 (1713441960) SKIP: replay-single test_111b needs >= 2 MDTs SKIP 111b (2s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 111c: DNE: unlink striped dir, uncommit on MDT1, fail client/MDT1/MDT2 ========================================================== 08:06:04 (1713441964) SKIP: replay-single test_111c needs >= 2 MDTs SKIP 111c (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 111d: DNE: unlink striped dir, uncommit on MDT2, fail client/MDT1/MDT2 ========================================================== 08:06:07 (1713441967) SKIP: replay-single test_111d needs >= 2 MDTs SKIP 111d (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 111e: DNE: unlink striped dir, uncommit on MDT2, fail MDT1/MDT2 ========================================================== 08:06:11 (1713441971) SKIP: replay-single test_111e needs >= 2 MDTs SKIP 111e (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 111f: DNE: unlink striped dir, uncommit on MDT1, fail MDT1/MDT2 ========================================================== 08:06:14 (1713441974) SKIP: replay-single test_111f needs >= 2 MDTs SKIP 111f (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 111g: DNE: unlink striped dir, fail MDT1/MDT2 ========================================================== 08:06:18 (1713441978) SKIP: replay-single test_111g needs >= 2 MDTs SKIP 111g (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112a: DNE: cross MDT rename, fail MDT1 ========================================================== 08:06:21 (1713441981) SKIP: replay-single test_112a needs >= 4 MDTs SKIP 112a (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112b: DNE: cross MDT rename, fail MDT2 ========================================================== 08:06:25 (1713441985) SKIP: replay-single test_112b needs >= 4 MDTs SKIP 112b (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112c: DNE: cross MDT rename, fail MDT3 ========================================================== 08:06:28 (1713441988) SKIP: replay-single test_112c needs >= 4 MDTs SKIP 112c (2s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112d: DNE: cross MDT rename, fail MDT4 ========================================================== 08:06:32 (1713441992) SKIP: replay-single test_112d needs >= 4 MDTs SKIP 112d (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112e: DNE: cross MDT rename, fail MDT1 and MDT2 ========================================================== 08:06:35 (1713441995) SKIP: replay-single test_112e needs >= 4 MDTs SKIP 112e (2s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112f: DNE: cross MDT rename, fail MDT1 and MDT3 ========================================================== 08:06:39 (1713441999) SKIP: replay-single test_112f needs >= 4 MDTs SKIP 112f (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112g: DNE: cross MDT rename, fail MDT1 and MDT4 ========================================================== 08:06:41 (1713442001) SKIP: replay-single test_112g needs >= 4 MDTs SKIP 112g (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112h: DNE: cross MDT rename, fail MDT2 and MDT3 ========================================================== 08:06:44 (1713442004) SKIP: replay-single test_112h needs >= 4 MDTs SKIP 112h (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112i: DNE: cross MDT rename, fail MDT2 and MDT4 ========================================================== 08:06:46 (1713442006) SKIP: replay-single test_112i needs >= 4 MDTs SKIP 112i (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112j: DNE: cross MDT rename, fail MDT3 and MDT4 ========================================================== 08:06:49 (1713442009) SKIP: replay-single test_112j needs >= 4 MDTs SKIP 112j (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112k: DNE: cross MDT rename, fail MDT1,MDT2,MDT3 ========================================================== 08:06:51 (1713442011) SKIP: replay-single test_112k needs >= 4 MDTs SKIP 112k (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112l: DNE: cross MDT rename, fail MDT1,MDT2,MDT4 ========================================================== 08:06:54 (1713442014) SKIP: replay-single test_112l needs >= 4 MDTs SKIP 112l (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112m: DNE: cross MDT rename, fail MDT1,MDT3,MDT4 ========================================================== 08:06:57 (1713442017) SKIP: replay-single test_112m needs >= 4 MDTs SKIP 112m (0s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112n: DNE: cross MDT rename, fail MDT2,MDT3,MDT4 ========================================================== 08:07:00 (1713442020) SKIP: replay-single test_112n needs >= 4 MDTs SKIP 112n (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 115: failover for create/unlink striped directory ========================================================== 08:07:02 (1713442022) SKIP: replay-single test_115 needs >= 2 MDTs SKIP 115 (2s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 116a: large update log master MDT recovery ========================================================== 08:07:06 (1713442026) SKIP: replay-single test_116a needs >= 2 MDTs SKIP 116a (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 116b: large update log slave MDT recovery ========================================================== 08:07:09 (1713442029) SKIP: replay-single test_116b needs >= 2 MDTs SKIP 116b (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 117: DNE: cross MDT unlink, fail MDT1 and MDT2 ========================================================== 08:07:12 (1713442032) SKIP: replay-single test_117 needs >= 4 MDTs SKIP 117 (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 118: invalidate osp update will not cause update log corruption ========================================================== 08:07:15 (1713442035) SKIP: replay-single test_118 needs >= 2 MDTs SKIP 118 (2s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 119: timeout of normal replay does not cause DNE replay fails ========================================================== 08:07:18 (1713442038) SKIP: replay-single test_119 needs >= 2 MDTs SKIP 119 (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 120: DNE fail abort should stop both normal and DNE replay ========================================================== 08:07:20 (1713442040) SKIP: replay-single test_120 needs >= 2 MDTs SKIP 120 (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 121: lock replay timed out and race ========================================================== 08:07:22 (1713442042) multiop /mnt/lustre/f121.replay-single vs_s TMPPIPE=/tmp/multiop_open_wait_pipe.6922 Stopping /mnt/lustre-mds1 (opts:) on oleg308-server Failover mds1 to oleg308-server fail_loc=0x721 fail_val=0 at_max=0 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 fail_loc=0x0 at_max=600 PASS 121 (27s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 130a: DoM file create (setstripe) replay ========================================================== 08:07:51 (1713442071) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3968 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 21504 3333120 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 25600 7098368 1% /mnt/lustre Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 08:07:55 (1713442075) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 08:08:09 (1713442089) targets are mounted 08:08:09 (1713442089) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 130a (23s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 130b: DoM file create (inherited) replay ========================================================== 08:08:16 (1713442096) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3968 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 21504 3333120 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 25600 7098368 1% /mnt/lustre Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 08:08:19 (1713442099) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 08:08:32 (1713442112) targets are mounted 08:08:32 (1713442112) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 130b (22s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 131a: DoM file write lock replay === 08:08:40 (1713442120) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3968 2204544 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 21504 3333120 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 25600 7098368 1% /mnt/lustre 1+0 records in 1+0 records out 8 bytes (8 B) copied, 0.00245916 s, 3.3 kB/s Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 08:08:44 (1713442124) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 08:08:58 (1713442138) targets are mounted 08:08:58 (1713442138) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 131a (24s) debug_raw_pointers=0 debug_raw_pointers=0 SKIP: replay-single test_131b skipping excluded test 131b debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 132a: PFL new component instantiate replay ========================================================== 08:09:07 (1713442147) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 4096 2204544 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 21504 3333120 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 25600 7098368 1% /mnt/lustre 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0238814 s, 43.9 MB/s /mnt/lustre/f132a.replay-single lcm_layout_gen: 3 lcm_mirror_count: 1 lcm_entry_count: 2 lcme_id: 1 lcme_mirror_id: 0 lcme_flags: init lcme_extent.e_start: 0 lcme_extent.e_end: 1048576 lmm_stripe_count: 1 lmm_stripe_size: 1048576 lmm_pattern: raid0 lmm_layout_gen: 0 lmm_stripe_offset: 0 lmm_objects: - 0: { l_ost_idx: 0, l_fid: [0x240000400:0x340a:0x0] } lcme_id: 2 lcme_mirror_id: 0 lcme_flags: init lcme_extent.e_start: 1048576 lcme_extent.e_end: EOF lmm_stripe_count: 2 lmm_stripe_size: 4194304 lmm_pattern: raid0 lmm_layout_gen: 0 lmm_stripe_offset: 1 lmm_objects: - 0: { l_ost_idx: 1, l_fid: [0x280000400:0x32e2:0x0] } - 1: { l_ost_idx: 0, l_fid: [0x240000400:0x340b:0x0] } Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 08:09:10 (1713442150) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 08:09:24 (1713442164) targets are mounted 08:09:24 (1713442164) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f132a.replay-single lcm_layout_gen: 1 lcm_mirror_count: 1 lcm_entry_count: 2 lcme_id: 1 lcme_mirror_id: 0 lcme_flags: init lcme_extent.e_start: 0 lcme_extent.e_end: 1048576 lmm_stripe_count: 1 lmm_stripe_size: 1048576 lmm_pattern: raid0 lmm_layout_gen: 65535 lmm_stripe_offset: 0 lmm_objects: - 0: { l_ost_idx: 0, l_fid: [0x240000400:0x340a:0x0] } lcme_id: 2 lcme_mirror_id: 0 lcme_flags: init lcme_extent.e_start: 1048576 lcme_extent.e_end: EOF lmm_stripe_count: 2 lmm_stripe_size: 4194304 lmm_pattern: raid0 lmm_layout_gen: 65535 lmm_stripe_offset: 1 lmm_objects: - 0: { l_ost_idx: 1, l_fid: [0x280000400:0x32e2:0x0] } - 1: { l_ost_idx: 0, l_fid: [0x240000400:0x340b:0x0] } PASS 132a (22s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 133: check resend of ongoing requests for lwp during failover ========================================================== 08:09:31 (1713442171) SKIP: replay-single test_133 needs >= 2 MDTs SKIP 133 (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 134: replay creation of a file created in a pool ========================================================== 08:09:34 (1713442174) Creating new pool oleg308-server: Pool lustre.pool_134 created Adding targets to pool oleg308-server: OST lustre-OST0001_UUID added to pool lustre.pool_134 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 4096 2204544 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 21504 3747840 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 5120 3764224 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 26624 7512064 1% /mnt/lustre Failing mds1 on oleg308-server Stopping /mnt/lustre-mds1 (opts:) on oleg308-server 08:09:43 (1713442183) shut down Failover mds1 to oleg308-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-MDT0000 08:09:57 (1713442197) targets are mounted 08:09:57 (1713442197) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Destroy the created pools: pool_134 lustre.pool_134 oleg308-server: OST lustre-OST0001_UUID removed from pool lustre.pool_134 oleg308-server: Pool lustre.pool_134 destroyed Waiting 90s for 'foo' PASS 134 (35s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 135: Server failure in lock replay phase ========================================================== 08:10:11 (1713442211) Failing ost1 on oleg308-server Stopping /mnt/lustre-ost1 (opts:) on oleg308-server 08:10:13 (1713442213) shut down Failover ost1 to oleg308-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-OST0000 08:10:27 (1713442227) targets are mounted 08:10:27 (1713442227) facet_failover done oleg308-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 4096 2204544 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 21504 3747840 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 5120 3764224 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 26624 7512064 1% /mnt/lustre ldlm.cancel_unused_locks_before_replay=0 Stopping /mnt/lustre-ost1 (opts:) on oleg308-server Failover ost1 to oleg308-server oleg308-server: oleg308-server.virtnet: executing load_module ../libcfs/libcfs/libcfs fail_loc=0x32d fail_val=20 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-OST0000 oleg308-client.virtnet: executing wait_import_state_mount REPLAY_LOCKS osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in REPLAY_LOCKS state after 0 sec Stopping /mnt/lustre-ost1 (opts:) on oleg308-server Failover ost1 to oleg308-server oleg308-server: oleg308-server.virtnet: executing load_module ../libcfs/libcfs/libcfs fail_loc=0 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 End of sync oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-OST0000 Stopping /mnt/lustre-ost1 (opts:-f) on oleg308-server Stopping /mnt/lustre-ost2 (opts:-f) on oleg308-server Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-OST0000 Starting ost2: -o localrecov lustre-ost2/ost2 /mnt/lustre-ost2 seq.cli-lustre-OST0001-super.width=65536 oleg308-server: oleg308-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg308-client: oleg308-server: ssh exited with exit code 1 Started lustre-OST0001 pdsh@oleg308-client: oleg308-client: ssh exited with exit code 5 ldlm.cancel_unused_locks_before_replay=1 PASS 135 (93s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 136: MDS to disconnect all OSPs first, then cleanup ldlm ========================================================== 08:11:46 (1713442306) SKIP: replay-single test_136 needs > 2 MDTs SKIP 136 (2s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 200: Dropping one OBD_PING should not cause disconnect ========================================================== 08:11:50 (1713442310) SKIP: replay-single test_200 Need remote client SKIP 200 (1s) debug_raw_pointers=0 debug_raw_pointers=0 == replay-single test complete, duration 5028 sec ======== 08:11:52 (1713442312) === replay-single: start cleanup 08:11:52 (1713442312) === === replay-single: finish cleanup 08:11:55 (1713442315) ===