-----============= acceptance-small: replay-single ============----- Wed Apr 17 17:01:25 EDT 2024 excepting tests: 110f 131b 59 oleg428-client.virtnet: executing check_config_client /mnt/lustre oleg428-client.virtnet: Checking config lustre mounted on /mnt/lustre Checking servers environments Checking clients oleg428-client.virtnet environments Using TIMEOUT=20 osc.lustre-OST0000-osc-ffff8800b0452800.idle_timeout=debug osc.lustre-OST0001-osc-ffff8800b0452800.idle_timeout=debug disable quota as required oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 == replay-single test 0a: empty replay =================== 17:01:35 (1713387695) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 0a (25s) == replay-single test 0b: ensure object created after recover exists. (3284) ========================================================== 17:02:00 (1713387720) Failing ost1 on oleg428-server Stopping /mnt/lustre-ost1 (opts:) on oleg428-server reboot facets: ost1 Failover ost1 to oleg428-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-OST0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec total: 20 open/close in 0.05 seconds: 433.67 ops/second - unlinked 0 (time 1713387738 ; total 0 ; last 0) total: 20 unlinks in 0 seconds: inf unlinks/second PASS 0b (19s) == replay-single test 0c: check replay-barrier =========== 17:02:19 (1713387739) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 Starting client: oleg428-client.virtnet: -o user_xattr,flock oleg428-server@tcp:/lustre /mnt/lustre rm: cannot remove '/mnt/lustre/f0c.replay-single': No such file or directory PASS 0c (81s) == replay-single test 0d: expired recovery with no clients ========================================================== 17:03:40 (1713387820) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 Starting client: oleg428-client.virtnet: -o user_xattr,flock oleg428-server@tcp:/lustre /mnt/lustre PASS 0d (77s) == replay-single test 1: simple create =================== 17:04:57 (1713387897) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f1.replay-single has type file OK PASS 1 (31s) == replay-single test 2a: touch ========================== 17:05:28 (1713387928) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f2a.replay-single has type file OK PASS 2a (26s) == replay-single test 2b: touch ========================== 17:05:54 (1713387954) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f2b.replay-single has type file OK PASS 2b (31s) == replay-single test 2c: setstripe replay =============== 17:06:25 (1713387985) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f2c.replay-single has type file OK PASS 2c (33s) == replay-single test 2d: setdirstripe replay ============ 17:06:58 (1713388018) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d2d.replay-single has type dir OK PASS 2d (26s) == replay-single test 2e: O_CREAT|O_EXCL create replay === 17:07:24 (1713388044) fail_loc=0x8000013b UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 Succeed in opening file "/mnt/lustre/f2e.replay-single"(flags=O_CREAT) oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f2e.replay-single has type file OK PASS 2e (36s) == replay-single test 3a: replay failed open(O_DIRECTORY) ========================================================== 17:08:00 (1713388080) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Error in opening file "/mnt/lustre/f3a.replay-single"(flags=O_DIRECTORY) 20: Not a directory Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f3a.replay-single has type file OK PASS 3a (34s) == replay-single test 3b: replay failed open -ENOMEM ===== 17:08:34 (1713388114) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre fail_loc=0x80000114 touch: cannot touch '/mnt/lustre/f3b.replay-single': Cannot allocate memory fail_loc=0 Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/f3b.replay-single: No such file or directory PASS 3b (33s) == replay-single test 3c: replay failed open -ENOMEM ===== 17:09:07 (1713388147) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre fail_loc=0x80000128 touch: cannot touch '/mnt/lustre/f3c.replay-single': Cannot allocate memory fail_loc=0 Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/f3c.replay-single: No such file or directory PASS 3c (35s) == replay-single test 4a: |x| 10 open(O_CREAT)s ========== 17:09:42 (1713388182) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 4a (23s) == replay-single test 4b: |x| rm 10 files ================ 17:10:05 (1713388205) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3761152 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3761152 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7522304 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/f4b.replay-single-*: No such file or directory PASS 4b (34s) == replay-single test 5: |x| 220 open(O_CREAT) =========== 17:10:39 (1713388239) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 5 (40s) == replay-single test 6a: mkdir + contained create ======= 17:11:19 (1713388279) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 7168 7531520 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d6a.replay-single has type dir OK /mnt/lustre/d6a.replay-single/f6a.replay-single has type file OK PASS 6a (35s) == replay-single test 6b: |X| rmdir ====================== 17:11:54 (1713388314) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/d6b.replay-single: No such file or directory PASS 6b (32s) == replay-single test 7: mkdir |X| contained create ====== 17:12:26 (1713388346) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d7.replay-single has type dir OK /mnt/lustre/d7.replay-single/f7.replay-single has type file OK PASS 7 (35s) == replay-single test 8: creat open |X| close ============ 17:13:01 (1713388381) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre multiop /mnt/lustre/f8.replay-single vmo_c TMPPIPE=/tmp/multiop_open_wait_pipe.6853 Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f8.replay-single /mnt/lustre/f8.replay-single has type file OK PASS 8 (33s) == replay-single test 9: |X| create (same inum/gen) ====== 17:13:34 (1713388414) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec old_inum == 144115305935798546, new_inum == 144115305935798546 old_inum and new_inum match PASS 9 (34s) == replay-single test 10: create |X| rename unlink ======= 17:14:08 (1713388448) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/f10.replay-single: No such file or directory PASS 10 (35s) == replay-single test 11: create open write rename |X| create-old-name read ========================================================== 17:14:43 (1713388483) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre new old Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec new old PASS 11 (26s) == replay-single test 12: open, unlink |X| close ========= 17:15:09 (1713388509) multiop /mnt/lustre/f12.replay-single vo_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6853 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 12 (25s) == replay-single test 13: open chmod 0 |x| write close === 17:15:35 (1713388535) multiop /mnt/lustre/f13.replay-single vO_wc TMPPIPE=/tmp/multiop_open_wait_pipe.6853 /mnt/lustre/f13.replay-single has perms 00 OK UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f13.replay-single has perms 00 OK /mnt/lustre/f13.replay-single has size 1 OK PASS 13 (25s) == replay-single test 14: open(O_CREAT), unlink |X| close ========================================================== 17:16:00 (1713388560) multiop /mnt/lustre/f14.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6853 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 14 (28s) == replay-single test 15: open(O_CREAT), unlink |X| touch new, close ========================================================== 17:16:28 (1713388588) multiop /mnt/lustre/f15.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6853 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 15 (26s) == replay-single test 16: |X| open(O_CREAT), unlink, touch new, unlink new ========================================================== 17:16:54 (1713388614) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 16 (34s) == replay-single test 17: |X| open(O_CREAT), |replay| close ========================================================== 17:17:28 (1713388648) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre multiop /mnt/lustre/f17.replay-single vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.6853 Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f17.replay-single has type file OK PASS 17 (25s) == replay-single test 18: open(O_CREAT), unlink, touch new, close, touch, unlink ========================================================== 17:17:53 (1713388673) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre multiop /mnt/lustre/f18.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6853 pid: 21472 will close Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 18 (36s) == replay-single test 19: mcreate, open, write, rename === 17:18:29 (1713388709) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre old Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec old PASS 19 (26s) == replay-single test 20a: |X| open(O_CREAT), unlink, replay, close (test mds_cleanup_orphans) ========================================================== 17:18:55 (1713388735) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre multiop /mnt/lustre/f20a.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6853 Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 20a (37s) == replay-single test 20b: write, unlink, eviction, replay (test mds_cleanup_orphans) ========================================================== 17:19:32 (1713388772) /mnt/lustre/f20b.replay-single lmm_stripe_count: 1 lmm_stripe_size: 1048576 lmm_pattern: raid0 lmm_layout_gen: 0 lmm_stripe_offset: 0 obdidx objid objid group 0 1122 0x462 0 10000+0 records in 10000+0 records out 40960000 bytes (41 MB) copied, 1.11439 s, 36.8 MB/s Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec affected facets: mds1 oleg428-server: oleg428-server.virtnet: executing _wait_recovery_complete *.lustre-MDT0000.recovery_status 1475 oleg428-server: *.lustre-MDT0000.recovery_status status: COMPLETE sleep 5 for ZFS zfs Waiting for MDT destroys to complete before 6144, after 6144 PASS 20b (43s) == replay-single test 20c: check that client eviction does not affect file content ========================================================== 17:20:15 (1713388815) multiop /mnt/lustre/f20c.replay-single vOw_c TMPPIPE=/tmp/multiop_open_wait_pipe.6853 -rw-r--r-- 1 root root 1 Apr 17 17:20 /mnt/lustre/f20c.replay-single oleg428-client: error: invalid path '/mnt/lustre': Input/output error pdsh@oleg428-client: oleg428-client: ssh exited with exit code 5 PASS 20c (5s) == replay-single test 21: |X| open(O_CREAT), unlink touch new, replay, close (test mds_cleanup_orphans) ========================================================== 17:20:20 (1713388820) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre multiop /mnt/lustre/f21.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6853 Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 21 (27s) == replay-single test 22: open(O_CREAT), |X| unlink, replay, close (test mds_cleanup_orphans) ========================================================== 17:20:47 (1713388847) multiop /mnt/lustre/f22.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6853 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3764224 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7530496 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 22 (36s) == replay-single test 23: open(O_CREAT), |X| unlink touch new, replay, close (test mds_cleanup_orphans) ========================================================== 17:21:23 (1713388883) multiop /mnt/lustre/f23.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6853 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 23 (34s) == replay-single test 24: open(O_CREAT), replay, unlink, close (test mds_cleanup_orphans) ========================================================== 17:21:57 (1713388917) multiop /mnt/lustre/f24.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6853 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 24 (34s) == replay-single test 25: open(O_CREAT), unlink, replay, close (test mds_cleanup_orphans) ========================================================== 17:22:31 (1713388951) multiop /mnt/lustre/f25.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6853 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 25 (26s) == replay-single test 26: |X| open(O_CREAT), unlink two, close one, replay, close one (test mds_cleanup_orphans) ========================================================== 17:22:57 (1713388977) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre multiop /mnt/lustre/f26.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6853 multiop /mnt/lustre/f26.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6853 Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 26 (36s) == replay-single test 27: |X| open(O_CREAT), unlink two, replay, close two (test mds_cleanup_orphans) ========================================================== 17:23:33 (1713389013) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre multiop /mnt/lustre/f27.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6853 multiop /mnt/lustre/f27.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6853 Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 27 (34s) == replay-single test 28: open(O_CREAT), |X| unlink two, close one, replay, close one (test mds_cleanup_orphans) ========================================================== 17:24:07 (1713389047) multiop /mnt/lustre/f28.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6853 multiop /mnt/lustre/f28.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6853 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 28 (33s) == replay-single test 29: open(O_CREAT), |X| unlink two, replay, close two (test mds_cleanup_orphans) ========================================================== 17:24:40 (1713389080) multiop /mnt/lustre/f29.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6853 multiop /mnt/lustre/f29.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6853 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 29 (35s) == replay-single test 30: open(O_CREAT) two, unlink two, replay, close two (test mds_cleanup_orphans) ========================================================== 17:25:15 (1713389115) multiop /mnt/lustre/f30.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6853 multiop /mnt/lustre/f30.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6853 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 30 (34s) == replay-single test 31: open(O_CREAT) two, unlink one, |X| unlink one, close two (test mds_cleanup_orphans) ========================================================== 17:25:49 (1713389149) multiop /mnt/lustre/f31.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6853 multiop /mnt/lustre/f31.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6853 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 31 (34s) == replay-single test 32: close() notices client eviction; close() after client eviction ========================================================== 17:26:23 (1713389183) multiop /mnt/lustre/f32.replay-single vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.6853 multiop /mnt/lustre/f32.replay-single vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.6853 oleg428-client: error: invalid path '/mnt/lustre': Input/output error pdsh@oleg428-client: oleg428-client: ssh exited with exit code 5 PASS 32 (5s) == replay-single test 33a: fid seq shouldn't be reused after abort recovery ========================================================== 17:26:28 (1713389188) total: 10 open/close in 0.06 seconds: 180.08 ops/second Replay barrier on lustre-MDT0000 Stopping /mnt/lustre-mds1 (opts:) on oleg428-server Failover mds1 to oleg428-server Starting mds1: -o localrecov -o abort_recovery lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client: error: invalid path '/mnt/lustre': Input/output error pdsh@oleg428-client: oleg428-client: ssh exited with exit code 5 first stat failed: 5 total: 10 open/close in 0.05 seconds: 186.80 ops/second PASS 33a (12s) == replay-single test 33b: test fid seq allocation ======= 17:26:40 (1713389200) fail_loc=0x1311 total: 10 open/close in 0.05 seconds: 195.75 ops/second Replay barrier on lustre-MDT0000 Stopping /mnt/lustre-mds1 (opts:) on oleg428-server Failover mds1 to oleg428-server Starting mds1: -o localrecov -o abort_recovery lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 total: 10 open/close in 0.06 seconds: 175.73 ops/second PASS 33b (12s) == replay-single test 34: abort recovery before client does replay (test mds_cleanup_orphans) ========================================================== 17:26:52 (1713389212) multiop /mnt/lustre/f34.replay-single vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.6853 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Stopping /mnt/lustre-mds1 (opts:) on oleg428-server Failover mds1 to oleg428-server Starting mds1: -o localrecov -o abort_recovery lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 PASS 34 (12s) == replay-single test 35: test recovery from llog for unlink op ========================================================== 17:27:04 (1713389224) fail_loc=0x80000119 Stopping /mnt/lustre-mds1 (opts:) on oleg428-server Failover mds1 to oleg428-server Starting mds1: -o localrecov -o abort_recovery lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 rm: cannot remove '/mnt/lustre/f35.replay-single': Input/output error pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 Can't lstat /mnt/lustre/f35.replay-single: No such file or directory PASS 35 (12s) == replay-single test 36: don't resend cancel ============ 17:27:16 (1713389236) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 PASS 36 (21s) == replay-single test 37: abort recovery before client does replay (test mds_cleanup_orphans for directories) ========================================================== 17:27:37 (1713389257) multiop /mnt/lustre/d37.replay-single/f37.replay-single vdD_c TMPPIPE=/tmp/multiop_open_wait_pipe.6853 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Stopping /mnt/lustre-mds1 (opts:) on oleg428-server Failover mds1 to oleg428-server Starting mds1: -o localrecov -o abort_recovery lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 PASS 37 (12s) == replay-single test 38: test recovery from unlink llog (test llog_gen_rec) ========================================================== 17:27:49 (1713389269) total: 800 open/close in 3.21 seconds: 249.40 ops/second - unlinked 0 (time 1713389275 ; total 0 ; last 0) total: 400 unlinks in 1 seconds: 400.000000 unlinks/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec - unlinked 0 (time 1713389308 ; total 0 ; last 0) total: 400 unlinks in 2 seconds: 200.000000 unlinks/second Can't lstat /mnt/lustre/f38.replay-single-*: No such file or directory PASS 38 (45s) == replay-single test 39: test recovery from unlink llog (test llog_gen_rec) ========================================================== 17:28:34 (1713389314) total: 800 open/close in 3.19 seconds: 250.72 ops/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre - unlinked 0 (time 1713389321 ; total 0 ; last 0) total: 400 unlinks in 1 seconds: 400.000000 unlinks/second Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec - unlinked 0 (time 1713389353 ; total 0 ; last 0) total: 400 unlinks in 2 seconds: 200.000000 unlinks/second Can't lstat /mnt/lustre/f39.replay-single-*: No such file or directory PASS 39 (45s) == replay-single test 40: cause recovery in ptlrpc, ensure IO continues ========================================================== 17:29:19 (1713389359) SKIP: replay-single test_40 layout_lock needs MDS connection for IO SKIP 40 (1s) == replay-single test 41: read from a valid osc while other oscs are invalid ========================================================== 17:29:20 (1713389360) 1+0 records in 1+0 records out 4096 bytes (4.1 kB) copied, 0.00268754 s, 1.5 MB/s 1+0 records in 1+0 records out 4096 bytes (4.1 kB) copied, 0.00631854 s, 648 kB/s PASS 41 (3s) == replay-single test 42: recovery after ost failure ===== 17:29:23 (1713389363) total: 800 open/close in 3.13 seconds: 255.69 ops/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3456 2205184 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre - unlinked 0 (time 1713389370 ; total 0 ; last 0) total: 400 unlinks in 1 seconds: 400.000000 unlinks/second debug=-1 Failing ost1 on oleg428-server Stopping /mnt/lustre-ost1 (opts:) on oleg428-server reboot facets: ost1 Failover ost1 to oleg428-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-OST0000 wait for MDS to timeout and recover debug=super ioctl neterror warning dlmtrace error emerg ha rpctrace vfstrace config console lfsck debug=super ioctl neterror warning dlmtrace error emerg ha rpctrace vfstrace config console lfsck - unlinked 0 (time 1713389428 ; total 0 ; last 0) total: 400 unlinks in 0 seconds: inf unlinks/second Can't lstat /mnt/lustre/f42.replay-single-*: No such file or directory PASS 42 (67s) == replay-single test 43: mds osc import failure during recovery; don't LBUG ========================================================== 17:30:30 (1713389430) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3456 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre fail_loc=0x80000204 Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 43 (37s) == replay-single test 44a: race in target handle connect ========================================================== 17:31:07 (1713389467) at_max=40 1 of 10 (1713389468) service : cur 1 worst 1 (at 1713389451, 17s ago) 1 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3456 2205184 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre 2 of 10 (1713389474) service : cur 5 worst 5 (at 1713389474, 0s ago) 5 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3456 2205184 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre 3 of 10 (1713389480) service : cur 5 worst 5 (at 1713389474, 6s ago) 5 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3456 2205184 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre 4 of 10 (1713389485) service : cur 5 worst 5 (at 1713389474, 11s ago) 5 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3456 2205184 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre 5 of 10 (1713389491) service : cur 5 worst 5 (at 1713389474, 17s ago) 5 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3456 2205184 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre 6 of 10 (1713389497) service : cur 5 worst 5 (at 1713389474, 23s ago) 5 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3456 2205184 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre 7 of 10 (1713389502) service : cur 5 worst 5 (at 1713389474, 28s ago) 5 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3456 2205184 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre 8 of 10 (1713389508) service : cur 5 worst 5 (at 1713389474, 34s ago) 5 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3456 2205184 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre 9 of 10 (1713389514) service : cur 5 worst 5 (at 1713389474, 40s ago) 5 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3456 2205184 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre 10 of 10 (1713389519) service : cur 5 worst 5 (at 1713389474, 45s ago) 5 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3456 2205184 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre fail_loc=0 at_max=600 PASS 44a (60s) == replay-single test 44b: race in target handle connect ========================================================== 17:32:07 (1713389527) 1 of 10 (1713389528) service : cur 5 worst 5 (at 1713389474, 54s ago) 5 0 0 0 fail_loc=0x80000704 Filesystem 1K-blocks Used Available Use% Mounted on 192.168.204.128@tcp:/lustre 7542784 6144 7532544 1% /mnt/lustre 2 of 10 (1713389570) service : cur 40 worst 40 (at 1713389568, 2s ago) 40 0 0 0 fail_loc=0x80000704 df: '/mnt/lustre': Operation already in progress 3 of 10 (1713389576) service : cur 40 worst 40 (at 1713389568, 8s ago) 40 0 0 0 fail_loc=0x80000704 df: '/mnt/lustre': Operation already in progress 4 of 10 (1713389577) service : cur 40 worst 40 (at 1713389568, 9s ago) 40 0 0 0 fail_loc=0x80000704 df: '/mnt/lustre': Operation already in progress 5 of 10 (1713389577) service : cur 40 worst 40 (at 1713389568, 9s ago) 40 0 0 0 fail_loc=0x80000704 df: '/mnt/lustre': Operation already in progress 6 of 10 (1713389578) service : cur 40 worst 40 (at 1713389568, 10s ago) 40 0 0 0 fail_loc=0x80000704 df: '/mnt/lustre': Operation already in progress 7 of 10 (1713389578) service : cur 40 worst 40 (at 1713389568, 10s ago) 40 0 0 0 fail_loc=0x80000704 df: '/mnt/lustre': Operation already in progress 8 of 10 (1713389579) service : cur 40 worst 40 (at 1713389568, 11s ago) 40 0 0 0 fail_loc=0x80000704 df: '/mnt/lustre': Operation already in progress 9 of 10 (1713389579) service : cur 40 worst 40 (at 1713389568, 11s ago) 40 0 0 0 fail_loc=0x80000704 df: '/mnt/lustre': Operation already in progress 10 of 10 (1713389580) service : cur 40 worst 40 (at 1713389568, 12s ago) 40 0 0 0 fail_loc=0x80000704 df: '/mnt/lustre': Operation already in progress PASS 44b (55s) == replay-single test 44c: race in target handle connect ========================================================== 17:33:02 (1713389582) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3456 2205184 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre total: 100 create in 0.26 seconds: 380.66 ops/second fail_loc=0x80000712 Stopping /mnt/lustre-mds1 (opts:) on oleg428-server Failover mds1 to oleg428-server Starting mds1: -o localrecov -o abort_recovery lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client: error: invalid path '/mnt/lustre': Input/output error pdsh@oleg428-client: oleg428-client: ssh exited with exit code 5 first stat failed: 5 unlink(/mnt/lustre/f44c.replay-single-0) error: No such file or directory total: 0 unlinks in 0 seconds: -nan unlinks/second Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec unlink(/mnt/lustre/f44c.replay-single-0) error: No such file or directory total: 0 unlinks in 0 seconds: -nan unlinks/second PASS 44c (43s) == replay-single test 45: Handle failed close ============ 17:33:45 (1713389625) multiop /mnt/lustre/f45.replay-single vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.6853 /mnt/lustre/f45.replay-single has type file OK PASS 45 (3s) == replay-single test 46: Don't leak file handle after open resend (3325) ========================================================== 17:33:48 (1713389628) fail_loc=0x122 fail_loc=0 Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec lfs path2fid: cannot get fid for 'f46.replay-single': No such file or directory PASS 46 (34s) == replay-single test 47: MDS->OSC failure during precreate cleanup (2824) ========================================================== 17:34:23 (1713389663) total: 20 open/close in 0.11 seconds: 181.23 ops/second Failing ost1 on oleg428-server Stopping /mnt/lustre-ost1 (opts:) on oleg428-server reboot facets: ost1 Failover ost1 to oleg428-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-OST0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec fail_loc=0x80000204 total: 20 open/close in 0.07 seconds: 273.34 ops/second - unlinked 0 (time 1713389745 ; total 0 ; last 0) total: 20 unlinks in 0 seconds: inf unlinks/second PASS 47 (84s) == replay-single test 48: MDS->OSC failure during precreate cleanup (2824) ========================================================== 17:35:47 (1713389747) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3456 2205184 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre total: 20 open/close in 0.10 seconds: 201.75 ops/second Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 fail_loc=0x80000216 total: 20 open/close in 0.09 seconds: 215.01 ops/second - unlinked 0 (time 1713389829 ; total 0 ; last 0) total: 40 unlinks in 0 seconds: inf unlinks/second PASS 48 (83s) == replay-single test 50: Double OSC recovery, don't LASSERT (3812) ========================================================== 17:37:11 (1713389831) PASS 50 (8s) == replay-single test 52: time out lock replay (3764) ==== 17:37:19 (1713389839) multiop /mnt/lustre/f52.replay-single vs_s TMPPIPE=/tmp/multiop_open_wait_pipe.6853 fail_loc=0x80000157 Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec fail_loc=0x0 fail_loc=0x0 PASS 52 (40s) == replay-single test 53a: |X| close request while two MDC requests in flight ========================================================== 17:37:59 (1713389879) fail_loc=0x80000115 fail_loc=0 Replay barrier on lustre-MDT0000 Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d53a.replay-single-1/f has type file OK /mnt/lustre/d53a.replay-single-2/f has type file OK PASS 53a (31s) == replay-single test 53b: |X| open request while two MDC requests in flight ========================================================== 17:38:30 (1713389910) multiop /mnt/lustre/d53b.replay-single-1/f vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.6853 fail_loc=0x80000107 fail_loc=0 Replay barrier on lustre-MDT0000 Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d53b.replay-single-1/f has type file OK /mnt/lustre/d53b.replay-single-2/f has type file OK PASS 53b (30s) == replay-single test 53c: |X| open request and close request while two MDC requests in flight ========================================================== 17:39:00 (1713389940) fail_loc=0x80000107 fail_loc=0x80000115 Replay barrier on lustre-MDT0000 Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 fail_loc=0 /mnt/lustre/d53c.replay-single-1/f has type file OK /mnt/lustre/d53c.replay-single-2/f has type file OK PASS 53c (29s) == replay-single test 53d: close reply while two MDC requests in flight ========================================================== 17:39:29 (1713389969) fail_loc=0x8000013b fail_loc=0 Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d53d.replay-single-1/f has type file OK /mnt/lustre/d53d.replay-single-2/f has type file OK PASS 53d (31s) == replay-single test 53e: |X| open reply while two MDC requests in flight ========================================================== 17:40:01 (1713390001) fail_loc=0x119 fail_loc=0 Replay barrier on lustre-MDT0000 Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d53e.replay-single-1/f has type file OK /mnt/lustre/d53e.replay-single-2/f has type file OK PASS 53e (31s) == replay-single test 53f: |X| open reply and close reply while two MDC requests in flight ========================================================== 17:40:32 (1713390032) fail_loc=0x119 fail_loc=0x8000013b Replay barrier on lustre-MDT0000 Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 fail_loc=0 /mnt/lustre/d53f.replay-single-1/f has type file OK /mnt/lustre/d53f.replay-single-2/f has type file OK PASS 53f (29s) == replay-single test 53g: |X| drop open reply and close request while close and open are both in flight ========================================================== 17:41:01 (1713390061) fail_loc=0x119 fail_loc=0x80000115 fail_loc=0 Replay barrier on lustre-MDT0000 Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 /mnt/lustre/d53g.replay-single-1/f has type file OK /mnt/lustre/d53g.replay-single-2/f has type file OK PASS 53g (28s) == replay-single test 53h: open request and close reply while two MDC requests in flight ========================================================== 17:41:29 (1713390089) fail_loc=0x80000107 fail_loc=0x8000013b Replay barrier on lustre-MDT0000 Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 fail_loc=0 /mnt/lustre/d53h.replay-single-1/f has type file OK /mnt/lustre/d53h.replay-single-2/f has type file OK PASS 53h (29s) == replay-single test 55: let MDS_CHECK_RESENT return the original return code instead of 0 ========================================================== 17:41:58 (1713390118) fail_loc=0x8000012b fail_loc=0x0 touch: cannot touch '/mnt/lustre/f55.replay-single': No such file or directory rm: cannot remove '/mnt/lustre/f55.replay-single': No such file or directory PASS 55 (10s) == replay-single test 56: don't replay a symlink open request (3440) ========================================================== 17:42:08 (1713390128) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3456 2205184 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 56 (42s) == replay-single test 57: test recovery from llog for setattr op ========================================================== 17:42:51 (1713390171) fail_loc=0x8000012c UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3456 2205184 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec affected facets: mds1 oleg428-server: oleg428-server.virtnet: executing _wait_recovery_complete *.lustre-MDT0000.recovery_status 1475 oleg428-server: *.lustre-MDT0000.recovery_status status: COMPLETE Waiting for orphan cleanup... osp.lustre-OST0000-osc-MDT0000.old_sync_processed osp.lustre-OST0001-osc-MDT0000.old_sync_processed wait 40 secs maximumly for oleg428-server mds-ost sync done. /mnt/lustre/f57.replay-single has type file OK fail_loc=0x0 PASS 57 (30s) == replay-single test 58a: test recovery from llog for setattr op (test llog_gen_rec) ========================================================== 17:43:21 (1713390201) fail_loc=0x8000012c - open/close 2498 (time 1713390212.88 total 10.00 last 249.69) total: 2500 open/close in 10.01 seconds: 249.79 ops/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3968 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec fail_loc=0x0 - unlinked 0 (time 1713390255 ; total 0 ; last 0) total: 2500 unlinks in 3 seconds: 833.333313 unlinks/second PASS 58a (58s) == replay-single test 58b: test replay of setxattr op ==== 17:44:19 (1713390259) Starting client: oleg428-client.virtnet: -o user_xattr,flock oleg428-server@tcp:/lustre /mnt/lustre2 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3968 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Stopping client oleg428-client.virtnet /mnt/lustre2 (opts:) oleg428-client.virtnet: executing wait_import_state_mount FULL mgc.*.mgs_server_uuid mgc.*.mgs_server_uuid in FULL state after 0 sec PASS 58b (28s) == replay-single test 58c: resend/reconstruct setxattr op ========================================================== 17:44:48 (1713390288) Starting client: oleg428-client.virtnet: -o user_xattr,flock oleg428-server@tcp:/lustre /mnt/lustre2 fail_val=0 fail_loc=0x123 fail_loc=0 fail_loc=0x119 fail_loc=0 Stopping client oleg428-client.virtnet /mnt/lustre2 (opts:) PASS 58c (23s) SKIP: replay-single test_59 skipping ALWAYS excluded test 59 == replay-single test 60: test llog post recovery init vs llog unlink ========================================================== 17:45:11 (1713390311) total: 200 open/close in 0.36 seconds: 548.29 ops/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre - unlinked 0 (time 1713390313 ; total 0 ; last 0) total: 100 unlinks in 1 seconds: 100.000000 unlinks/second Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec - unlinked 0 (time 1713390343 ; total 0 ; last 0) total: 100 unlinks in 0 seconds: inf unlinks/second PASS 60 (34s) == replay-single test 61a: test race llog recovery vs llog cleanup ========================================================== 17:45:45 (1713390345) total: 800 open/close in 1.88 seconds: 426.09 ops/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre - unlinked 0 (time 1713390350 ; total 0 ; last 0) total: 800 unlinks in 1 seconds: 800.000000 unlinks/second fail_val=0 fail_loc=0x80000221 Failing ost1 on oleg428-server Stopping /mnt/lustre-ost1 (opts:) on oleg428-server reboot facets: ost1 Failover ost1 to oleg428-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-OST0000 Failing ost1 on oleg428-server Stopping /mnt/lustre-ost1 (opts:) on oleg428-server reboot facets: ost1 Failover ost1 to oleg428-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-OST0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec fail_val=0 fail_loc=0x0 Can't lstat /mnt/lustre/d61a.replay-single/f61a.replay-single-*: No such file or directory PASS 61a (85s) == replay-single test 61b: test race mds llog sync vs llog cleanup ========================================================== 17:47:10 (1713390430) fail_loc=0x8000013a Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec 1+0 records in 1+0 records out 4096 bytes (4.1 kB) copied, 0.00432736 s, 947 kB/s PASS 61b (63s) == replay-single test 61c: test race mds llog sync vs llog cleanup ========================================================== 17:48:13 (1713390493) fail_val=0 fail_loc=0x80000222 Failing ost1 on oleg428-server Stopping /mnt/lustre-ost1 (opts:) on oleg428-server reboot facets: ost1 Failover ost1 to oleg428-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-OST0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec fail_val=0 fail_loc=0x0 PASS 61c (33s) == replay-single test 61d: error in llog_setup should cleanup the llog context correctly ========================================================== 17:48:46 (1713390526) Stopping /mnt/lustre-mds1 (opts:) on oleg428-server fail_loc=0x80000605 Starting mgs: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: mount.lustre: mount lustre-mdt1/mdt1 at /mnt/lustre-mds1 failed: Operation not supported pdsh@oleg428-client: oleg428-server: ssh exited with exit code 95 Start of lustre-mdt1/mdt1 on mgs failed 95 fail_loc=0 Starting mgs: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 PASS 61d (10s) == replay-single test 62: don't mis-drop resent replay === 17:48:56 (1713390536) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre total: 25 open/close in 0.13 seconds: 190.13 ops/second fail_loc=0x80000707 Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec fail_loc=0 - unlinked 0 (time 1713390578 ; total 0 ; last 0) total: 25 unlinks in 0 seconds: inf unlinks/second PASS 62 (43s) == replay-single test 65a: AT: verify early replies ====== 17:49:39 (1713390579) at_history=8 at_history=8 debug=other fail_val=6000 fail_loc=0x8000050a 00000100:00001000:3.0:1713390604.634707:0:5864:0:(client.c:514:ptlrpc_at_recv_early_reply()) @@@ Early reply #1, new deadline in 31s (25s) req@ffff88013695b900 x1796617153137408/t0(0) o101->lustre-MDT0000-mdc-ffff8800aada4800@192.168.204.128@tcp:12/10 lens 664/66264 e 1 to 0 dl 1713390635 ref 2 fl Rpc:PQr/0/ffffffff rc 0/-1 job:'createmany.0' portal 12 : cur 31 worst 40 (at 1713389569, 1041s ago) 31 1 1 1 portal 23 : cur 1 worst 10 (at 1713387947, 2663s ago) 1 10 1 1 portal 30 : cur 1 worst 1 (at 1713387898, 2712s ago) 1 0 1 1 portal 17 : cur 1 worst 1 (at 1713387987, 2623s ago) 1 1 1 1 portal 12 : cur 6 worst 40 (at 1713389569, 1050s ago) 31 1 1 1 portal 23 : cur 1 worst 10 (at 1713387947, 2672s ago) 1 10 1 1 portal 30 : cur 1 worst 1 (at 1713387898, 2721s ago) 1 0 1 1 portal 17 : cur 1 worst 1 (at 1713387987, 2632s ago) 1 1 1 1 PASS 65a (41s) == replay-single test 65b: AT: verify early replies on packed reply / bulk ========================================================== 17:50:20 (1713390620) at_history=8 at_history=8 debug=other trace fail_val=6 fail_loc=0x224 fail_loc=0 00000100:00001000:3.0:1713390645.700474:0:1826:0:(client.c:514:ptlrpc_at_recv_early_reply()) @@@ Early reply #1, new deadline in 31s (25s) req@ffff88013695a600 x1796617153145024/t0(0) o4->lustre-OST0000-osc-ffff8800aada4800@192.168.204.128@tcp:6/4 lens 4584/448 e 1 to 0 dl 1713390676 ref 2 fl Rpc:Qr/0/ffffffff rc 0/-1 job:'multiop.0' debug=super ioctl neterror warning dlmtrace error emerg ha rpctrace vfstrace config console lfsck debug=super ioctl neterror warning dlmtrace error emerg ha rpctrace vfstrace config console lfsck portal 28 : cur 1 worst 1 (at 1713387895, 2756s ago) 1 1 1 1 portal 7 : cur 1 worst 1 (at 1713387896, 2755s ago) 1 1 1 1 portal 17 : cur 1 worst 1 (at 1713387955, 2696s ago) 1 1 1 1 portal 6 : cur 31 worst 31 (at 1713390650, 1s ago) 31 0 0 0 PASS 65b (33s) == replay-single test 66a: AT: verify MDT service time adjusts with no early replies ========================================================== 17:50:53 (1713390653) at_history=8 at_history=8 portal 12 : cur 1 worst 40 (at 1713389569, 1107s ago) 31 1 1 1 fail_val=5000 fail_loc=0x8000050a portal 12 : cur 31 worst 40 (at 1713389569, 1112s ago) 31 1 1 1 fail_val=10000 fail_loc=0x8000050a portal 12 : cur 10 worst 40 (at 1713389569, 1123s ago) 31 1 1 1 fail_loc=0 portal 12 : cur 1 worst 40 (at 1713389569, 1133s ago) 31 1 1 1 Current MDT timeout 1, worst 40 PASS 66a (50s) == replay-single test 66b: AT: verify net latency adjusts ========================================================== 17:51:43 (1713390703) at_history=8 at_history=8 fail_val=6 fail_loc=0x50c fail_loc=0 network timeout orig 1, cur 6, worst 6 PASS 66b (49s) == replay-single test 67a: AT: verify slow request processing doesn't induce reconnects ========================================================== 17:52:32 (1713390752) at_history=8 at_history=8 fail_val=400 fail_loc=0x50a fail_loc=0 0 osc reconnect attempts on gradual slow PASS 67a (55s) == replay-single test 67b: AT: verify instant slowdown doesn't induce reconnects ========================================================== 17:53:27 (1713390807) at_history=8 at_history=8 Creating to objid 6945 on ost lustre-OST0000... fail_val=20000 fail_loc=0x80000223 total: 18 open/close in 0.05 seconds: 340.09 ops/second Connected clients: oleg428-client.virtnet oleg428-client.virtnet service : cur 1 worst 1 (at 1713387661, 3170s ago) 1 0 0 1 phase 2 0 osc reconnect attempts on instant slow fail_loc=0x80000223 fail_loc=0 Connected clients: oleg428-client.virtnet oleg428-client.virtnet service : cur 1 worst 1 (at 1713387661, 3171s ago) 1 0 0 1 0 osc reconnect attempts on 2nd slow PASS 67b (27s) == replay-single test 68: AT: verify slowing locks ======= 17:53:54 (1713390834) at_history=8 at_history=8 fail_val=19 fail_loc=0x80000312 fail_val=25 fail_loc=0x80000312 fail_loc=0 PASS 68 (70s) Cleaning up AT ... == replay-single test 70a: check multi client t-f ======== 17:55:04 (1713390904) SKIP: replay-single test_70a Need two or more clients, have 1 SKIP 70a (1s) == replay-single test 70b: dbench 1mdts recovery; 1 clients ========================================================== 17:55:05 (1713390905) Starting client oleg428-client.virtnet: -o user_xattr,flock oleg428-server@tcp:/lustre /mnt/lustre Started clients oleg428-client.virtnet: 192.168.204.128@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt) + MISSING_DBENCH_OK= + PATH=/opt/iozone/bin:/opt/iozone/bin:/opt/iozone/bin:/home/green/git/lustre-release/lustre/tests/mpi:/home/green/git/lustre-release/lustre/tests/racer:/home/green/git/lustre-release/lustre/../lustre-iokit/sgpdd-survey:/home/green/git/lustre-release/lustre/tests:/home/green/git/lustre-release/lustre/utils/gss:/home/green/git/lustre-release/lustre/utils:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin::/home/green/git/lustre-release/lustre/utils:/home/green/git/lustre-release/lustre/tests:/sbin:/usr/sbin:/home/green/git/lustre-release/lustre/utils:/home/green/git/lustre-release/lustre/tests/: + DBENCH_LIB= + TESTSUITE=replay-single + TESTNAME=test_70b + MOUNT=/mnt/lustre ++ hostname + DIR=/mnt/lustre/d70b.replay-single/oleg428-client.virtnet + LCTL=/home/green/git/lustre-release/lustre/utils/lctl + rundbench 1 -t 300 dbench: no process found Started rundbench load pid=11998 ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3716096 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 30720 3680256 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 33792 7396352 1% /mnt/lustre test_70b fail mds1 1 times Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec oleg428-client.virtnet: looking for dbench program oleg428-client.virtnet: /usr/bin/dbench oleg428-client.virtnet: creating output directory /mnt/lustre/d70b.replay-single/oleg428-client.virtnet oleg428-client.virtnet: mkdir: created directory '/mnt/lustre/d70b.replay-single' oleg428-client.virtnet: mkdir: created directory '/mnt/lustre/d70b.replay-single/oleg428-client.virtnet' oleg428-client.virtnet: found dbench client file /usr/share/dbench/client.txt oleg428-client.virtnet: '/usr/share/dbench/client.txt' -> 'client.txt' oleg428-client.virtnet: running 'dbench 1 -t 300' on /mnt/lustre/d70b.replay-single/oleg428-client.virtnet at Wed Apr 17 17:55:06 EDT 2024 oleg428-client.virtnet: waiting for dbench pid 12022 oleg428-client.virtnet: dbench version 4.00 - Copyright Andrew Tridgell 1999-2004 oleg428-client.virtnet: oleg428-client.virtnet: Running for 300 seconds with load 'client.txt' and minimum warmup 60 secs oleg428-client.virtnet: failed to create barrier semaphore oleg428-client.virtnet: 0 of 1 processes prepared for launch 0 sec oleg428-client.virtnet: 1 of 1 processes prepared for launch 0 sec oleg428-client.virtnet: releasing clients oleg428-client.virtnet: 1 313 11.48 MB/sec warmup 1 sec latency 23.706 ms oleg428-client.virtnet: 1 664 10.58 MB/sec warmup 2 sec latency 19.810 ms oleg428-client.virtnet: 1 885 7.22 MB/sec warmup 3 sec latency 51.484 ms oleg428-client.virtnet: 1 1202 5.76 MB/sec warmup 4 sec latency 45.556 ms oleg428-client.virtnet: 1 1619 5.23 MB/sec warmup 5 sec latency 41.491 ms oleg428-client.virtnet: 1 1841 4.39 MB/sec warmup 6 sec latency 358.219 ms oleg428-client.virtnet: 1 1841 3.76 MB/sec warmup 7 sec latency 1358.454 ms oleg428-client.virtnet: 1 1841 3.29 MB/sec warmup 8 sec latency 2358.730 ms oleg428-client.virtnet: 1 1841 2.93 MB/sec warmup 9 sec latency 3358.975 ms oleg428-client.virtnet: 1 1841 2.64 MB/sec warmup 10 sec latency 4359.228 ms oleg428-client.virtnet: 1 1841 2.40 MB/sec warmup 11 sec latency 5359.455 ms oleg428-client.virtnet: 1 1841 2.20 MB/sec warmup 12 sec latency 6359.648 ms oleg428-client.virtnet: 1 1841 2.03 MB/sec warmup 13 sec latency 7359.935 ms oleg428-client.virtnet: 1 1841 1.88 MB/sec warmup 14 sec latency 8360.186 ms oleg428-client.virtnet: 1 1841 1.76 MB/sec warmup 15 sec latency 9360.385 ms oleg428-client.virtnet: 1 1841 1.65 MB/sec warmup 16 sec latency 10360.574 ms oleg428-client.virtnet: 1 1841 1.55 MB/sec warmup 17 sec latency 11360.788 ms oleg428-client.virtnet: 1 1841 1.46 MB/sec warmup 18 sec latency 12360.982 ms oleg428-client.virtnet: 1 1841 1.39 MB/sec warmup 19 sec latency 13361.221 ms oleg428-client.virtnet: 1 1841 1.32 MB/sec warmup 20 sec latency 14361.486 ms oleg428-client.virtnet: 1 1841 1.25 MB/sec warmup 21 sec latency 15361.749 ms oleg428-client.virtnet: 1 1841 1.20 MB/sec warmup 22 sec latency 16361.967 ms oleg428-client.virtnet: 1 1841 1.15 MB/sec warmup 23 sec latency 17362.156 ms oleg428-client.virtnet: 1 1841 1.10 MB/sec warmup 24 sec latency 18362.363 ms oleg428-client.virtnet: 1 1841 1.05 MB/sec warmup 25 sec latency 19362.554 ms oleg428-client.virtnet: 1 1841 1.01 MB/sec warmup 26 sec latency 20362.761 ms oleg428-client.virtnet: 1 1904 0.98 MB/sec warmup 27 sec latency 21139.359 ms oleg428-client.virtnet: 1 2343 0.99 MB/sec warmup 28 sec latency 13.009 ms oleg428-client.virtnet: 1 2752 1.10 MB/sec warmup 29 sec latency 49.549 ms oleg428-client.virtnet: 1 3297 1.18 MB/sec warmup 30 sec latency 45.436 ms oleg428-client.virtnet: 1 3731 1.26 MB/sec warmup 31 sec latency 46.311 ms oleg428-client.virtnet: 1 3983 1.25 MB/sec warmup 32 sec latency 17.UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 13312 3720192 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 31744 3664896 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 45056 7385088 1% /mnt/lustre test_70b fail mds1 2 times Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 14336 3735552 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 40960 3704832 2% /mnt/lustre[OST:1] filesystem_summary: 7542784 55296 7440384 1% /mnt/lustre test_70b fail mds1 3 times Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 869 ms oleg428-client.virtnet: 1 4281 1.22 MB/sec warmup 33 sec latency 57.175 ms oleg428-client.virtnet: 1 4493 1.20 MB/sec warmup 34 sec latency 55.567 ms oleg428-client.virtnet: 1 4834 1.21 MB/sec warmup 35 sec latency 40.982 ms oleg428-client.virtnet: 1 5212 1.26 MB/sec warmup 36 sec latency 50.577 ms oleg428-client.virtnet: 1 5216 1.22 MB/sec warmup 37 sec latency 997.587 ms oleg428-client.virtnet: 1 5216 1.19 MB/sec warmup 38 sec latency 1997.760 ms oleg428-client.virtnet: 1 5216 1.16 MB/sec warmup 39 sec latency 2997.938 ms oleg428-client.virtnet: 1 5216 1.13 MB/sec warmup 40 sec latency 3998.164 ms oleg428-client.virtnet: 1 5216 1.11 MB/sec warmup 41 sec latency 4998.454 ms oleg428-client.virtnet: 1 5216 1.08 MB/sec warmup 42 sec latency 5998.596 ms oleg428-client.virtnet: 1 5216 1.05 MB/sec warmup 43 sec latency 6998.772 ms oleg428-client.virtnet: 1 5216 1.03 MB/sec warmup 44 sec latency 7998.952 ms oleg428-client.virtnet: 1 5216 1.01 MB/sec warmup 45 sec latency 8999.138 ms oleg428-client.virtnet: 1 5216 0.99 MB/sec warmup 46 sec latency 9999.345 ms oleg428-client.virtnet: 1 5216 0.96 MB/sec warmup 47 sec latency 10999.570 ms oleg428-client.virtnet: 1 5216 0.94 MB/sec warmup 48 sec latency 11999.817 ms oleg428-client.virtnet: 1 5216 0.92 MB/sec warmup 49 sec latency 13000.069 ms oleg428-client.virtnet: 1 5216 0.91 MB/sec warmup 50 sec latency 14000.323 ms oleg428-client.virtnet: 1 5216 0.89 MB/sec warmup 51 sec latency 15000.457 ms oleg428-client.virtnet: 1 5216 0.87 MB/sec warmup 52 sec latency 16000.687 ms oleg428-client.virtnet: 1 5216 0.86 MB/sec warmup 53 sec latency 17000.923 ms oleg428-client.virtnet: 1 5216 0.84 MB/sec warmup 54 sec latency 18001.172 ms oleg428-client.virtnet: 1 5216 0.82 MB/sec warmup 55 sec latency 19001.379 ms oleg428-client.virtnet: 1 5216 0.81 MB/sec warmup 56 sec latency 20001.613 ms oleg428-client.virtnet: 1 5216 0.79 MB/sec warmup 57 sec latency 21001.875 ms oleg428-client.virtnet: 1 5216 0.78 MB/sec warmup 58 sec latency 22002.134 ms oleg428-client.virtnet: 1 5216 0.77 MB/sec warmup 59 sec latency 23002.333 ms oleg428-client.virtnet: 1 5216 0.00 MB/sec execute 1 sec latency 25002.750 ms oleg428-client.virtnet: 1 5216 0.00 MB/sec execute 2 sec latency 26002.945 ms oleg428-client.virtnet: 1 5216 0.00 MB/sec execute 3 sec latency 27003.162 ms oleg428-client.virtnet: 1 5216 0.00 MB/sec execute 4 sec latency 28003.391 ms oleg428-client.virtnet: 1 5216 0.00 MB/sec execute 5 sec latency 29003.614 ms oleg428-client.virtnet: 1 5216 0.00 MB/sec execute 6 sec latency 30003.805 ms oleg428-client.virtnet: 1 5216 0.00 MB/sec execute 7 sec latency 31004.021 ms oleg428-client.virtnet: 1 5278 0.00 MB/sec execute 8 sec latency 31768.606 ms oleg428-client.virtnet: 1 5644 0.05 MB/sec execute 9 sec latency 48.504 ms oleg428-client.virtnet: 1 5996 0.19 MB/sec execute 10 sec latency 44.056 ms oleg428-client.virtnet: 1 6555 0.67 MB/sec execute 11 sec latency 46.899 ms oleg428-client.virtnet: 1 7171 1.03 MB/sec execute 12 sec latency 41.943 ms oleg428-client.virtnet: 1 7467 1.07 MB/sec execute 13 sec latency 19.006 ms oleg428-client.virtnet: 1 7836 1.02 MB/sec execute 14 sec latency 42.152 ms oleg428-client.virtnet: 1 8080 0.98 MB/sec execute 15 sec latency 43.103 ms oleg428-client.virtnet: 1 8502 1.18 MB/sec execute 16 sec latency 39.441 ms oleg428-client.virtnet: 1 8556 1.11 MB/sec execute 17 sec latency 833.580 ms oleg428-client.virtnet: 1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 13312 3737600 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 40960 3716096 2% /mnt/lustre[OST:1] filesystem_summary: 7542784 54272 7453696 1% /mnt/lustre test_70b fail mds1 4 times Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 8556 1.05 MB/sec execute 18 sec latency 1833.793 ms oleg428-client.virtnet: 1 8556 1.00 MB/sec execute 19 sec latency 2834.009 ms oleg428-client.virtnet: 1 8556 0.95 MB/sec execute 20 sec latency 3834.252 ms oleg428-client.virtnet: 1 8556 0.90 MB/sec execute 21 sec latency 4834.490 ms oleg428-client.virtnet: 1 8556 0.86 MB/sec execute 22 sec latency 5834.720 ms oleg428-client.virtnet: 1 8556 0.82 MB/sec execute 23 sec latency 6834.990 ms oleg428-client.virtnet: 1 8556 0.79 MB/sec execute 24 sec latency 7835.261 ms oleg428-client.virtnet: 1 8556 0.76 MB/sec execute 25 sec latency 8835.542 ms oleg428-client.virtnet: 1 8556 0.73 MB/sec execute 26 sec latency 9835.831 ms oleg428-client.virtnet: 1 8556 0.70 MB/sec execute 27 sec latency 10836.046 ms oleg428-client.virtnet: 1 8556 0.68 MB/sec execute 28 sec latency 11836.309 ms oleg428-client.virtnet: 1 8556 0.65 MB/sec execute 29 sec latency 12836.530 ms oleg428-client.virtnet: 1 8556 0.63 MB/sec execute 30 sec latency 13836.805 ms oleg428-client.virtnet: 1 8556 0.61 MB/sec execute 31 sec latency 14837.035 ms oleg428-client.virtnet: 1 8556 0.59 MB/sec execute 32 sec latency 15837.269 ms oleg428-client.virtnet: 1 8556 0.57 MB/sec execute 33 sec latency 16837.490 ms oleg428-client.virtnet: 1 8556 0.56 MB/sec execute 34 sec latency 17837.751 ms oleg428-client.virtnet: 1 8556 0.54 MB/sec execute 35 sec latency 18837.978 ms oleg428-client.virtnet: 1 8556 0.53 MB/sec execute 36 sec latency 19838.172 ms oleg428-client.virtnet: 1 8556 0.51 MB/sec execute 37 sec latency 20838.381 ms oleg428-client.virtnet: 1 8556 0.50 MB/sec execute 38 sec latency 21838.565 ms oleg428-client.virtnet: 1 8556 0.49 MB/sec execute 39 sec latency 22838.736 ms oleg428-client.virtnet: 1 8556 0.47 MB/sec execute 40 sec latency 23838.965 ms oleg428-client.virtnet: 1 8556 0.46 MB/sec execute 41 sec latency 24839.162 ms oleg428-client.virtnet: 1 8556 0.45 MB/sec execute 42 sec latency 25839.304 ms oleg428-client.virtnet: 1 8556 0.44 MB/sec execute 43 sec latency 26839.553 ms oleg428-client.virtnet: 1 8556 0.43 MB/sec execute 44 sec latency 27839.735 ms oleg428-client.virtnet: 1 8597 0.42 MB/sec execute 45 sec latency 28641.825 ms oleg428-client.virtnet: 1 8891 0.42 MB/sec execute 46 sec latency 49.730 ms oleg428-client.virtnet: 1 9262 0.42 MB/sec execute 47 sec latency 50.490 ms oleg428-client.virtnet: 1 9574 0.45 MB/sec execute 48 sec latency 45.952 ms oleg428-client.virtnet: 1 10088 0.54 MB/sec execute 49 sec latency 47.995 ms oleg428-client.virtnet: 1 10652 0.63 MB/sec execute 50 sec latency 47.847 ms oleg428-client.virtnet: 1 10951 0.64 MB/sec execute 51 sec latency 16.869 ms oleg428-client.virtnet: 1 11202 0.63 MB/sec execute 52 sec latency 17.108 ms oleg428-client.virtnet: 1 11449 0.63 MB/sec execute 53 sec latency 68.913 ms oleg428-client.virtnet: 1 11529 0.62 MB/sec execute 54 sec latency 724.865 ms oleg428-client.virtnet: 1 11529 0.61 MB/sec execute 55 sec latency 1725.066 ms oleg428-client.virtnet: 1 11529 0.60 MB/sec execute 56 sec latency 2725.279 ms oleg428-client.virtnet: 1 11529 0.59 MB/sec execute 57 sec latency 3725.499 ms oleg428-client.virtnet: 1 11529 0.58 MB/sec execute 58 sec latency 4725.701 ms oleg428-client.virtnet: 1 11529 0.57 MB/sec execute 59 sec latency 5725.880 ms oleg428-client.virtnet: 1 11529 0.56 MB/sec execute 60 sec latency 6726.081 ms oleg428-client.virtnet: 1 11529 0.55 MB/sec execute 61 sec latency 7726.269 ms oleg428-client.virtnet: 1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 13312 3738624 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 40960 3700736 2% /mnt/lustre[OST:1] filesystem_summary: 7542784 54272 7439360 1% /mnt/lustre test_70b fail mds1 5 times Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 11529 0.54 MB/sec execute 62 sec latency 8726.472 ms oleg428-client.virtnet: 1 11529 0.53 MB/sec execute 63 sec latency 9726.678 ms oleg428-client.virtnet: 1 11529 0.53 MB/sec execute 64 sec latency 10726.931 ms oleg428-client.virtnet: 1 11529 0.52 MB/sec execute 65 sec latency 11727.145 ms oleg428-client.virtnet: 1 11529 0.51 MB/sec execute 66 sec latency 12727.351 ms oleg428-client.virtnet: 1 11529 0.50 MB/sec execute 67 sec latency 13727.586 ms oleg428-client.virtnet: 1 11529 0.50 MB/sec execute 68 sec latency 14727.768 ms oleg428-client.virtnet: 1 11529 0.49 MB/sec execute 69 sec latency 15727.933 ms oleg428-client.virtnet: 1 11529 0.48 MB/sec execute 70 sec latency 16728.113 ms oleg428-client.virtnet: 1 11529 0.47 MB/sec execute 71 sec latency 17728.301 ms oleg428-client.virtnet: 1 11529 0.47 MB/sec execute 72 sec latency 18728.529 ms oleg428-client.virtnet: 1 11529 0.46 MB/sec execute 73 sec latency 19728.725 ms oleg428-client.virtnet: 1 11529 0.46 MB/sec execute 74 sec latency 20728.863 ms oleg428-client.virtnet: 1 11529 0.45 MB/sec execute 75 sec latency 21729.079 ms oleg428-client.virtnet: 1 11562 0.44 MB/sec execute 76 sec latency 22564.981 ms oleg428-client.virtnet: 1 12040 0.49 MB/sec execute 77 sec latency 48.291 ms oleg428-client.virtnet: 1 12406 0.49 MB/sec execute 78 sec latency 40.807 ms oleg428-client.virtnet: 1 12864 0.49 MB/sec execute 79 sec latency 37.319 ms oleg428-client.virtnet: 1 13369 0.55 MB/sec execute 80 sec latency 49.089 ms oleg428-client.virtnet: 1 14010 0.59 MB/sec execute 81 sec latency 40.376 ms oleg428-client.virtnet: 1 14475 0.63 MB/sec execute 82 sec latency 14.393 ms oleg428-client.virtnet: 1 14844 0.63 MB/sec execute 83 sec latency 12.967 ms oleg428-client.virtnet: 1 15109 0.63 MB/sec execute 84 sec latency 55.068 ms oleg428-client.virtnet: 1 15359 0.64 MB/sec execute 85 sec latency 414.878 ms oleg428-client.virtnet: 1 15359 0.63 MB/sec execute 86 sec latency 1415.024 ms oleg428-client.virtnet: 1 15359 0.62 MB/sec execute 87 sec latency 2415.212 ms oleg428-client.virtnet: 1 15359 0.62 MB/sec execute 88 sec latency 3415.368 ms oleg428-client.virtnet: 1 15359 0.61 MB/sec execute 89 sec latency 4415.573 ms oleg428-client.virtnet: 1 15359 0.60 MB/sec execute 90 sec latency 5415.702 ms oleg428-client.virtnet: 1 15359 0.60 MB/sec execute 91 sec latency 6415.860 ms oleg428-client.virtnet: 1 15359 0.59 MB/sec execute 92 sec latency 7415.997 ms oleg428-client.virtnet: 1 15359 0.58 MB/sec execute 93 sec latency 8416.155 ms oleg428-client.virtnet: 1 15359 0.58 MB/sec execute 94 sec latency 9416.331 ms oleg428-client.virtnet: 1 15359 0.57 MB/sec execute 95 sec latency 10416.489 ms oleg428-client.virtnet: 1 15359 0.56 MB/sec execute 96 sec latency 11416.613 ms oleg428-client.virtnet: 1 15359 0.56 MB/sec execute 97 sec latency 12416.773 ms oleg428-client.virtnet: 1 15359 0.55 MB/sec execute 98 sec latency 13416.969 ms oleg428-client.virtnet: 1 15359 0.55 MB/sec execute 99 sec latency 14417.109 ms oleg428-client.virtnet: 1 15359 0.54 MB/sec execute 100 sec latency 15417.225 ms oleg428-client.virtnet: 1 15359 0.54 MB/sec execute 101 sec latency 16417.357 ms oleg428-client.virtnet: 1 15359 0.53 MB/sec execute 102 sec latency 17417.500 ms oleg428-client.virtnet: 1 15359 0.53 MB/sec execute 103 sec latency 18417.686 ms oleg428-client.virtnet: 1 15359 0.52 MB/sec execute 104 sec latency 19417.937 ms oleg428-client.virtnet: 1 15359 0.52 MB/sec execute 105 sec latency 20418.105 ms oleg428-client.virtnet: 1oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 13312 3740672 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 40960 3713024 2% /mnt/lustre[OST:1] filesystem_summary: 7542784 54272 7453696 1% /mnt/lustre test_70b fail mds1 6 times Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 15359 0.51 MB/sec execute 106 sec latency 21418.326 ms oleg428-client.virtnet: 1 15359 0.51 MB/sec execute 107 sec latency 22418.535 ms oleg428-client.virtnet: 1 15359 0.50 MB/sec execute 108 sec latency 23418.695 ms oleg428-client.virtnet: 1 15359 0.50 MB/sec execute 109 sec latency 24418.837 ms oleg428-client.virtnet: 1 15359 0.49 MB/sec execute 110 sec latency 25419.029 ms oleg428-client.virtnet: 1 15359 0.49 MB/sec execute 111 sec latency 26419.227 ms oleg428-client.virtnet: 1 15359 0.48 MB/sec execute 112 sec latency 27419.486 ms oleg428-client.virtnet: 1 15366 0.48 MB/sec execute 113 sec latency 28375.417 ms oleg428-client.virtnet: 1 15761 0.50 MB/sec execute 114 sec latency 50.990 ms oleg428-client.virtnet: 1 16050 0.50 MB/sec execute 115 sec latency 53.976 ms oleg428-client.virtnet: 1 16455 0.50 MB/sec execute 116 sec latency 40.844 ms oleg428-client.virtnet: 1 16777 0.51 MB/sec execute 117 sec latency 60.554 ms oleg428-client.virtnet: 1 17382 0.55 MB/sec execute 118 sec latency 51.931 ms oleg428-client.virtnet: 1 17885 0.59 MB/sec execute 119 sec latency 43.721 ms oleg428-client.virtnet: 1 18145 0.59 MB/sec execute 120 sec latency 17.407 ms oleg428-client.virtnet: 1 18432 0.59 MB/sec execute 121 sec latency 56.313 ms oleg428-client.virtnet: 1 18522 0.59 MB/sec execute 122 sec latency 612.942 ms oleg428-client.virtnet: 1 18522 0.58 MB/sec execute 123 sec latency 1613.141 ms oleg428-client.virtnet: 1 18522 0.58 MB/sec execute 124 sec latency 2613.355 ms oleg428-client.virtnet: 1 18522 0.57 MB/sec execute 125 sec latency 3613.588 ms oleg428-client.virtnet: 1 18522 0.57 MB/sec execute 126 sec latency 4613.734 ms oleg428-client.virtnet: 1 18522 0.56 MB/sec execute 127 sec latency 5613.914 ms oleg428-client.virtnet: 1 18522 0.56 MB/sec execute 128 sec latency 6614.166 ms oleg428-client.virtnet: 1 18522 0.56 MB/sec execute 129 sec latency 7614.372 ms oleg428-client.virtnet: 1 18522 0.55 MB/sec execute 130 sec latency 8614.550 ms oleg428-client.virtnet: 1 18522 0.55 MB/sec execute 131 sec latency 9614.750 ms oleg428-client.virtnet: 1 18522 0.54 MB/sec execute 132 sec latency 10614.995 ms oleg428-client.virtnet: 1 18522 0.54 MB/sec execute 133 sec latency 11615.283 ms oleg428-client.virtnet: 1 18522 0.53 MB/sec execute 134 sec latency 12615.472 ms oleg428-client.virtnet: 1 18522 0.53 MB/sec execute 135 sec latency 13615.744 ms oleg428-client.virtnet: 1 18522 0.53 MB/sec execute 136 sec latency 14616.013 ms oleg428-client.virtnet: 1 18522 0.52 MB/sec execute 137 sec latency 15616.304 ms oleg428-client.virtnet: 1 18522 0.52 MB/sec execute 138 sec latency 16616.486 ms oleg428-client.virtnet: 1 18522 0.52 MB/sec execute 139 sec latency 17616.657 ms oleg428-client.virtnet: 1 18522 0.51 MB/sec execute 140 sec latency 18616.924 ms oleg428-client.virtnet: 1 18522 0.51 MB/sec execute 141 sec latency 19617.200 ms oleg428-client.virtnet: 1 18522 0.50 MB/sec execute 142 sec latency 20617.442 ms oleg428-client.virtnet: 1 18522 0.50 MB/sec execute 143 sec latency 21617.643 ms oleg428-client.virtnet: 1 18522 0.50 MB/sec execute 144 sec latency 22617.872 ms oleg428-client.virtnet: 1 18522 0.49 MB/sec execute 145 sec latency 23618.056 ms oleg428-client.virtnet: 1 18522 0.49 MB/sec execute 146 sec latency 24618.294 ms oleg428-client.virtnet: 1 18522 0.49 MB/sec execute 147 sec latency 25618.515 ms oleg428-client.virtnet: 1 18522 0.48 MB/sec execute 148 sec latency 26618.787 ms oleg428-client.virtnet: 1 18522 0.48 MB/sec execute 149 sec latency 27619.042 ms oleg428-client.virtnet:oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 13312 3748864 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 40960 3710976 2% /mnt/lustre[OST:1] filesystem_summary: 7542784 54272 7459840 1% /mnt/lustre test_70b fail mds1 7 times Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 14336 3744768 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 43008 3714048 2% /mnt/lustre[OST:1] filesystem_summary: 7542784 57344 7458816 1% /mnt/lustre 1 18522 0.48 MB/sec execute 150 sec latency 28619.217 ms oleg428-client.virtnet: 1 18736 0.48 MB/sec execute 151 sec latency 28651.945 ms oleg428-client.virtnet: 1 19177 0.50 MB/sec execute 152 sec latency 28.233 ms oleg428-client.virtnet: 1 19473 0.50 MB/sec execute 153 sec latency 44.746 ms oleg428-client.virtnet: 1 19831 0.50 MB/sec execute 154 sec latency 41.151 ms oleg428-client.virtnet: 1 20190 0.51 MB/sec execute 155 sec latency 63.900 ms oleg428-client.virtnet: 1 20694 0.54 MB/sec execute 156 sec latency 47.177 ms oleg428-client.virtnet: 1 21153 0.55 MB/sec execute 157 sec latency 44.449 ms oleg428-client.virtnet: 1 21555 0.57 MB/sec execute 158 sec latency 18.490 ms oleg428-client.virtnet: 1 21673 0.57 MB/sec execute 159 sec latency 487.103 ms oleg428-client.virtnet: 1 21673 0.56 MB/sec execute 160 sec latency 1487.336 ms oleg428-client.virtnet: 1 21673 0.56 MB/sec execute 161 sec latency 2487.653 ms oleg428-client.virtnet: 1 21673 0.56 MB/sec execute 162 sec latency 3487.988 ms oleg428-client.virtnet: 1 21673 0.55 MB/sec execute 163 sec latency 4488.239 ms oleg428-client.virtnet: 1 21673 0.55 MB/sec execute 164 sec latency 5488.523 ms oleg428-client.virtnet: 1 21673 0.55 MB/sec execute 165 sec latency 6488.878 ms oleg428-client.virtnet: 1 21673 0.54 MB/sec execute 166 sec latency 7489.138 ms oleg428-client.virtnet: 1 21673 0.54 MB/sec execute 167 sec latency 8489.400 ms oleg428-client.virtnet: 1 21673 0.54 MB/sec execute 168 sec latency 9489.706 ms oleg428-client.virtnet: 1 21673 0.53 MB/sec execute 169 sec latency 10490.065 ms oleg428-client.virtnet: 1 21673 0.53 MB/sec execute 170 sec latency 11490.277 ms oleg428-client.virtnet: 1 21673 0.53 MB/sec execute 171 sec latency 12490.467 ms oleg428-client.virtnet: 1 21673 0.53 MB/sec execute 172 sec latency 13490.661 ms oleg428-client.virtnet: 1 21673 0.52 MB/sec execute 173 sec latency 14490.932 ms oleg428-client.virtnet: 1 21673 0.52 MB/sec execute 174 sec latency 15491.151 ms oleg428-client.virtnet: 1 21673 0.52 MB/sec execute 175 sec latency 16491.377 ms oleg428-client.virtnet: 1 21673 0.51 MB/sec execute 176 sec latency 17491.710 ms oleg428-client.virtnet: 1 21673 0.51 MB/sec execute 177 sec latency 18491.971 ms oleg428-client.virtnet: 1 21673 0.51 MB/sec execute 178 sec latency 19492.191 ms oleg428-client.virtnet: 1 21673 0.50 MB/sec execute 179 sec latency 20492.427 ms oleg428-client.virtnet: 1 21673 0.50 MB/sec execute 180 sec latency 21492.750 ms oleg428-client.virtnet: 1 21673 0.50 MB/sec execute 181 sec latency 22492.960 ms oleg428-client.virtnet: 1 21673 0.50 MB/sec execute 182 sec latency 23493.172 ms oleg428-client.virtnet: 1 21673 0.49 MB/sec execute 183 sec latency 24493.397 ms oleg428-client.virtnet: 1 21673 0.49 MB/sec execute 184 sec latency 25493.776 ms oleg428-client.virtnet: 1 21673 0.49 MB/sec execute 185 sec latency 26494.008 ms oleg428-client.virtnet: 1 21673 0.49 MB/sec execute 186 sec latency 27494.178 ms oleg428-client.virtnet: 1 21679 0.48 MB/sec execute 187 sec latency 28451.771 ms oleg428-client.virtnet: 1 21918 0.48 MB/sec execute 188 sec latency 18.783 ms oleg428-client.virtnet: 1 22139 0.48 MB/sec execute 189 sec latency 68.895 ms oleg428-client.virtnet: 1 22450 0.49 MB/sec execute 190 sec latency 41.604 ms oleg428-client.virtnet: 1 22828 0.50 MB/sec execute 191 sec latency 49.944 ms oleg428-client.virtnet: 1 23123 0.50 MB/sec execute 192 sec latency 48.193 ms oleg428-client.virtnet: 1 23507 0.50 MB/sec execute 193 sec latency 39.290 ms oleg428-client.virtnet: 1 test_70b fail mds1 8 times Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 16384 3725312 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 40960 3715072 2% /mnt/lustre[OST:1] filesystem_summary: 7542784 57344 7440384 1% /mnt/lustre test_70b fail mds1 9 times Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 23828 0.51 MB/sec execute 194 sec latency 40.264 ms oleg428-client.virtnet: 1 24342 0.53 MB/sec execute 195 sec latency 43.102 ms oleg428-client.virtnet: 1 24667 0.53 MB/sec execute 196 sec latency 544.313 ms oleg428-client.virtnet: 1 24667 0.53 MB/sec execute 197 sec latency 1544.560 ms oleg428-client.virtnet: 1 24667 0.53 MB/sec execute 198 sec latency 2544.790 ms oleg428-client.virtnet: 1 24667 0.53 MB/sec execute 199 sec latency 3545.076 ms oleg428-client.virtnet: 1 24667 0.52 MB/sec execute 200 sec latency 4545.407 ms oleg428-client.virtnet: 1 24667 0.52 MB/sec execute 201 sec latency 5545.670 ms oleg428-client.virtnet: 1 24667 0.52 MB/sec execute 202 sec latency 6545.898 ms oleg428-client.virtnet: 1 24667 0.52 MB/sec execute 203 sec latency 7546.157 ms oleg428-client.virtnet: 1 24667 0.51 MB/sec execute 204 sec latency 8546.364 ms oleg428-client.virtnet: 1 24667 0.51 MB/sec execute 205 sec latency 9546.620 ms oleg428-client.virtnet: 1 24667 0.51 MB/sec execute 206 sec latency 10546.892 ms oleg428-client.virtnet: 1 24667 0.51 MB/sec execute 207 sec latency 11547.114 ms oleg428-client.virtnet: 1 24667 0.50 MB/sec execute 208 sec latency 12547.295 ms oleg428-client.virtnet: 1 24667 0.50 MB/sec execute 209 sec latency 13547.509 ms oleg428-client.virtnet: 1 24667 0.50 MB/sec execute 210 sec latency 14547.701 ms oleg428-client.virtnet: 1 24667 0.50 MB/sec execute 211 sec latency 15547.819 ms oleg428-client.virtnet: 1 24667 0.49 MB/sec execute 212 sec latency 16548.020 ms oleg428-client.virtnet: 1 24667 0.49 MB/sec execute 213 sec latency 17548.269 ms oleg428-client.virtnet: 1 24667 0.49 MB/sec execute 214 sec latency 18548.489 ms oleg428-client.virtnet: 1 24667 0.49 MB/sec execute 215 sec latency 19548.754 ms oleg428-client.virtnet: 1 24667 0.49 MB/sec execute 216 sec latency 20548.971 ms oleg428-client.virtnet: 1 24667 0.48 MB/sec execute 217 sec latency 21549.168 ms oleg428-client.virtnet: 1 24667 0.48 MB/sec execute 218 sec latency 22549.357 ms oleg428-client.virtnet: 1 24667 0.48 MB/sec execute 219 sec latency 23549.582 ms oleg428-client.virtnet: 1 24667 0.48 MB/sec execute 220 sec latency 24549.798 ms oleg428-client.virtnet: 1 24667 0.47 MB/sec execute 221 sec latency 25549.987 ms oleg428-client.virtnet: 1 24667 0.47 MB/sec execute 222 sec latency 26550.201 ms oleg428-client.virtnet: 1 24667 0.47 MB/sec execute 223 sec latency 27550.463 ms oleg428-client.virtnet: 1 24676 0.47 MB/sec execute 224 sec latency 28498.992 ms oleg428-client.virtnet: 1 25107 0.49 MB/sec execute 225 sec latency 17.262 ms oleg428-client.virtnet: 1 25332 0.48 MB/sec execute 226 sec latency 17.844 ms oleg428-client.virtnet: 1 25607 0.48 MB/sec execute 227 sec latency 58.405 ms oleg428-client.virtnet: 1 25806 0.48 MB/sec execute 228 sec latency 49.706 ms oleg428-client.virtnet: 1 26134 0.49 MB/sec execute 229 sec latency 46.455 ms oleg428-client.virtnet: 1 26513 0.50 MB/sec execute 230 sec latency 59.471 ms oleg428-client.virtnet: 1 26839 0.50 MB/sec execute 231 sec latency 40.906 ms oleg428-client.virtnet: 1 27245 0.50 MB/sec execute 232 sec latency 42.188 ms oleg428-client.virtnet: 1 27355 0.50 MB/sec execute 233 sec latency 472.279 ms oleg428-client.virtnet: 1 27355 0.50 MB/sec execute 234 sec latency 1472.529 ms oleg428-client.virtnet: 1 27355 0.50 MB/sec execute 235 sec latency 2472.852 ms oleg428-client.virtnet: 1 27355 0.50 MB/sec execute 236 sec latency 3473.110 ms oleg428-client.virtnet: 1 27355 0.49 MB/sec execute 237 sec latency 4473.378 ms oleg428-client.virtnet: 1 27355 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec 0.49 MB/sec execute 238 sec latency 5473.659 ms oleg428-client.virtnet: 1 27355 0.49 MB/sec execute 239 sec latency 6473.952 ms oleg428-client.virtnet: 1 27355 0.49 MB/sec execute 240 sec latency 7474.206 ms oleg428-client.virtnet: 1 27355 0.49 MB/sec execute 241 sec latency 8474.537 ms oleg428-client.virtnet: 1 27355 0.48 MB/sec execute 242 sec latency 9474.796 ms oleg428-client.virtnet: 1 27355 0.48 MB/sec execute 243 sec latency 10475.059 ms oleg428-client.virtnet: 1 27355 0.48 MB/sec execute 244 sec latency 11475.290 ms oleg428-client.virtnet: 1 27355 0.48 MB/sec execute 245 sec latency 12475.500 ms oleg428-client.virtnet: 1 27355 0.48 MB/sec execute 246 sec latency 13475.665 ms oleg428-client.virtnet: 1 27355 0.47 MB/sec execute 247 sec latency 14475.786 ms oleg428-client.virtnet: 1 27355 0.47 MB/sec execute 248 sec latency 15475.975 ms oleg428-client.virtnet: 1 27355 0.47 MB/sec execute 249 sec latency 16476.124 ms oleg428-client.virtnet: 1 27355 0.47 MB/sec execute 250 sec latency 17476.394 ms oleg428-client.virtnet: 1 27355 0.47 MB/sec execute 251 sec latency 18476.655 ms oleg428-client.virtnet: 1 27355 0.47 MB/sec execute 252 sec latency 19476.836 ms oleg428-client.virtnet: 1 27355 0.46 MB/sec execute 253 sec latency 20477.061 ms oleg428-client.virtnet: 1 27355 0.46 MB/sec execute 254 sec latency 21477.295 ms oleg428-client.virtnet: 1 27355 0.46 MB/sec execute 255 sec latency 22477.505 ms oleg428-client.virtnet: 1 27355 0.46 MB/sec execute 256 sec latency 23477.750 ms oleg428-client.virtnet: 1 27355 0.46 MB/sec execute 257 sec latency 24477.981 ms oleg428-client.virtnet: 1 27355 0.45 MB/sec execute 258 sec latency 25478.270 ms oleg428-client.virtnet: 1 27355 0.45 MB/sec execute 259 sec latency 26478.497 ms oleg428-client.virtnet: 1 27355 0.45 MB/sec execute 260 sec latency 27478.703 ms oleg428-client.virtnet: 1 27441 0.45 MB/sec execute 261 sec latency 28314.069 ms oleg428-client.virtnet: 1 27896 0.47 MB/sec execute 262 sec latency 54.396 ms oleg428-client.virtnet: 1 28495 0.48 MB/sec execute 263 sec latency 43.698 ms oleg428-client.virtnet: 1 28768 0.49 MB/sec execute 264 sec latency 19.432 ms oleg428-client.virtnet: 1 29029 0.49 MB/sec execute 265 sec latency 17.502 ms oleg428-client.virtnet: 1 29226 0.49 MB/sec execute 266 sec latency 83.579 ms oleg428-client.virtnet: 1 29488 0.49 MB/sec execute 267 sec latency 47.261 ms oleg428-client.virtnet: 1 30020 0.50 MB/sec execute 268 sec latency 37.232 ms oleg428-client.virtnet: 1 30402 0.50 MB/sec execute 269 sec latency 46.343 ms oleg428-client.virtnet: 1 30796 0.50 MB/sec execute 270 sec latency 44.969 ms oleg428-client.virtnet: 1 31211 0.52 MB/sec execute 271 sec latency 46.546 ms oleg428-client.virtnet: 1 31762 0.53 MB/sec execute 272 sec latency 44.263 ms oleg428-client.virtnet: 1 32179 0.54 MB/sec execute 273 sec latency 17.012 ms oleg428-client.virtnet: 1 32397 0.54 MB/sec execute 274 sec latency 17.979 ms oleg428-client.virtnet: 1 32686 0.54 MB/sec execute 275 sec latency 66.190 ms oleg428-client.virtnet: 1 32893 0.54 MB/sec execute 276 sec latency 50.552 ms oleg428-client.virtnet: 1 33267 0.54 MB/sec execute 277 sec latency 33.789 ms oleg428-client.virtnet: 1 33623 0.55 MB/sec execute 278 sec latency 53.761 ms oleg428-client.virtnet: 1 33984 0.55 MB/sec execute 279 sec latency 27.634 ms oleg428-client.virtnet: 1 34373 0.55 MB/sec execute 280 sec latency 47.838 ms oleg428-client.virtnet: 1 34867 0.57 MB/sec execute 281 sec latency 48.879 ms oleg428-client.virtnet: 1 35334 0.58 MB/sec execute 282 sec latency 46.622 ms oleg428-client.virtnet: 1 35752 0.59 MB/sec execute 283 sec latency 17.452 ms oleg428-client.virtnet: 1 35973 0.59 MB/sec execute 284 sec latency 17.583 ms oleg428-client.virtnet: 1 36247 0.59 MB/sec execute 285 sec latency 69.158 ms oleg428-client.virtnet: 1 36473 0.59 MB/sec execute 286 sec latency 51.103 ms oleg428-client.virtnet: 1 36800 0.59 MB/sec execute 287 sec latency 41.418 ms oleg428-client.virtnet: 1 37167 0.60 MB/sec execute 288 sec latency 44.350 ms oleg428-client.virtnet: 1 37492 0.60 MB/sec execute 289 sec latency 52.965 ms oleg428-client.virtnet: 1 37885 0.60 MB/sec execute 290 sec latency 47.707 ms oleg428-client.virtnet: 1 38278 0.61 MB/sec execute 291 sec latency 45.097 ms oleg428-client.virtnet: 1 38843 0.62 MB/sec execute 292 sec latency 45.580 ms oleg428-client.virtnet: 1 39255 0.63 MB/sec execute 293 sec latency 28.331 ms oleg428-client.virtnet: 1 39494 0.63 MB/sec execute 294 sec latency 17.405 ms oleg428-client.virtnet: 1 39788 0.63 MB/sec execute 295 sec latency 69.965 ms oleg428-client.virtnet: 1 39982 0.63 MB/sec execute 296 sec latency 57.962 ms oleg428-client.virtnet: 1 40305 0.63 MB/sec execute 297 sec latency 41.641 ms oleg428-client.virtnet: 1 40687 0.64 MB/sec execute 298 sec latency 45.927 ms oleg428-client.virtnet: 1 41071 0.64 MB/sec execute 299 sec latency 46.276 ms oleg428-client.virtnet: 1 cleanup 300 sec oleg428-client.virtnet: 0 cleanup 300 sec oleg428-client.virtnet: oleg428-client.virtnet: Operation Count AvgLat MaxLat oleg428-client.virtnet: ---------------------------------------- oleg428-client.virtnet: NTCreateX 6288 21.139 31768.593 oleg428-client.virtnet: Close 4653 13.995 28498.962 oleg428-client.virtnet: Rename 263 11.211 16.425 oleg428-client.virtnet: Unlink 1251 2.928 7.568 oleg428-client.virtnet: Qpathinfo 5727 7.033 28375.395 oleg428-client.virtnet: Qfileinfo 996 0.155 2.496 oleg428-client.virtnet: Qfsinfo 1025 0.166 5.575 oleg428-client.virtnet: Sfileinfo 507 117.551 28651.931 oleg428-client.virtnet: Find 2194 0.603 24.795 oleg428-client.virtnet: WriteX 3095 1.100 5.902 oleg428-client.virtnet: ReadX 9759 0.015 4.396 oleg428-client.virtnet: LockX 20 1.697 1.936 oleg428-client.virtnet: UnlockX 20 1.809 2.455 oleg428-client.virtnet: Flush 435 32.064 83.561 oleg428-client.virtnet: oleg428-client.virtnet: Throughput 0.640598 MB/sec 1 clients 1 procs max_latency=31768.606 ms oleg428-client.virtnet: stopping dbench on /mnt/lustre/d70b.replay-single/oleg428-client.virtnet at Wed Apr 17 18:01:07 EDT 2024 with return code 0 oleg428-client.virtnet: clean dbench files on /mnt/lustre/d70b.replay-single/oleg428-client.virtnet oleg428-client.virtnet: /mnt/lustre/d70b.replay-single/oleg428-client.virtnet /mnt/lustre/d70b.replay-single/oleg428-client.virtnet oleg428-client.virtnet: removed directory: 'clients/client0/~dmtmp/PWRPNT' oleg428-client.virtnet: removed directory: 'clients/client0/~dmtmp/WORDPRO' oleg428-client.virtnet: removed directory: 'clients/client0/~dmtmp/COREL' oleg428-client.virtnet: removed directory: 'clients/client0/~dmtmp/ACCESS' oleg428-client.virtnet: removed directory: 'clients/client0/~dmtmp/PM' oleg428-client.virtnet: removed directory: 'clients/client0/~dmtmp/SEED' oleg428-client.virtnet: removed directory: 'clients/client0/~dmtmp/WORD' oleg428-client.virtnet: removed directory: 'clients/client0/~dmtmp/PARADOX' oleg428-client.virtnet: removed directory: 'clients/client0/~dmtmp/EXCEL' oleg428-client.virtnet: removed directory: 'clients/client0/~dmtmp' oleg428-client.virtnet: removed directory: 'clients/client0' oleg428-client.virtnet: removed directory: 'clients' oleg428-client.virtnet: removed 'client.txt' oleg428-client.virtnet: /mnt/lustre/d70b.replay-single/oleg428-client.virtnet oleg428-client.virtnet: dbench successfully finished PASS 70b (364s) == replay-single test 70c: tar 1mdts recovery ============ 18:01:09 (1713391269) Starting client oleg428-client.virtnet: -o user_xattr,flock oleg428-server@tcp:/lustre /mnt/lustre Started clients oleg428-client.virtnet: 192.168.204.128@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt) Started tar 23043 tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 11776 2196864 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 17408 3187712 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 18432 3200000 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 35840 6387712 1% /mnt/lustre test_70c fail mds1 1 times Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 11776 2196864 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 30720 3214336 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 10240 3234816 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 40960 6449152 1% /mnt/lustre test_70c fail mds1 2 times Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 70c (404s) == replay-single test 70d: mkdir/rmdir striped dir 1mdts recovery ========================================================== 18:07:53 (1713391673) SKIP: replay-single test_70d needs >= 2 MDTs SKIP 70d (1s) == replay-single test 70e: rename cross-MDT with random fails ========================================================== 18:07:54 (1713391674) SKIP: replay-single test_70e needs >= 2 MDTs SKIP 70e (1s) == replay-single test 70f: OSS O_DIRECT recovery with 1 clients ========================================================== 18:07:55 (1713391675) mount clients oleg428-client.virtnet ... Starting client oleg428-client.virtnet: -o user_xattr,flock oleg428-server@tcp:/lustre /mnt/lustre Started clients oleg428-client.virtnet: 192.168.204.128@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt) Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3840 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 4096 3761152 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3761152 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 8192 7522304 1% /mnt/lustre Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... test_70f failing OST 1 times Failing ost1 on oleg428-server Stopping /mnt/lustre-ost1 (opts:) on oleg428-server reboot facets: ost1 Failover ost1 to oleg428-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... Started lustre-OST0000 ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 8192 3741696 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 8192 3750912 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 16384 7492608 1% /mnt/lustre ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... test_70f failing OST 2 times Failing ost1 on oleg428-server Stopping /mnt/lustre-ost1 (opts:) on oleg428-server reboot facets: ost1 Failover ost1 to oleg428-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-OST0000 ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 8192 3750912 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 8192 3750912 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 16384 7501824 1% /mnt/lustre ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... test_70f failing OST 3 times Failing ost1 on oleg428-server Stopping /mnt/lustre-ost1 (opts:) on oleg428-server reboot facets: ost1 Failover ost1 to oleg428-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-OST0000 ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3840 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 8192 3750912 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 8192 3750912 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 16384 7501824 1% /mnt/lustre ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... test_70f failing OST 4 times Failing ost1 on oleg428-server Stopping /mnt/lustre-ost1 (opts:) on oleg428-server reboot facets: ost1 Failover ost1 to oleg428-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-OST0000 ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3840 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 7168 3744768 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 8192 3755008 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 15360 7499776 1% /mnt/lustre ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... test_70f failing OST 5 times Failing ost1 on oleg428-server Stopping /mnt/lustre-ost1 (opts:) on oleg428-server reboot facets: ost1 Failover ost1 to oleg428-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... Started lustre-OST0000 ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg428-client.virtnet' ... ldlm.namespaces.MGC192.168.204.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aada4800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aada4800.lru_size=clear PASS 70f (142s) == replay-single test 71a: mkdir/rmdir striped dir with 2 mdts recovery ========================================================== 18:10:17 (1713391817) SKIP: replay-single test_71a needs >= 2 MDTs SKIP 71a (2s) == replay-single test 73a: open(O_CREAT), unlink, replay, reconnect before open replay, close ========================================================== 18:10:19 (1713391819) multiop /mnt/lustre/f73a.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6853 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3840 2204672 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 8192 3751936 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 8192 3751936 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 16384 7503872 1% /mnt/lustre fail_loc=0x80000302 Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 73a (39s) == replay-single test 73b: open(O_CREAT), unlink, replay, reconnect at open_replay reply, close ========================================================== 18:10:58 (1713391858) multiop /mnt/lustre/f73b.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6853 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre fail_loc=0x80000157 Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 73b (33s) == replay-single test 74: Ensure applications don't fail waiting for OST recovery ========================================================== 18:11:31 (1713391891) Stopping clients: oleg428-client.virtnet /mnt/lustre (opts:) Stopping client oleg428-client.virtnet /mnt/lustre opts: Stopping /mnt/lustre-ost1 (opts:) on oleg428-server Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 Starting client oleg428-client.virtnet: -o user_xattr,flock oleg428-server@tcp:/lustre /mnt/lustre Started clients oleg428-client.virtnet: 192.168.204.128@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt) Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-OST0000 PASS 74 (31s) == replay-single test 80a: DNE: create remote dir, drop update rep from MDT0, fail MDT0 ========================================================== 18:12:02 (1713391922) SKIP: replay-single test_80a needs >= 2 MDTs SKIP 80a (1s) == replay-single test 80b: DNE: create remote dir, drop update rep from MDT0, fail MDT1 ========================================================== 18:12:04 (1713391924) SKIP: replay-single test_80b needs >= 2 MDTs SKIP 80b (2s) == replay-single test 80c: DNE: create remote dir, drop update rep from MDT1, fail MDT[0,1] ========================================================== 18:12:05 (1713391925) SKIP: replay-single test_80c needs >= 2 MDTs SKIP 80c (1s) == replay-single test 80d: DNE: create remote dir, drop update rep from MDT1, fail 2 MDTs ========================================================== 18:12:06 (1713391926) SKIP: replay-single test_80d needs >= 2 MDTs SKIP 80d (1s) == replay-single test 80e: DNE: create remote dir, drop MDT1 rep, fail MDT0 ========================================================== 18:12:07 (1713391927) SKIP: replay-single test_80e needs >= 2 MDTs SKIP 80e (1s) == replay-single test 80f: DNE: create remote dir, drop MDT1 rep, fail MDT1 ========================================================== 18:12:08 (1713391928) SKIP: replay-single test_80f needs >= 2 MDTs SKIP 80f (1s) == replay-single test 80g: DNE: create remote dir, drop MDT1 rep, fail MDT0, then MDT1 ========================================================== 18:12:09 (1713391929) SKIP: replay-single test_80g needs >= 2 MDTs SKIP 80g (1s) == replay-single test 80h: DNE: create remote dir, drop MDT1 rep, fail 2 MDTs ========================================================== 18:12:10 (1713391930) SKIP: replay-single test_80h needs >= 2 MDTs SKIP 80h (1s) == replay-single test 81a: DNE: unlink remote dir, drop MDT0 update rep, fail MDT1 ========================================================== 18:12:11 (1713391931) SKIP: replay-single test_81a needs >= 2 MDTs SKIP 81a (1s) == replay-single test 81b: DNE: unlink remote dir, drop MDT0 update reply, fail MDT0 ========================================================== 18:12:12 (1713391932) SKIP: replay-single test_81b needs >= 2 MDTs SKIP 81b (1s) == replay-single test 81c: DNE: unlink remote dir, drop MDT0 update reply, fail MDT0,MDT1 ========================================================== 18:12:13 (1713391933) SKIP: replay-single test_81c needs >= 2 MDTs SKIP 81c (1s) == replay-single test 81d: DNE: unlink remote dir, drop MDT0 update reply, fail 2 MDTs ========================================================== 18:12:14 (1713391934) SKIP: replay-single test_81d needs >= 2 MDTs SKIP 81d (1s) == replay-single test 81e: DNE: unlink remote dir, drop MDT1 req reply, fail MDT0 ========================================================== 18:12:15 (1713391935) SKIP: replay-single test_81e needs >= 2 MDTs SKIP 81e (1s) == replay-single test 81f: DNE: unlink remote dir, drop MDT1 req reply, fail MDT1 ========================================================== 18:12:16 (1713391936) SKIP: replay-single test_81f needs >= 2 MDTs SKIP 81f (1s) == replay-single test 81g: DNE: unlink remote dir, drop req reply, fail M0, then M1 ========================================================== 18:12:18 (1713391938) SKIP: replay-single test_81g needs >= 2 MDTs SKIP 81g (1s) == replay-single test 81h: DNE: unlink remote dir, drop request reply, fail 2 MDTs ========================================================== 18:12:19 (1713391939) SKIP: replay-single test_81h needs >= 2 MDTs SKIP 81h (1s) == replay-single test 84a: stale open during export disconnect ========================================================== 18:12:20 (1713391940) fail_loc=0x80000144 total: 1 open/close in 0.01 seconds: 153.26 ops/second oleg428-client: error: invalid path '/mnt/lustre': Input/output error pdsh@oleg428-client: oleg428-client: ssh exited with exit code 5 PASS 84a (5s) == replay-single test 85a: check the cancellation of unused locks during recovery(IBITS) ========================================================== 18:12:25 (1713391945) before recovery: unused locks count = 301 Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec after recovery: unused locks count = 101 PASS 85a (25s) == replay-single test 85b: check the cancellation of unused locks during recovery(EXTENT) ========================================================== 18:12:50 (1713391970) before recovery: unused locks count = 100 Failing ost1 on oleg428-server Stopping /mnt/lustre-ost1 (opts:) on oleg428-server reboot facets: ost1 Failover ost1 to oleg428-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-OST0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec after recovery: unused locks count = 0 PASS 85b (27s) == replay-single test 86: umount server after clear nid_stats should not hit LBUG ========================================================== 18:13:17 (1713391997) Stopping clients: oleg428-client.virtnet /mnt/lustre (opts:) Stopping client oleg428-client.virtnet /mnt/lustre opts: mdt.lustre-MDT0000.exports.clear=0 Stopping /mnt/lustre-mds1 (opts:) on oleg428-server Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 Starting client oleg428-client.virtnet: -o user_xattr,flock oleg428-server@tcp:/lustre /mnt/lustre Started clients oleg428-client.virtnet: 192.168.204.128@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt) PASS 86 (8s) == replay-single test 87a: write replay ================== 18:13:25 (1713392005) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3712 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 15360 3753984 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 18432 7520256 1% /mnt/lustre 8+0 records in 8+0 records out 8388608 bytes (8.4 MB) copied, 0.126898 s, 66.1 MB/s Failing ost1 on oleg428-server Stopping /mnt/lustre-ost1 (opts:) on oleg428-server reboot facets: ost1 Failover ost1 to oleg428-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-OST0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec 8+0 records in 8+0 records out 8388608 bytes (8.4 MB) copied, 0.0652964 s, 128 MB/s PASS 87a (25s) == replay-single test 87b: write replay with changed data (checksum resend) ========================================================== 18:13:50 (1713392030) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3712 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 11264 3758080 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 14336 7524352 1% /mnt/lustre 8+0 records in 8+0 records out 8388608 bytes (8.4 MB) copied, 0.126785 s, 66.2 MB/s 8+0 records in 8+0 records out 8 bytes (8 B) copied, 0.00271554 s, 2.9 kB/s Failing ost1 on oleg428-server Stopping /mnt/lustre-ost1 (opts:) on oleg428-server reboot facets: ost1 Failover ost1 to oleg428-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-OST0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec 0+1 records in 0+1 records out 72 bytes (72 B) copied, 0.00520453 s, 13.8 kB/s PASS 87b (25s) == replay-single test 88: MDS should not assign same objid to different files ========================================================== 18:14:15 (1713392055) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3712 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 12288 3757056 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 15360 7523328 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3712 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 12288 3757056 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 15360 7523328 1% /mnt/lustre before test: last_id = 12417, next_id = 12386 Creating to objid 12417 on ost lustre-OST0000... total: 33 open/close in 0.11 seconds: 304.35 ops/second total: 8 open/close in 0.04 seconds: 220.95 ops/second before recovery: last_id = 12417, next_id = 12386 Stopping /mnt/lustre-mds1 (opts:) on oleg428-server Stopping /mnt/lustre-ost1 (opts:) on oleg428-server Failover mds1 to oleg428-server Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 Failover ost1 to oleg428-server Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-OST0000 after recovery: last_id = 12449, next_id = 12418 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0391497 s, 13.4 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0389458 s, 13.5 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0380577 s, 13.8 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0347356 s, 15.1 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0341119 s, 15.4 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0366927 s, 14.3 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0349186 s, 15.0 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0307972 s, 17.0 MB/s -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12386 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12387 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12388 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12389 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12390 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12391 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12392 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12393 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12394 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12395 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12396 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12397 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12398 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12399 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12400 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12401 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12402 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12403 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12404 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12405 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12406 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12407 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12408 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12409 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12410 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12411 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12412 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12413 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12414 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12415 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12416 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12417 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12418 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12419 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12420 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12421 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12422 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12423 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12424 -rw-r--r-- 1 root root 0 Apr 17 18:14 /mnt/lustre/d88.replay-single/f-12425 -rw-r--r-- 1 root root 524288 Apr 17 18:15 /mnt/lustre/d88.replay-single/f-12429 -rw-r--r-- 1 root root 524288 Apr 17 18:15 /mnt/lustre/d88.replay-single/f-12430 -rw-r--r-- 1 root root 524288 Apr 17 18:15 /mnt/lustre/d88.replay-single/f-12431 -rw-r--r-- 1 root root 524288 Apr 17 18:15 /mnt/lustre/d88.replay-single/f-12432 -rw-r--r-- 1 root root 524288 Apr 17 18:15 /mnt/lustre/d88.replay-single/f-12433 -rw-r--r-- 1 root root 524288 Apr 17 18:15 /mnt/lustre/d88.replay-single/f-12434 -rw-r--r-- 1 root root 524288 Apr 17 18:15 /mnt/lustre/d88.replay-single/f-12435 -rw-r--r-- 1 root root 524288 Apr 17 18:15 /mnt/lustre/d88.replay-single/f-12436 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0352265 s, 14.9 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0348848 s, 15.0 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0277679 s, 18.9 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0203682 s, 25.7 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0172226 s, 30.4 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0171018 s, 30.7 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0202084 s, 25.9 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0327559 s, 16.0 MB/s PASS 88 (61s) == replay-single test 89: no disk space leak on late ost connection ========================================================== 18:15:16 (1713392116) Waiting for orphan cleanup... osp.lustre-OST0000-osc-MDT0000.old_sync_processed osp.lustre-OST0001-osc-MDT0000.old_sync_processed wait 40 secs maximumly for oleg428-server mds-ost sync done. sleep 5 for ZFS zfs Waiting for MDT destroys to complete 10+0 records in 10+0 records out 41943040 bytes (42 MB) copied, 0.247617 s, 169 MB/s Stopping /mnt/lustre-ost1 (opts:) on oleg428-server Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-OST0000 Starting client: oleg428-client.virtnet: -o user_xattr,flock oleg428-server@tcp:/lustre /mnt/lustre osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 68 sec Waiting for orphan cleanup... osp.lustre-OST0000-osc-MDT0000.old_sync_processed osp.lustre-OST0001-osc-MDT0000.old_sync_processed wait 40 secs maximumly for oleg428-server mds-ost sync done. sleep 5 for ZFS zfs sleep 5 for ZFS zfs Waiting for MDT destroys to complete free_before: 7518208 free_after: 7518208 PASS 89 (128s) == replay-single test 90: lfs find identifies the missing striped file segments ========================================================== 18:17:24 (1713392244) Create the files Fail ost2 lustre-OST0001_UUID, display the list of affected files Stopping /mnt/lustre-ost2 (opts:) on oleg428-server General Query: lfs find /mnt/lustre/d90.replay-single /mnt/lustre/d90.replay-single /mnt/lustre/d90.replay-single/f1 /mnt/lustre/d90.replay-single/all /mnt/lustre/d90.replay-single/f0 Querying files on shutdown ost2: lfs find --obd lustre-OST0001_UUID /mnt/lustre/d90.replay-single/f1 /mnt/lustre/d90.replay-single/all Check getstripe: /home/green/git/lustre-release/lustre/utils/lfs getstripe -r --obd lustre-OST0001_UUID /mnt/lustre/d90.replay-single/f1 lmm_stripe_count: 1 lmm_stripe_size: 1048576 lmm_pattern: raid0 lmm_layout_gen: 0 lmm_stripe_offset: 1 obdidx objid objid group 1 9315 0x2463 0 * /mnt/lustre/d90.replay-single/all lmm_stripe_count: 2 lmm_stripe_size: 1048576 lmm_pattern: raid0 lmm_layout_gen: 0 lmm_stripe_offset: 1 obdidx objid objid group 1 9314 0x2462 0 * /mnt/lustre/d90.replay-single/all /mnt/lustre/d90.replay-single/f1 Failover ost2 to oleg428-server Starting ost2: -o localrecov lustre-ost2/ost2 /mnt/lustre-ost2 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-OST0001 PASS 90 (20s) == replay-single test 93a: replay + reconnect ============ 18:17:44 (1713392264) 1+0 records in 1+0 records out 1024 bytes (1.0 kB) copied, 0.00267702 s, 383 kB/s fail_val=40 fail_loc=0x715 Failing ost1 on oleg428-server Stopping /mnt/lustre-ost1 (opts:) on oleg428-server reboot facets: ost1 Failover ost1 to oleg428-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-OST0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec PASS 93a (61s) == replay-single test 93b: replay + reconnect on mds ===== 18:18:45 (1713392325) total: 20 open/close in 0.11 seconds: 189.13 ops/second fail_val=80 fail_loc=0x715 Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 93b (113s) == replay-single test 100a: DNE: create striped dir, drop update rep from MDT1, fail MDT1 ========================================================== 18:20:38 (1713392438) SKIP: replay-single test_100a needs >= 2 MDTs SKIP 100a (1s) == replay-single test 100b: DNE: create striped dir, fail MDT0 ========================================================== 18:20:39 (1713392439) SKIP: replay-single test_100b needs >= 2 MDTs SKIP 100b (1s) == replay-single test 100c: DNE: create striped dir, fail MDT0 ========================================================== 18:20:40 (1713392440) SKIP: replay-single test_100c needs >= 2 MDTs SKIP 100c (1s) == replay-single test 101: Shouldn't reassign precreated objs to other files after recovery ========================================================== 18:20:42 (1713392442) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3584 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 17408 3751936 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 7168 3762176 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 24576 7514112 1% /mnt/lustre Stopping /mnt/lustre-mds1 (opts:) on oleg428-server Failover mds1 to oleg428-server Starting mds1: -o localrecov -o abort_recovery lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client: error: invalid path '/mnt/lustre': Input/output error pdsh@oleg428-client: oleg428-client: ssh exited with exit code 5 first stat failed: 5 PASS 101 (37s) == replay-single test 102a: check resend (request lost) with multiple modify RPCs in flight ========================================================== 18:21:20 (1713392480) creating 7 files ... fail_loc=0x159 launch 7 chmod in parallel (18:21:21) ... fail_loc=0 done (18:21:28) /mnt/lustre/d102a.replay-single/file-1 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-2 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-3 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-4 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-5 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-6 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-7 has perms 0600 OK PASS 102a (9s) == replay-single test 102b: check resend (reply lost) with multiple modify RPCs in flight ========================================================== 18:21:29 (1713392489) creating 7 files ... fail_loc=0x15a launch 7 chmod in parallel (18:21:30) ... fail_loc=0 done (18:21:37) /mnt/lustre/d102b.replay-single/file-1 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-2 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-3 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-4 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-5 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-6 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-7 has perms 0600 OK PASS 102b (10s) == replay-single test 102c: check replay w/o reconstruction with multiple mod RPCs in flight ========================================================== 18:21:39 (1713392499) creating 7 files ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 17408 3337216 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 7168 3347456 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 24576 6684672 1% /mnt/lustre fail_loc=0x15a launch 7 chmod in parallel (18:21:42) ... fail_loc=0 Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec done (18:22:10) /mnt/lustre/d102c.replay-single/file-1 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-2 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-3 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-4 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-5 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-6 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-7 has perms 0600 OK PASS 102c (33s) == replay-single test 102d: check replay & reconstruction with multiple mod RPCs in flight ========================================================== 18:22:12 (1713392532) creating 7 files ... fail_loc=0x15a launch 7 chmod in parallel (18:22:13) ... fail_loc=0 Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec done (18:22:41) /mnt/lustre/d102d.replay-single/file-1 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-2 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-3 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-4 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-5 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-6 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-7 has perms 0600 OK PASS 102d (30s) == replay-single test 103: Check otr_next_id overflow ==== 18:22:42 (1713392562) fail_loc=0x80000162 total: 30 open/close in 0.15 seconds: 199.27 ops/second Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 103 (34s) == replay-single test 110a: DNE: create striped dir, fail MDT1 ========================================================== 18:23:16 (1713392596) SKIP: replay-single test_110a needs >= 2 MDTs SKIP 110a (1s) == replay-single test 110b: DNE: create striped dir, fail MDT1 and client ========================================================== 18:23:17 (1713392597) SKIP: replay-single test_110b needs >= 2 MDTs SKIP 110b (1s) == replay-single test 110c: DNE: create striped dir, fail MDT2 ========================================================== 18:23:19 (1713392599) SKIP: replay-single test_110c needs >= 2 MDTs SKIP 110c (1s) == replay-single test 110d: DNE: create striped dir, fail MDT2 and client ========================================================== 18:23:20 (1713392600) SKIP: replay-single test_110d needs >= 2 MDTs SKIP 110d (1s) == replay-single test 110e: DNE: create striped dir, uncommit on MDT2, fail client/MDT1/MDT2 ========================================================== 18:23:21 (1713392601) SKIP: replay-single test_110e needs >= 2 MDTs SKIP 110e (1s) SKIP: replay-single test_110f skipping excluded test 110f == replay-single test 110g: DNE: create striped dir, uncommit on MDT1, fail client/MDT1/MDT2 ========================================================== 18:23:23 (1713392603) SKIP: replay-single test_110g needs >= 2 MDTs SKIP 110g (1s) == replay-single test 111a: DNE: unlink striped dir, fail MDT1 ========================================================== 18:23:24 (1713392604) SKIP: replay-single test_111a needs >= 2 MDTs SKIP 111a (1s) == replay-single test 111b: DNE: unlink striped dir, fail MDT2 ========================================================== 18:23:25 (1713392605) SKIP: replay-single test_111b needs >= 2 MDTs SKIP 111b (1s) == replay-single test 111c: DNE: unlink striped dir, uncommit on MDT1, fail client/MDT1/MDT2 ========================================================== 18:23:26 (1713392606) SKIP: replay-single test_111c needs >= 2 MDTs SKIP 111c (1s) == replay-single test 111d: DNE: unlink striped dir, uncommit on MDT2, fail client/MDT1/MDT2 ========================================================== 18:23:27 (1713392607) SKIP: replay-single test_111d needs >= 2 MDTs SKIP 111d (1s) == replay-single test 111e: DNE: unlink striped dir, uncommit on MDT2, fail MDT1/MDT2 ========================================================== 18:23:28 (1713392608) SKIP: replay-single test_111e needs >= 2 MDTs SKIP 111e (1s) == replay-single test 111f: DNE: unlink striped dir, uncommit on MDT1, fail MDT1/MDT2 ========================================================== 18:23:29 (1713392609) SKIP: replay-single test_111f needs >= 2 MDTs SKIP 111f (1s) == replay-single test 111g: DNE: unlink striped dir, fail MDT1/MDT2 ========================================================== 18:23:31 (1713392611) SKIP: replay-single test_111g needs >= 2 MDTs SKIP 111g (1s) == replay-single test 112a: DNE: cross MDT rename, fail MDT1 ========================================================== 18:23:32 (1713392612) SKIP: replay-single test_112a needs >= 4 MDTs SKIP 112a (1s) == replay-single test 112b: DNE: cross MDT rename, fail MDT2 ========================================================== 18:23:33 (1713392613) SKIP: replay-single test_112b needs >= 4 MDTs SKIP 112b (1s) == replay-single test 112c: DNE: cross MDT rename, fail MDT3 ========================================================== 18:23:34 (1713392614) SKIP: replay-single test_112c needs >= 4 MDTs SKIP 112c (1s) == replay-single test 112d: DNE: cross MDT rename, fail MDT4 ========================================================== 18:23:35 (1713392615) SKIP: replay-single test_112d needs >= 4 MDTs SKIP 112d (1s) == replay-single test 112e: DNE: cross MDT rename, fail MDT1 and MDT2 ========================================================== 18:23:36 (1713392616) SKIP: replay-single test_112e needs >= 4 MDTs SKIP 112e (1s) == replay-single test 112f: DNE: cross MDT rename, fail MDT1 and MDT3 ========================================================== 18:23:37 (1713392617) SKIP: replay-single test_112f needs >= 4 MDTs SKIP 112f (2s) == replay-single test 112g: DNE: cross MDT rename, fail MDT1 and MDT4 ========================================================== 18:23:39 (1713392619) SKIP: replay-single test_112g needs >= 4 MDTs SKIP 112g (1s) == replay-single test 112h: DNE: cross MDT rename, fail MDT2 and MDT3 ========================================================== 18:23:40 (1713392620) SKIP: replay-single test_112h needs >= 4 MDTs SKIP 112h (1s) == replay-single test 112i: DNE: cross MDT rename, fail MDT2 and MDT4 ========================================================== 18:23:41 (1713392621) SKIP: replay-single test_112i needs >= 4 MDTs SKIP 112i (1s) == replay-single test 112j: DNE: cross MDT rename, fail MDT3 and MDT4 ========================================================== 18:23:42 (1713392622) SKIP: replay-single test_112j needs >= 4 MDTs SKIP 112j (1s) == replay-single test 112k: DNE: cross MDT rename, fail MDT1,MDT2,MDT3 ========================================================== 18:23:43 (1713392623) SKIP: replay-single test_112k needs >= 4 MDTs SKIP 112k (1s) == replay-single test 112l: DNE: cross MDT rename, fail MDT1,MDT2,MDT4 ========================================================== 18:23:44 (1713392624) SKIP: replay-single test_112l needs >= 4 MDTs SKIP 112l (1s) == replay-single test 112m: DNE: cross MDT rename, fail MDT1,MDT3,MDT4 ========================================================== 18:23:45 (1713392625) SKIP: replay-single test_112m needs >= 4 MDTs SKIP 112m (1s) == replay-single test 112n: DNE: cross MDT rename, fail MDT2,MDT3,MDT4 ========================================================== 18:23:46 (1713392626) SKIP: replay-single test_112n needs >= 4 MDTs SKIP 112n (1s) == replay-single test 115: failover for create/unlink striped directory ========================================================== 18:23:47 (1713392627) SKIP: replay-single test_115 needs >= 2 MDTs SKIP 115 (1s) == replay-single test 116a: large update log master MDT recovery ========================================================== 18:23:49 (1713392629) SKIP: replay-single test_116a needs >= 2 MDTs SKIP 116a (2s) == replay-single test 116b: large update log slave MDT recovery ========================================================== 18:23:50 (1713392630) SKIP: replay-single test_116b needs >= 2 MDTs SKIP 116b (1s) == replay-single test 117: DNE: cross MDT unlink, fail MDT1 and MDT2 ========================================================== 18:23:51 (1713392631) SKIP: replay-single test_117 needs >= 4 MDTs SKIP 117 (1s) == replay-single test 118: invalidate osp update will not cause update log corruption ========================================================== 18:23:52 (1713392632) SKIP: replay-single test_118 needs >= 2 MDTs SKIP 118 (1s) == replay-single test 119: timeout of normal replay does not cause DNE replay fails ========================================================== 18:23:53 (1713392633) SKIP: replay-single test_119 needs >= 2 MDTs SKIP 119 (1s) == replay-single test 120: DNE fail abort should stop both normal and DNE replay ========================================================== 18:23:54 (1713392634) SKIP: replay-single test_120 needs >= 2 MDTs SKIP 120 (1s) == replay-single test 121: lock replay timed out and race ========================================================== 18:23:55 (1713392635) multiop /mnt/lustre/f121.replay-single vs_s TMPPIPE=/tmp/multiop_open_wait_pipe.6853 Stopping /mnt/lustre-mds1 (opts:) on oleg428-server Failover mds1 to oleg428-server fail_loc=0x721 fail_val=0 at_max=0 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 fail_loc=0x0 at_max=600 PASS 121 (16s) == replay-single test 130a: DoM file create (setstripe) replay ========================================================== 18:24:11 (1713392651) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 17408 3337216 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 7168 3762176 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 24576 7099392 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 130a (24s) == replay-single test 130b: DoM file create (inherited) replay ========================================================== 18:24:35 (1713392675) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3712 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 17408 3337216 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 7168 3762176 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 24576 7099392 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 130b (34s) == replay-single test 131a: DoM file write lock replay === 18:25:09 (1713392709) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 17408 3337216 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 7168 3762176 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 24576 7099392 1% /mnt/lustre 1+0 records in 1+0 records out 8 bytes (8 B) copied, 0.00215826 s, 3.7 kB/s Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 131a (33s) SKIP: replay-single test_131b skipping excluded test 131b == replay-single test 132a: PFL new component instantiate replay ========================================================== 18:25:42 (1713392742) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 17408 3337216 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 7168 3762176 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 24576 7099392 1% /mnt/lustre 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0138064 s, 75.9 MB/s /mnt/lustre/f132a.replay-single lcm_layout_gen: 3 lcm_mirror_count: 1 lcm_entry_count: 2 lcme_id: 1 lcme_mirror_id: 0 lcme_flags: init lcme_extent.e_start: 0 lcme_extent.e_end: 1048576 lmm_stripe_count: 1 lmm_stripe_size: 1048576 lmm_pattern: raid0 lmm_layout_gen: 0 lmm_stripe_offset: 0 lmm_objects: - 0: { l_ost_idx: 0, l_fid: [0x100000000:0x35e2:0x0] } lcme_id: 2 lcme_mirror_id: 0 lcme_flags: init lcme_extent.e_start: 1048576 lcme_extent.e_end: EOF lmm_stripe_count: 2 lmm_stripe_size: 1048576 lmm_pattern: raid0 lmm_layout_gen: 0 lmm_stripe_offset: 1 lmm_objects: - 0: { l_ost_idx: 1, l_fid: [0x100010000:0x2982:0x0] } - 1: { l_ost_idx: 0, l_fid: [0x100000000:0x35e3:0x0] } Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f132a.replay-single lcm_layout_gen: 1 lcm_mirror_count: 1 lcm_entry_count: 2 lcme_id: 1 lcme_mirror_id: 0 lcme_flags: init lcme_extent.e_start: 0 lcme_extent.e_end: 1048576 lmm_stripe_count: 1 lmm_stripe_size: 1048576 lmm_pattern: raid0 lmm_layout_gen: 65535 lmm_stripe_offset: 0 lmm_objects: - 0: { l_ost_idx: 0, l_fid: [0x100000000:0x35e2:0x0] } lcme_id: 2 lcme_mirror_id: 0 lcme_flags: init lcme_extent.e_start: 1048576 lcme_extent.e_end: EOF lmm_stripe_count: 2 lmm_stripe_size: 1048576 lmm_pattern: raid0 lmm_layout_gen: 65535 lmm_stripe_offset: 1 lmm_objects: - 0: { l_ost_idx: 1, l_fid: [0x100010000:0x2982:0x0] } - 1: { l_ost_idx: 0, l_fid: [0x100000000:0x35e3:0x0] } PASS 132a (25s) == replay-single test 133: check resend of ongoing requests for lwp during failover ========================================================== 18:26:07 (1713392767) SKIP: replay-single test_133 needs >= 2 MDTs SKIP 133 (0s) == replay-single test 134: replay creation of a file created in a pool ========================================================== 18:26:07 (1713392767) Creating new pool oleg428-server: Pool lustre.pool_134 created Adding targets to pool oleg428-server: OST lustre-OST0001_UUID added to pool lustre.pool_134 Waiting 90s for 'lustre-OST0001_UUID ' UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 18432 3750912 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 7168 3762176 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 25600 7513088 1% /mnt/lustre Failing mds1 on oleg428-server Stopping /mnt/lustre-mds1 (opts:) on oleg428-server reboot facets: mds1 Failover mds1 to oleg428-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg428-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Destroy the created pools: pool_134 lustre.pool_134 oleg428-server: OST lustre-OST0001_UUID removed from pool lustre.pool_134 oleg428-server: Pool lustre.pool_134 destroyed Waiting 90s for 'foo' PASS 134 (44s) == replay-single test 135: Server failure in lock replay phase ========================================================== 18:26:51 (1713392811) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 18432 3750912 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 7168 3762176 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 25600 7513088 1% /mnt/lustre Stopping /mnt/lustre-ost1 (opts:) on oleg428-server Failover ost1 to oleg428-server oleg428-server: oleg428-server.virtnet: executing load_module ../libcfs/libcfs/libcfs fail_val=20 fail_loc=0x32d Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-OST0000 Stopping /mnt/lustre-ost1 (opts:) on oleg428-server Failover ost1 to oleg428-server oleg428-server: oleg428-server.virtnet: executing load_module ../libcfs/libcfs/libcfs fail_loc=0 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-OST0000 Stopping /mnt/lustre-ost1 (opts:-f) on oleg428-server Stopping /mnt/lustre-ost2 (opts:-f) on oleg428-server Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-OST0000 Starting ost2: -o localrecov lustre-ost2/ost2 /mnt/lustre-ost2 oleg428-server: oleg428-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg428-client: oleg428-server: ssh exited with exit code 1 Started lustre-OST0001 PASS 135 (79s) == replay-single test complete, duration 5205 sec ======== 18:28:10 (1713392890)