-----============= acceptance-small: replay-single ============----- Wed Apr 17 04:43:46 EDT 2024 excepting tests: 110f 131b 59 oleg412-client.virtnet: executing check_config_client /mnt/lustre oleg412-client.virtnet: Checking config lustre mounted on /mnt/lustre Checking servers environments Checking clients oleg412-client.virtnet environments Using TIMEOUT=20 osc.lustre-OST0000-osc-ffff8800b61dc800.idle_timeout=debug osc.lustre-OST0001-osc-ffff8800b61dc800.idle_timeout=debug disable quota as required oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 == replay-single test 0a: empty replay =================== 04:43:55 (1713343435) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 0a (26s) == replay-single test 0b: ensure object created after recover exists. (3284) ========================================================== 04:44:21 (1713343461) Failing ost1 on oleg412-server Stopping /mnt/lustre-ost1 (opts:) on oleg412-server reboot facets: ost1 Failover ost1 to oleg412-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-OST0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec total: 20 open/close in 0.04 seconds: 452.07 ops/second - unlinked 0 (time 1713343479 ; total 0 ; last 0) total: 20 unlinks in 0 seconds: inf unlinks/second PASS 0b (19s) == replay-single test 0c: check replay-barrier =========== 04:44:40 (1713343480) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 Starting client: oleg412-client.virtnet: -o user_xattr,flock oleg412-server@tcp:/lustre /mnt/lustre rm: cannot remove '/mnt/lustre/f0c.replay-single': No such file or directory PASS 0c (81s) == replay-single test 0d: expired recovery with no clients ========================================================== 04:46:01 (1713343561) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 Starting client: oleg412-client.virtnet: -o user_xattr,flock oleg412-server@tcp:/lustre /mnt/lustre PASS 0d (77s) == replay-single test 1: simple create =================== 04:47:18 (1713343638) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f1.replay-single has type file OK PASS 1 (31s) == replay-single test 2a: touch ========================== 04:47:49 (1713343669) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f2a.replay-single has type file OK PASS 2a (23s) == replay-single test 2b: touch ========================== 04:48:12 (1713343692) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f2b.replay-single has type file OK PASS 2b (22s) == replay-single test 2c: setstripe replay =============== 04:48:34 (1713343714) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f2c.replay-single has type file OK PASS 2c (23s) == replay-single test 2d: setdirstripe replay ============ 04:48:57 (1713343737) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d2d.replay-single has type dir OK PASS 2d (22s) == replay-single test 2e: O_CREAT|O_EXCL create replay === 04:49:19 (1713343759) fail_loc=0x8000013b UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 Succeed in opening file "/mnt/lustre/f2e.replay-single"(flags=O_CREAT) oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f2e.replay-single has type file OK PASS 2e (34s) == replay-single test 3a: replay failed open(O_DIRECTORY) ========================================================== 04:49:53 (1713343793) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Error in opening file "/mnt/lustre/f3a.replay-single"(flags=O_DIRECTORY) 20: Not a directory Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f3a.replay-single has type file OK PASS 3a (24s) == replay-single test 3b: replay failed open -ENOMEM ===== 04:50:17 (1713343817) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre fail_loc=0x80000114 touch: cannot touch '/mnt/lustre/f3b.replay-single': Cannot allocate memory fail_loc=0 Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/f3b.replay-single: No such file or directory PASS 3b (24s) == replay-single test 3c: replay failed open -ENOMEM ===== 04:50:41 (1713343841) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre fail_loc=0x80000128 touch: cannot touch '/mnt/lustre/f3c.replay-single': Cannot allocate memory fail_loc=0 Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/f3c.replay-single: No such file or directory PASS 3c (32s) == replay-single test 4a: |x| 10 open(O_CREAT)s ========== 04:51:13 (1713343873) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 4a (24s) == replay-single test 4b: |x| rm 10 files ================ 04:51:37 (1713343897) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3764224 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3764224 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7528448 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/f4b.replay-single-*: No such file or directory PASS 4b (22s) == replay-single test 5: |x| 220 open(O_CREAT) =========== 04:51:59 (1713343919) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3200 2205440 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 5 (40s) == replay-single test 6a: mkdir + contained create ======= 04:52:39 (1713343959) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d6a.replay-single has type dir OK /mnt/lustre/d6a.replay-single/f6a.replay-single has type file OK PASS 6a (35s) == replay-single test 6b: |X| rmdir ====================== 04:53:14 (1713343994) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/d6b.replay-single: No such file or directory PASS 6b (32s) == replay-single test 7: mkdir |X| contained create ====== 04:53:47 (1713344027) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d7.replay-single has type dir OK /mnt/lustre/d7.replay-single/f7.replay-single has type file OK PASS 7 (33s) == replay-single test 8: creat open |X| close ============ 04:54:20 (1713344060) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre multiop /mnt/lustre/f8.replay-single vmo_c TMPPIPE=/tmp/multiop_open_wait_pipe.6842 Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f8.replay-single /mnt/lustre/f8.replay-single has type file OK PASS 8 (23s) == replay-single test 9: |X| create (same inum/gen) ====== 04:54:43 (1713344083) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec old_inum == 144115305935798546, new_inum == 144115305935798546 old_inum and new_inum match PASS 9 (34s) == replay-single test 10: create |X| rename unlink ======= 04:55:17 (1713344117) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/f10.replay-single: No such file or directory PASS 10 (34s) == replay-single test 11: create open write rename |X| create-old-name read ========================================================== 04:55:51 (1713344151) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre new old Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec new old PASS 11 (27s) == replay-single test 12: open, unlink |X| close ========= 04:56:18 (1713344178) multiop /mnt/lustre/f12.replay-single vo_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6842 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 12 (26s) == replay-single test 13: open chmod 0 |x| write close === 04:56:44 (1713344204) multiop /mnt/lustre/f13.replay-single vO_wc TMPPIPE=/tmp/multiop_open_wait_pipe.6842 /mnt/lustre/f13.replay-single has perms 00 OK UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f13.replay-single has perms 00 OK /mnt/lustre/f13.replay-single has size 1 OK PASS 13 (34s) == replay-single test 14: open(O_CREAT), unlink |X| close ========================================================== 04:57:19 (1713344239) multiop /mnt/lustre/f14.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6842 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 14 (34s) == replay-single test 15: open(O_CREAT), unlink |X| touch new, close ========================================================== 04:57:52 (1713344272) multiop /mnt/lustre/f15.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6842 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 15 (33s) == replay-single test 16: |X| open(O_CREAT), unlink, touch new, unlink new ========================================================== 04:58:25 (1713344305) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 16 (26s) == replay-single test 17: |X| open(O_CREAT), |replay| close ========================================================== 04:58:51 (1713344331) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre multiop /mnt/lustre/f17.replay-single vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.6842 Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f17.replay-single has type file OK PASS 17 (30s) == replay-single test 18: open(O_CREAT), unlink, touch new, close, touch, unlink ========================================================== 04:59:21 (1713344361) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3200 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre multiop /mnt/lustre/f18.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6842 pid: 21420 will close Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 18 (36s) == replay-single test 19: mcreate, open, write, rename === 04:59:57 (1713344397) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre old Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec old PASS 19 (36s) == replay-single test 20a: |X| open(O_CREAT), unlink, replay, close (test mds_cleanup_orphans) ========================================================== 05:00:33 (1713344433) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre multiop /mnt/lustre/f20a.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6842 Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 20a (38s) == replay-single test 20b: write, unlink, eviction, replay (test mds_cleanup_orphans) ========================================================== 05:01:11 (1713344471) /mnt/lustre/f20b.replay-single lmm_stripe_count: 1 lmm_stripe_size: 1048576 lmm_pattern: raid0 lmm_layout_gen: 0 lmm_stripe_offset: 0 obdidx objid objid group 0 1122 0x462 0 dd: error writing '/mnt/lustre/f20b.replay-single': Cannot send after transport endpoint shutdown oleg412-client: error: invalid path '/mnt/lustre': Input/output error 9357+0 records in 9356+0 records out 38322176 bytes (38 MB) copied, 2.07016 s, 18.5 MB/s pdsh@oleg412-client: oleg412-client: ssh exited with exit code 5 Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec affected facets: mds1 oleg412-server: oleg412-server.virtnet: executing _wait_recovery_complete *.lustre-MDT0000.recovery_status 1475 oleg412-server: *.lustre-MDT0000.recovery_status status: COMPLETE sleep 5 for ZFS zfs Waiting for MDT destroys to complete before 3072, after 6144 PASS 20b (48s) == replay-single test 20c: check that client eviction does not affect file content ========================================================== 05:01:59 (1713344519) multiop /mnt/lustre/f20c.replay-single vOw_c TMPPIPE=/tmp/multiop_open_wait_pipe.6842 -rw-r--r-- 1 root root 1 Apr 17 05:02 /mnt/lustre/f20c.replay-single oleg412-client: error: invalid path '/mnt/lustre': Input/output error pdsh@oleg412-client: oleg412-client: ssh exited with exit code 5 PASS 20c (6s) == replay-single test 21: |X| open(O_CREAT), unlink touch new, replay, close (test mds_cleanup_orphans) ========================================================== 05:02:05 (1713344525) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre multiop /mnt/lustre/f21.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6842 Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 21 (35s) == replay-single test 22: open(O_CREAT), |X| unlink, replay, close (test mds_cleanup_orphans) ========================================================== 05:02:40 (1713344560) multiop /mnt/lustre/f22.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6842 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 22 (37s) == replay-single test 23: open(O_CREAT), |X| unlink touch new, replay, close (test mds_cleanup_orphans) ========================================================== 05:03:17 (1713344597) multiop /mnt/lustre/f23.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6842 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 23 (34s) == replay-single test 24: open(O_CREAT), replay, unlink, close (test mds_cleanup_orphans) ========================================================== 05:03:51 (1713344631) multiop /mnt/lustre/f24.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6842 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 24 (34s) == replay-single test 25: open(O_CREAT), unlink, replay, close (test mds_cleanup_orphans) ========================================================== 05:04:25 (1713344665) multiop /mnt/lustre/f25.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6842 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 25 (35s) == replay-single test 26: |X| open(O_CREAT), unlink two, close one, replay, close one (test mds_cleanup_orphans) ========================================================== 05:05:00 (1713344700) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre multiop /mnt/lustre/f26.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6842 multiop /mnt/lustre/f26.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6842 Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 26 (29s) == replay-single test 27: |X| open(O_CREAT), unlink two, replay, close two (test mds_cleanup_orphans) ========================================================== 05:05:29 (1713344729) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre multiop /mnt/lustre/f27.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6842 multiop /mnt/lustre/f27.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6842 Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 27 (28s) == replay-single test 28: open(O_CREAT), |X| unlink two, close one, replay, close one (test mds_cleanup_orphans) ========================================================== 05:05:57 (1713344757) multiop /mnt/lustre/f28.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6842 multiop /mnt/lustre/f28.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6842 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 28 (32s) == replay-single test 29: open(O_CREAT), |X| unlink two, replay, close two (test mds_cleanup_orphans) ========================================================== 05:06:29 (1713344789) multiop /mnt/lustre/f29.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6842 multiop /mnt/lustre/f29.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6842 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 29 (35s) == replay-single test 30: open(O_CREAT) two, unlink two, replay, close two (test mds_cleanup_orphans) ========================================================== 05:07:04 (1713344824) multiop /mnt/lustre/f30.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6842 multiop /mnt/lustre/f30.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6842 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 30 (35s) == replay-single test 31: open(O_CREAT) two, unlink one, |X| unlink one, close two (test mds_cleanup_orphans) ========================================================== 05:07:39 (1713344859) multiop /mnt/lustre/f31.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6842 multiop /mnt/lustre/f31.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6842 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 31 (33s) == replay-single test 32: close() notices client eviction; close() after client eviction ========================================================== 05:08:12 (1713344892) multiop /mnt/lustre/f32.replay-single vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.6842 multiop /mnt/lustre/f32.replay-single vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.6842 oleg412-client: error: invalid path '/mnt/lustre': Input/output error pdsh@oleg412-client: oleg412-client: ssh exited with exit code 5 PASS 32 (6s) == replay-single test 33a: fid seq shouldn't be reused after abort recovery ========================================================== 05:08:18 (1713344898) total: 10 open/close in 0.07 seconds: 145.77 ops/second Replay barrier on lustre-MDT0000 Stopping /mnt/lustre-mds1 (opts:) on oleg412-server Failover mds1 to oleg412-server Starting mds1: -o localrecov -o abort_recovery lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 total: 10 open/close in 0.07 seconds: 136.94 ops/second PASS 33a (12s) == replay-single test 33b: test fid seq allocation ======= 05:08:30 (1713344910) fail_loc=0x1311 total: 10 open/close in 0.08 seconds: 128.64 ops/second Replay barrier on lustre-MDT0000 Stopping /mnt/lustre-mds1 (opts:) on oleg412-server Failover mds1 to oleg412-server Starting mds1: -o localrecov -o abort_recovery lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 total: 10 open/close in 0.07 seconds: 137.34 ops/second PASS 33b (12s) == replay-single test 34: abort recovery before client does replay (test mds_cleanup_orphans) ========================================================== 05:08:42 (1713344922) multiop /mnt/lustre/f34.replay-single vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.6842 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Stopping /mnt/lustre-mds1 (opts:) on oleg412-server Failover mds1 to oleg412-server Starting mds1: -o localrecov -o abort_recovery lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 PASS 34 (13s) == replay-single test 35: test recovery from llog for unlink op ========================================================== 05:08:55 (1713344935) fail_loc=0x80000119 Stopping /mnt/lustre-mds1 (opts:) on oleg412-server Failover mds1 to oleg412-server Starting mds1: -o localrecov -o abort_recovery lustre-mdt1/mdt1 /mnt/lustre-mds1 rm: cannot remove '/mnt/lustre/f35.replay-single': Input/output error oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 Can't lstat /mnt/lustre/f35.replay-single: No such file or directory PASS 35 (13s) == replay-single test 36: don't resend cancel ============ 05:09:08 (1713344948) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 PASS 36 (21s) == replay-single test 37: abort recovery before client does replay (test mds_cleanup_orphans for directories) ========================================================== 05:09:29 (1713344969) multiop /mnt/lustre/d37.replay-single/f37.replay-single vdD_c TMPPIPE=/tmp/multiop_open_wait_pipe.6842 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Stopping /mnt/lustre-mds1 (opts:) on oleg412-server Failover mds1 to oleg412-server Starting mds1: -o localrecov -o abort_recovery lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 PASS 37 (12s) == replay-single test 38: test recovery from unlink llog (test llog_gen_rec) ========================================================== 05:09:41 (1713344981) total: 800 open/close in 3.79 seconds: 211.30 ops/second - unlinked 0 (time 1713344987 ; total 0 ; last 0) total: 400 unlinks in 1 seconds: 400.000000 unlinks/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec - unlinked 0 (time 1713345020 ; total 0 ; last 0) total: 400 unlinks in 2 seconds: 200.000000 unlinks/second Can't lstat /mnt/lustre/f38.replay-single-*: No such file or directory PASS 38 (45s) == replay-single test 39: test recovery from unlink llog (test llog_gen_rec) ========================================================== 05:10:26 (1713345026) total: 800 open/close in 3.64 seconds: 219.57 ops/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre - unlinked 0 (time 1713345033 ; total 0 ; last 0) total: 400 unlinks in 1 seconds: 400.000000 unlinks/second Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec - unlinked 0 (time 1713345061 ; total 0 ; last 0) total: 400 unlinks in 1 seconds: 400.000000 unlinks/second Can't lstat /mnt/lustre/f39.replay-single-*: No such file or directory PASS 39 (40s) == replay-single test 40: cause recovery in ptlrpc, ensure IO continues ========================================================== 05:11:06 (1713345066) SKIP: replay-single test_40 layout_lock needs MDS connection for IO SKIP 40 (1s) == replay-single test 41: read from a valid osc while other oscs are invalid ========================================================== 05:11:07 (1713345067) 1+0 records in 1+0 records out 4096 bytes (4.1 kB) copied, 0.00279746 s, 1.5 MB/s 1+0 records in 1+0 records out 4096 bytes (4.1 kB) copied, 0.00565385 s, 724 kB/s PASS 41 (3s) == replay-single test 42: recovery after ost failure ===== 05:11:10 (1713345070) total: 800 open/close in 3.84 seconds: 208.15 ops/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3456 2205184 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre - unlinked 0 (time 1713345077 ; total 0 ; last 0) total: 400 unlinks in 1 seconds: 400.000000 unlinks/second debug=-1 Failing ost1 on oleg412-server Stopping /mnt/lustre-ost1 (opts:) on oleg412-server reboot facets: ost1 Failover ost1 to oleg412-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-OST0000 wait for MDS to timeout and recover debug=super ioctl neterror warning dlmtrace error emerg ha rpctrace vfstrace config console lfsck debug=super ioctl neterror warning dlmtrace error emerg ha rpctrace vfstrace config console lfsck - unlinked 0 (time 1713345135 ; total 0 ; last 0) total: 400 unlinks in 0 seconds: inf unlinks/second Can't lstat /mnt/lustre/f42.replay-single-*: No such file or directory PASS 42 (67s) == replay-single test 43: mds osc import failure during recovery; don't LBUG ========================================================== 05:12:17 (1713345137) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3456 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre fail_loc=0x80000204 Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 43 (36s) == replay-single test 44a: race in target handle connect ========================================================== 05:12:53 (1713345173) at_max=40 1 of 10 (1713345175) service : cur 1 worst 1 (at 1713345156, 18s ago) 1 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3456 2205184 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre 2 of 10 (1713345181) service : cur 5 worst 5 (at 1713345180, 0s ago) 5 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3456 2205184 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre 3 of 10 (1713345186) service : cur 5 worst 5 (at 1713345180, 6s ago) 5 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3456 2205184 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre 4 of 10 (1713345192) service : cur 5 worst 5 (at 1713345180, 11s ago) 5 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3456 2205184 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre 5 of 10 (1713345198) service : cur 5 worst 5 (at 1713345180, 17s ago) 5 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3456 2205184 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre 6 of 10 (1713345203) service : cur 5 worst 5 (at 1713345180, 23s ago) 5 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3456 2205184 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre 7 of 10 (1713345209) service : cur 5 worst 5 (at 1713345180, 28s ago) 5 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3456 2205184 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre 8 of 10 (1713345215) service : cur 5 worst 5 (at 1713345180, 34s ago) 5 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3456 2205184 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre 9 of 10 (1713345220) service : cur 5 worst 5 (at 1713345180, 39s ago) 5 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3456 2205184 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre 10 of 10 (1713345226) service : cur 5 worst 5 (at 1713345180, 45s ago) 5 0 0 0 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3456 2205184 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre fail_loc=0 at_max=600 PASS 44a (60s) == replay-single test 44b: race in target handle connect ========================================================== 05:13:53 (1713345233) 1 of 10 (1713345234) service : cur 5 worst 5 (at 1713345180, 53s ago) 5 0 0 0 fail_loc=0x80000704 Filesystem 1K-blocks Used Available Use% Mounted on 192.168.204.112@tcp:/lustre 7542784 6144 7532544 1% /mnt/lustre 2 of 10 (1713345275) service : cur 40 worst 40 (at 1713345273, 1s ago) 40 0 0 0 fail_loc=0x80000704 df: '/mnt/lustre': Operation already in progress 3 of 10 (1713345282) service : cur 40 worst 40 (at 1713345273, 8s ago) 40 0 0 0 fail_loc=0x80000704 df: '/mnt/lustre': Operation already in progress 4 of 10 (1713345282) service : cur 40 worst 40 (at 1713345273, 9s ago) 40 0 0 0 fail_loc=0x80000704 df: '/mnt/lustre': Operation already in progress 5 of 10 (1713345283) service : cur 40 worst 40 (at 1713345273, 9s ago) 40 0 0 0 fail_loc=0x80000704 df: '/mnt/lustre': Operation already in progress 6 of 10 (1713345283) service : cur 40 worst 40 (at 1713345273, 10s ago) 40 0 0 0 fail_loc=0x80000704 df: '/mnt/lustre': Operation already in progress 7 of 10 (1713345284) service : cur 40 worst 40 (at 1713345273, 10s ago) 40 0 0 0 fail_loc=0x80000704 df: '/mnt/lustre': Operation already in progress 8 of 10 (1713345285) service : cur 40 worst 40 (at 1713345273, 11s ago) 40 0 0 0 fail_loc=0x80000704 df: '/mnt/lustre': Operation already in progress 9 of 10 (1713345285) service : cur 40 worst 40 (at 1713345273, 11s ago) 40 0 0 0 fail_loc=0x80000704 df: '/mnt/lustre': Operation already in progress 10 of 10 (1713345286) service : cur 40 worst 40 (at 1713345273, 12s ago) 40 0 0 0 fail_loc=0x80000704 df: '/mnt/lustre': Operation already in progress PASS 44b (55s) == replay-single test 44c: race in target handle connect ========================================================== 05:14:48 (1713345288) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3456 2205184 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre total: 100 create in 0.18 seconds: 554.72 ops/second fail_loc=0x80000712 Stopping /mnt/lustre-mds1 (opts:) on oleg412-server Failover mds1 to oleg412-server Starting mds1: -o localrecov -o abort_recovery lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 unlink(/mnt/lustre/f44c.replay-single-0) error: No such file or directory total: 0 unlinks in 0 seconds: -nan unlinks/second Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec unlink(/mnt/lustre/f44c.replay-single-0) error: No such file or directory total: 0 unlinks in 0 seconds: -nan unlinks/second PASS 44c (32s) == replay-single test 45: Handle failed close ============ 05:15:21 (1713345321) multiop /mnt/lustre/f45.replay-single vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.6842 /mnt/lustre/f45.replay-single has type file OK PASS 45 (2s) == replay-single test 46: Don't leak file handle after open resend (3325) ========================================================== 05:15:23 (1713345323) fail_loc=0x122 fail_loc=0 Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec lfs path2fid: cannot get fid for 'f46.replay-single': No such file or directory PASS 46 (31s) == replay-single test 47: MDS->OSC failure during precreate cleanup (2824) ========================================================== 05:15:54 (1713345354) total: 20 open/close in 0.08 seconds: 258.27 ops/second Failing ost1 on oleg412-server Stopping /mnt/lustre-ost1 (opts:) on oleg412-server reboot facets: ost1 Failover ost1 to oleg412-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-OST0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec fail_loc=0x80000204 total: 20 open/close in 0.06 seconds: 316.79 ops/second - unlinked 0 (time 1713345434 ; total 0 ; last 0) total: 20 unlinks in 0 seconds: inf unlinks/second PASS 47 (81s) == replay-single test 48: MDS->OSC failure during precreate cleanup (2824) ========================================================== 05:17:15 (1713345435) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3328 2205312 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre total: 20 open/close in 0.05 seconds: 364.06 ops/second Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 fail_loc=0x80000216 total: 20 open/close in 0.05 seconds: 429.26 ops/second - unlinked 0 (time 1713345523 ; total 0 ; last 0) total: 40 unlinks in 0 seconds: inf unlinks/second PASS 48 (89s) == replay-single test 50: Double OSC recovery, don't LASSERT (3812) ========================================================== 05:18:45 (1713345525) PASS 50 (7s) == replay-single test 52: time out lock replay (3764) ==== 05:18:52 (1713345532) multiop /mnt/lustre/f52.replay-single vs_s TMPPIPE=/tmp/multiop_open_wait_pipe.6842 fail_loc=0x80000157 Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec fail_loc=0x0 fail_loc=0x0 PASS 52 (36s) == replay-single test 53a: |X| close request while two MDC requests in flight ========================================================== 05:19:29 (1713345569) fail_loc=0x80000115 fail_loc=0 Replay barrier on lustre-MDT0000 Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d53a.replay-single-1/f has type file OK /mnt/lustre/d53a.replay-single-2/f has type file OK PASS 53a (29s) == replay-single test 53b: |X| open request while two MDC requests in flight ========================================================== 05:19:58 (1713345598) multiop /mnt/lustre/d53b.replay-single-1/f vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.6842 fail_loc=0x80000107 fail_loc=0 Replay barrier on lustre-MDT0000 Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d53b.replay-single-1/f has type file OK /mnt/lustre/d53b.replay-single-2/f has type file OK PASS 53b (29s) == replay-single test 53c: |X| open request and close request while two MDC requests in flight ========================================================== 05:20:27 (1713345627) fail_loc=0x80000107 fail_loc=0x80000115 Replay barrier on lustre-MDT0000 Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 fail_loc=0 /mnt/lustre/d53c.replay-single-1/f has type file OK /mnt/lustre/d53c.replay-single-2/f has type file OK PASS 53c (27s) == replay-single test 53d: close reply while two MDC requests in flight ========================================================== 05:20:54 (1713345654) fail_loc=0x8000013b fail_loc=0 Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d53d.replay-single-1/f has type file OK /mnt/lustre/d53d.replay-single-2/f has type file OK PASS 53d (30s) == replay-single test 53e: |X| open reply while two MDC requests in flight ========================================================== 05:21:24 (1713345684) fail_loc=0x119 fail_loc=0 Replay barrier on lustre-MDT0000 Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d53e.replay-single-1/f has type file OK /mnt/lustre/d53e.replay-single-2/f has type file OK PASS 53e (29s) == replay-single test 53f: |X| open reply and close reply while two MDC requests in flight ========================================================== 05:21:53 (1713345713) fail_loc=0x119 fail_loc=0x8000013b Replay barrier on lustre-MDT0000 Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 fail_loc=0 /mnt/lustre/d53f.replay-single-1/f has type file OK /mnt/lustre/d53f.replay-single-2/f has type file OK PASS 53f (27s) == replay-single test 53g: |X| drop open reply and close request while close and open are both in flight ========================================================== 05:22:20 (1713345740) fail_loc=0x119 fail_loc=0x80000115 fail_loc=0 Replay barrier on lustre-MDT0000 Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 /mnt/lustre/d53g.replay-single-1/f has type file OK /mnt/lustre/d53g.replay-single-2/f has type file OK PASS 53g (28s) == replay-single test 53h: open request and close reply while two MDC requests in flight ========================================================== 05:22:48 (1713345768) fail_loc=0x80000107 fail_loc=0x8000013b Replay barrier on lustre-MDT0000 Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 fail_loc=0 /mnt/lustre/d53h.replay-single-1/f has type file OK /mnt/lustre/d53h.replay-single-2/f has type file OK PASS 53h (28s) == replay-single test 55: let MDS_CHECK_RESENT return the original return code instead of 0 ========================================================== 05:23:16 (1713345796) fail_loc=0x8000012b fail_loc=0x0 touch: cannot touch '/mnt/lustre/f55.replay-single': No such file or directory rm: cannot remove '/mnt/lustre/f55.replay-single': No such file or directory PASS 55 (9s) == replay-single test 56: don't replay a symlink open request (3440) ========================================================== 05:23:25 (1713345805) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3456 2205184 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 56 (33s) == replay-single test 57: test recovery from llog for setattr op ========================================================== 05:23:58 (1713345838) fail_loc=0x8000012c UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3328 2205184 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec affected facets: mds1 oleg412-server: oleg412-server.virtnet: executing _wait_recovery_complete *.lustre-MDT0000.recovery_status 1475 oleg412-server: *.lustre-MDT0000.recovery_status status: COMPLETE Waiting for orphan cleanup... osp.lustre-OST0000-osc-MDT0000.old_sync_processed osp.lustre-OST0001-osc-MDT0000.old_sync_processed wait 40 secs maximumly for oleg412-server mds-ost sync done. /mnt/lustre/f57.replay-single has type file OK fail_loc=0x0 PASS 57 (35s) == replay-single test 58a: test recovery from llog for setattr op (test llog_gen_rec) ========================================================== 05:24:33 (1713345873) fail_loc=0x8000012c total: 2500 open/close in 5.14 seconds: 486.79 ops/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3712 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec fail_loc=0x0 - unlinked 0 (time 1713345909 ; total 0 ; last 0) total: 2500 unlinks in 3 seconds: 833.333313 unlinks/second PASS 58a (41s) == replay-single test 58b: test replay of setxattr op ==== 05:25:14 (1713345914) Starting client: oleg412-client.virtnet: -o user_xattr,flock oleg412-server@tcp:/lustre /mnt/lustre2 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3712 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Stopping client oleg412-client.virtnet /mnt/lustre2 (opts:) oleg412-client.virtnet: executing wait_import_state_mount FULL mgc.*.mgs_server_uuid mgc.*.mgs_server_uuid in FULL state after 0 sec PASS 58b (35s) == replay-single test 58c: resend/reconstruct setxattr op ========================================================== 05:25:49 (1713345949) Starting client: oleg412-client.virtnet: -o user_xattr,flock oleg412-server@tcp:/lustre /mnt/lustre2 fail_val=0 fail_loc=0x123 fail_loc=0 fail_loc=0x119 fail_loc=0 Stopping client oleg412-client.virtnet /mnt/lustre2 (opts:) PASS 58c (22s) SKIP: replay-single test_59 skipping ALWAYS excluded test 59 == replay-single test 60: test llog post recovery init vs llog unlink ========================================================== 05:26:11 (1713345971) total: 200 open/close in 0.49 seconds: 407.52 ops/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre - unlinked 0 (time 1713345974 ; total 0 ; last 0) total: 100 unlinks in 0 seconds: inf unlinks/second Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec - unlinked 0 (time 1713346004 ; total 0 ; last 0) total: 100 unlinks in 0 seconds: inf unlinks/second PASS 60 (35s) == replay-single test 61a: test race llog recovery vs llog cleanup ========================================================== 05:26:46 (1713346006) total: 800 open/close in 1.72 seconds: 464.80 ops/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre - unlinked 0 (time 1713346010 ; total 0 ; last 0) total: 800 unlinks in 1 seconds: 800.000000 unlinks/second fail_val=0 fail_loc=0x80000221 Failing ost1 on oleg412-server Stopping /mnt/lustre-ost1 (opts:) on oleg412-server reboot facets: ost1 Failover ost1 to oleg412-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-OST0000 Failing ost1 on oleg412-server Stopping /mnt/lustre-ost1 (opts:) on oleg412-server reboot facets: ost1 Failover ost1 to oleg412-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-OST0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec fail_val=0 fail_loc=0x0 Can't lstat /mnt/lustre/d61a.replay-single/f61a.replay-single-*: No such file or directory PASS 61a (80s) == replay-single test 61b: test race mds llog sync vs llog cleanup ========================================================== 05:28:06 (1713346086) fail_loc=0x8000013a Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec 1+0 records in 1+0 records out 4096 bytes (4.1 kB) copied, 0.00243498 s, 1.7 MB/s PASS 61b (56s) == replay-single test 61c: test race mds llog sync vs llog cleanup ========================================================== 05:29:02 (1713346142) fail_val=0 fail_loc=0x80000222 Failing ost1 on oleg412-server Stopping /mnt/lustre-ost1 (opts:) on oleg412-server reboot facets: ost1 Failover ost1 to oleg412-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-OST0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec fail_val=0 fail_loc=0x0 PASS 61c (30s) == replay-single test 61d: error in llog_setup should cleanup the llog context correctly ========================================================== 05:29:32 (1713346172) Stopping /mnt/lustre-mds1 (opts:) on oleg412-server fail_loc=0x80000605 Starting mgs: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: mount.lustre: mount lustre-mdt1/mdt1 at /mnt/lustre-mds1 failed: Operation not supported pdsh@oleg412-client: oleg412-server: ssh exited with exit code 95 Start of lustre-mdt1/mdt1 on mgs failed 95 fail_loc=0 Starting mgs: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 PASS 61d (8s) == replay-single test 62: don't mis-drop resent replay === 05:29:40 (1713346180) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3584 2205056 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre total: 25 open/close in 0.08 seconds: 331.84 ops/second fail_loc=0x80000707 Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec fail_loc=0 - unlinked 0 (time 1713346207 ; total 0 ; last 0) total: 25 unlinks in 0 seconds: inf unlinks/second PASS 62 (28s) == replay-single test 65a: AT: verify early replies ====== 05:30:08 (1713346208) at_history=8 at_history=8 debug=other fail_val=6000 fail_loc=0x8000050a 00000100:00001000:3.0:1713346232.099458:0:5822:0:(client.c:514:ptlrpc_at_recv_early_reply()) @@@ Early reply #1, new deadline in 31s (25s) req@ffff88012b08b440 x1796570744208512/t0(0) o101->lustre-MDT0000-mdc-ffff8800aaf30800@192.168.204.112@tcp:12/10 lens 664/66264 e 1 to 0 dl 1713346263 ref 2 fl Rpc:PQr/0/ffffffff rc 0/-1 job:'createmany.0' portal 12 : cur 31 worst 40 (at 1713345275, 962s ago) 31 1 1 1 portal 23 : cur 1 worst 10 (at 1713343688, 2549s ago) 10 1 1 10 portal 30 : cur 1 worst 1 (at 1713343639, 2598s ago) 1 0 0 1 portal 17 : cur 1 worst 1 (at 1713343715, 2522s ago) 1 1 1 0 portal 12 : cur 6 worst 40 (at 1713345275, 971s ago) 31 1 1 1 portal 23 : cur 1 worst 10 (at 1713343688, 2558s ago) 10 1 1 10 portal 30 : cur 1 worst 1 (at 1713343639, 2607s ago) 1 0 0 1 portal 17 : cur 1 worst 1 (at 1713343715, 2531s ago) 1 1 1 0 PASS 65a (39s) == replay-single test 65b: AT: verify early replies on packed reply / bulk ========================================================== 05:30:47 (1713346247) at_history=8 at_history=8 debug=other trace fail_val=6 fail_loc=0x224 fail_loc=0 00000100:00001000:1.0:1713346271.425380:0:1835:0:(client.c:514:ptlrpc_at_recv_early_reply()) @@@ Early reply #1, new deadline in 31s (25s) req@ffff8800a7974280 x1796570744216064/t0(0) o4->lustre-OST0000-osc-ffff8800aaf30800@192.168.204.112@tcp:6/4 lens 4584/448 e 1 to 0 dl 1713346302 ref 2 fl Rpc:Qr/0/ffffffff rc 0/-1 job:'multiop.0' debug=super ioctl neterror warning dlmtrace error emerg ha rpctrace vfstrace config console lfsck debug=super ioctl neterror warning dlmtrace error emerg ha rpctrace vfstrace config console lfsck portal 28 : cur 1 worst 31 (at 1713344277, 2000s ago) 1 1 1 1 portal 7 : cur 1 worst 1 (at 1713343637, 2640s ago) 1 1 1 1 portal 17 : cur 1 worst 1 (at 1713343693, 2584s ago) 1 0 1 1 portal 6 : cur 31 worst 31 (at 1713346276, 1s ago) 31 1 0 0 PASS 65b (31s) == replay-single test 66a: AT: verify MDT service time adjusts with no early replies ========================================================== 05:31:18 (1713346278) at_history=8 at_history=8 portal 12 : cur 1 worst 40 (at 1713345275, 1025s ago) 31 1 1 1 fail_val=5000 fail_loc=0x8000050a portal 12 : cur 31 worst 40 (at 1713345275, 1030s ago) 31 1 1 1 fail_val=10000 fail_loc=0x8000050a portal 12 : cur 10 worst 40 (at 1713345275, 1041s ago) 31 1 1 1 fail_loc=0 portal 12 : cur 1 worst 40 (at 1713345275, 1050s ago) 31 1 1 1 Current MDT timeout 1, worst 40 PASS 66a (48s) == replay-single test 66b: AT: verify net latency adjusts ========================================================== 05:32:06 (1713346326) at_history=8 at_history=8 fail_val=6 fail_loc=0x50c fail_loc=0 network timeout orig 1, cur 6, worst 6 PASS 66b (47s) == replay-single test 67a: AT: verify slow request processing doesn't induce reconnects ========================================================== 05:32:53 (1713346373) at_history=8 at_history=8 fail_val=400 fail_loc=0x50a fail_loc=0 0 osc reconnect attempts on gradual slow PASS 67a (54s) == replay-single test 67b: AT: verify instant slowdown doesn't induce reconnects ========================================================== 05:33:47 (1713346427) at_history=8 at_history=8 Creating to objid 5665 on ost lustre-OST0000... fail_val=20000 fail_loc=0x80000223 total: 17 open/close in 0.04 seconds: 407.79 ops/second Connected clients: oleg412-client.virtnet oleg412-client.virtnet service : cur 1 worst 1 (at 1713343404, 3045s ago) 1 0 1 0 phase 2 1 osc reconnect attempts on instant slow fail_loc=0x80000223 fail_loc=0 Connected clients: oleg412-client.virtnet oleg412-client.virtnet service : cur 1 worst 1 (at 1713343404, 3046s ago) 1 1 0 1 0 osc reconnect attempts on 2nd slow PASS 67b (25s) == replay-single test 68: AT: verify slowing locks ======= 05:34:12 (1713346452) at_history=8 at_history=8 fail_val=19 fail_loc=0x80000312 fail_val=25 fail_loc=0x80000312 fail_loc=0 PASS 68 (69s) Cleaning up AT ... == replay-single test 70a: check multi client t-f ======== 05:35:21 (1713346521) SKIP: replay-single test_70a Need two or more clients, have 1 SKIP 70a (0s) == replay-single test 70b: dbench 1mdts recovery; 1 clients ========================================================== 05:35:22 (1713346522) Starting client oleg412-client.virtnet: -o user_xattr,flock oleg412-server@tcp:/lustre /mnt/lustre Started clients oleg412-client.virtnet: 192.168.204.112@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt) + MISSING_DBENCH_OK= + PATH=/opt/iozone/bin:/opt/iozone/bin:/opt/iozone/bin:/home/green/git/lustre-release/lustre/tests/mpi:/home/green/git/lustre-release/lustre/tests/racer:/home/green/git/lustre-release/lustre/../lustre-iokit/sgpdd-survey:/home/green/git/lustre-release/lustre/tests:/home/green/git/lustre-release/lustre/utils/gss:/home/green/git/lustre-release/lustre/utils:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin::/home/green/git/lustre-release/lustre/utils:/home/green/git/lustre-release/lustre/tests:/sbin:/usr/sbin:/home/green/git/lustre-release/lustre/utils:/home/green/git/lustre-release/lustre/tests/: + DBENCH_LIB= + TESTSUITE=replay-single + TESTNAME=test_70b + MOUNT=/mnt/lustre ++ hostname + DIR=/mnt/lustre/d70b.replay-single/oleg412-client.virtnet + LCTL=/home/green/git/lustre-release/lustre/utils/lctl + rundbench 1 -t 300 dbench: no process found Started rundbench load pid=11946 ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3712 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 32768 3687424 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3716096 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 35840 7403520 1% /mnt/lustre test_70b fail mds1 1 times Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: looking for dbench program oleg412-client.virtnet: /usr/bin/dbench oleg412-client.virtnet: creating output directory /mnt/lustre/d70b.replay-single/oleg412-client.virtnet oleg412-client.virtnet: mkdir: created directory '/mnt/lustre/d70b.replay-single' oleg412-client.virtnet: mkdir: created directory '/mnt/lustre/d70b.replay-single/oleg412-client.virtnet' oleg412-client.virtnet: found dbench client file /usr/share/dbench/client.txt oleg412-client.virtnet: '/usr/share/dbench/client.txt' -> 'client.txt' oleg412-client.virtnet: running 'dbench 1 -t 300' on /mnt/lustre/d70b.replay-single/oleg412-client.virtnet at Wed Apr 17 05:35:22 EDT 2024 oleg412-client.virtnet: waiting for dbench pid 11971 oleg412-client.virtnet: dbench version 4.00 - Copyright Andrew Tridgell 1999-2004 oleg412-client.virtnet: oleg412-client.virtnet: Running for 300 seconds with load 'client.txt' and minimum warmup 60 secs oleg412-client.virtnet: failed to create barrier semaphore oleg412-client.virtnet: 0 of 1 processes prepared for launch 0 sec oleg412-client.virtnet: 1 of 1 processes prepared for launch 0 sec oleg412-client.virtnet: releasing clients oleg412-client.virtnet: 1 615 20.98 MB/sec warmup 1 sec latency 11.748 ms oleg412-client.virtnet: 1 1027 11.44 MB/sec warmup 2 sec latency 59.707 ms oleg412-client.virtnet: 1 1749 8.76 MB/sec warmup 3 sec latency 39.269 ms oleg412-client.virtnet: 1 2447 7.03 MB/sec warmup 4 sec latency 23.190 ms oleg412-client.virtnet: 1 3237 6.90 MB/sec warmup 5 sec latency 271.563 ms oleg412-client.virtnet: 1 3237 5.75 MB/sec warmup 6 sec latency 1271.732 ms oleg412-client.virtnet: 1 3237 4.93 MB/sec warmup 7 sec latency 2271.990 ms oleg412-client.virtnet: 1 3237 4.31 MB/sec warmup 8 sec latency 3272.140 ms oleg412-client.virtnet: 1 3237 3.83 MB/sec warmup 9 sec latency 4272.295 ms oleg412-client.virtnet: 1 3237 3.45 MB/sec warmup 10 sec latency 5272.450 ms oleg412-client.virtnet: 1 3237 3.14 MB/sec warmup 11 sec latency 6272.623 ms oleg412-client.virtnet: 1 3237 2.87 MB/sec warmup 12 sec latency 7272.815 ms oleg412-client.virtnet: 1 3237 2.65 MB/sec warmup 13 sec latency 8272.981 ms oleg412-client.virtnet: 1 3237 2.46 MB/sec warmup 14 sec latency 9273.147 ms oleg412-client.virtnet: 1 3237 2.30 MB/sec warmup 15 sec latency 10273.288 ms oleg412-client.virtnet: 1 3237 2.16 MB/sec warmup 16 sec latency 11273.462 ms oleg412-client.virtnet: 1 3237 2.03 MB/sec warmup 17 sec latency 12273.651 ms oleg412-client.virtnet: 1 3237 1.92 MB/sec warmup 18 sec latency 13273.823 ms oleg412-client.virtnet: 1 3237 1.82 MB/sec warmup 19 sec latency 14273.974 ms oleg412-client.virtnet: 1 3237 1.72 MB/sec warmup 20 sec latency 15274.122 ms oleg412-client.virtnet: 1 3237 1.64 MB/sec warmup 21 sec latency 16274.297 ms oleg412-client.virtnet: 1 3237 1.57 MB/sec warmup 22 sec latency 17274.490 ms oleg412-client.virtnet: 1 3237 1.50 MB/sec warmup 23 sec latency 18274.663 ms oleg412-client.virtnet: 1 3237 1.44 MB/sec warmup 24 sec latency 19274.816 ms oleg412-client.virtnet: 1 3237 1.38 MB/sec warmup 25 sec latency 20274.992 ms oleg412-client.virtnet: 1 3237 1.33 MB/sec warmup 26 sec latency 21275.173 ms oleg412-client.virtnet: 1 3237 1.28 MB/sec warmup 27 sec latency 22275.349 ms oleg412-client.virtnet: 1 3237 1.23 MB/sec warmup 28 sec latency 23275.535 ms oleg412-client.virtnet: 1 3237 1.19 MB/sec warmup 29 sec latency 24275.707 ms oleg412-client.virtnet: 1 3237 1.15 MB/sec warmup 30 sec latency 25275.870 ms oleg412-client.virtnet: 1 3237 1.11 MB/sec warmup 31 sec latency 26276.014 ms oleg412-client.virtnet: 1 3237 1.08 MB/sec warmup 32 soleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 40960 3719168 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 13312 3747840 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 54272 7467008 1% /mnt/lustre test_70b fail mds1 2 times Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3712 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 40960 3698688 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 13312 3741696 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 54272 7440384 1% /mnt/lustre ec latency 27276.168 ms oleg412-client.virtnet: 1 3237 1.05 MB/sec warmup 33 sec latency 28276.369 ms oleg412-client.virtnet: 1 3237 1.01 MB/sec warmup 34 sec latency 29276.518 ms oleg412-client.virtnet: 1 3237 0.99 MB/sec warmup 35 sec latency 30276.669 ms oleg412-client.virtnet: 1 3645 1.07 MB/sec warmup 36 sec latency 30845.405 ms oleg412-client.virtnet: 1 4214 1.09 MB/sec warmup 37 sec latency 9.284 ms oleg412-client.virtnet: 1 4645 1.11 MB/sec warmup 38 sec latency 36.308 ms oleg412-client.virtnet: 1 5354 1.16 MB/sec warmup 39 sec latency 32.589 ms oleg412-client.virtnet: 1 6066 1.20 MB/sec warmup 40 sec latency 30.130 ms oleg412-client.virtnet: 1 7180 1.41 MB/sec warmup 41 sec latency 25.471 ms oleg412-client.virtnet: 1 7715 1.41 MB/sec warmup 42 sec latency 10.106 ms oleg412-client.virtnet: 1 7898 1.39 MB/sec warmup 43 sec latency 552.167 ms oleg412-client.virtnet: 1 7898 1.36 MB/sec warmup 44 sec latency 1552.331 ms oleg412-client.virtnet: 1 7898 1.33 MB/sec warmup 45 sec latency 2552.506 ms oleg412-client.virtnet: 1 7898 1.30 MB/sec warmup 46 sec latency 3552.701 ms oleg412-client.virtnet: 1 7898 1.27 MB/sec warmup 47 sec latency 4552.853 ms oleg412-client.virtnet: 1 7898 1.24 MB/sec warmup 48 sec latency 5553.012 ms oleg412-client.virtnet: 1 7898 1.22 MB/sec warmup 49 sec latency 6553.140 ms oleg412-client.virtnet: 1 7898 1.19 MB/sec warmup 50 sec latency 7553.288 ms oleg412-client.virtnet: 1 7898 1.17 MB/sec warmup 51 sec latency 8553.453 ms oleg412-client.virtnet: 1 7898 1.15 MB/sec warmup 52 sec latency 9553.643 ms oleg412-client.virtnet: 1 7898 1.13 MB/sec warmup 53 sec latency 10553.816 ms oleg412-client.virtnet: 1 7898 1.11 MB/sec warmup 54 sec latency 11554.010 ms oleg412-client.virtnet: 1 7898 1.09 MB/sec warmup 55 sec latency 12554.168 ms oleg412-client.virtnet: 1 7898 1.07 MB/sec warmup 56 sec latency 13554.330 ms oleg412-client.virtnet: 1 7898 1.05 MB/sec warmup 57 sec latency 14554.551 ms oleg412-client.virtnet: 1 7898 1.03 MB/sec warmup 58 sec latency 15554.723 ms oleg412-client.virtnet: 1 7898 1.01 MB/sec warmup 59 sec latency 16554.882 ms oleg412-client.virtnet: 1 7898 0.00 MB/sec execute 1 sec latency 18555.129 ms oleg412-client.virtnet: 1 7898 0.00 MB/sec execute 2 sec latency 19555.287 ms oleg412-client.virtnet: 1 7898 0.00 MB/sec execute 3 sec latency 20555.510 ms oleg412-client.virtnet: 1 7898 0.00 MB/sec execute 4 sec latency 21555.661 ms oleg412-client.virtnet: 1 7898 0.00 MB/sec execute 5 sec latency 22555.828 ms oleg412-client.virtnet: 1 7898 0.00 MB/sec execute 6 sec latency 23555.983 ms oleg412-client.virtnet: 1 7898 0.00 MB/sec execute 7 sec latency 24556.137 ms oleg412-client.virtnet: 1 7898 0.00 MB/sec execute 8 sec latency 25556.297 ms oleg412-client.virtnet: 1 7898 0.00 MB/sec execute 9 sec latency 26556.442 ms oleg412-client.virtnet: 1 7898 0.00 MB/sec execute 10 sec latency 27556.603 ms oleg412-client.virtnet: 1 7898 0.00 MB/sec execute 11 sec latency 28556.805 ms oleg412-client.virtnet: 1 7898 0.00 MB/sec execute 12 sec latency 29556.959 ms oleg412-client.virtnet: 1 8025 0.02 MB/sec execute 13 sec latency 30240.954 ms oleg412-client.virtnet: 1 8705 0.34 MB/sec execute 14 sec latency 35.204 ms oleg412-client.virtnet: 1 9460 0.43 MB/sec execute 15 sec latency 29.074 ms oleg412-client.virtnet: 1 10404 0.87 MB/sec execute 16 sec latency 24.888 ms oleg412-client.virtnet: 1 11076 1.09 MB/sec execute 17 sec latency 15.018 ms olegtest_70b fail mds1 3 times Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3712 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 40960 3712000 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 13312 3728384 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 54272 7440384 1% /mnt/lustre test_70b fail mds1 4 times Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 412-client.virtnet: 1 11561 1.08 MB/sec execute 18 sec latency 34.422 ms oleg412-client.virtnet: 1 12262 1.25 MB/sec execute 19 sec latency 28.863 ms oleg412-client.virtnet: 1 12546 1.20 MB/sec execute 20 sec latency 504.357 ms oleg412-client.virtnet: 1 12546 1.15 MB/sec execute 21 sec latency 1504.546 ms oleg412-client.virtnet: 1 12546 1.09 MB/sec execute 22 sec latency 2504.741 ms oleg412-client.virtnet: 1 12546 1.05 MB/sec execute 23 sec latency 3504.902 ms oleg412-client.virtnet: 1 12546 1.00 MB/sec execute 24 sec latency 4505.065 ms oleg412-client.virtnet: 1 12546 0.96 MB/sec execute 25 sec latency 5505.262 ms oleg412-client.virtnet: 1 12546 0.93 MB/sec execute 26 sec latency 6505.430 ms oleg412-client.virtnet: 1 12546 0.89 MB/sec execute 27 sec latency 7505.590 ms oleg412-client.virtnet: 1 12546 0.86 MB/sec execute 28 sec latency 8505.743 ms oleg412-client.virtnet: 1 12546 0.83 MB/sec execute 29 sec latency 9505.915 ms oleg412-client.virtnet: 1 12546 0.80 MB/sec execute 30 sec latency 10506.104 ms oleg412-client.virtnet: 1 12546 0.78 MB/sec execute 31 sec latency 11506.226 ms oleg412-client.virtnet: 1 12546 0.75 MB/sec execute 32 sec latency 12506.381 ms oleg412-client.virtnet: 1 12546 0.73 MB/sec execute 33 sec latency 13506.610 ms oleg412-client.virtnet: 1 12546 0.71 MB/sec execute 34 sec latency 14506.826 ms oleg412-client.virtnet: 1 12546 0.69 MB/sec execute 35 sec latency 15507.006 ms oleg412-client.virtnet: 1 12546 0.67 MB/sec execute 36 sec latency 16507.206 ms oleg412-client.virtnet: 1 12546 0.65 MB/sec execute 37 sec latency 17507.429 ms oleg412-client.virtnet: 1 12546 0.63 MB/sec execute 38 sec latency 18507.594 ms oleg412-client.virtnet: 1 12546 0.62 MB/sec execute 39 sec latency 19507.777 ms oleg412-client.virtnet: 1 12546 0.60 MB/sec execute 40 sec latency 20507.937 ms oleg412-client.virtnet: 1 12546 0.59 MB/sec execute 41 sec latency 21508.162 ms oleg412-client.virtnet: 1 12546 0.57 MB/sec execute 42 sec latency 22508.330 ms oleg412-client.virtnet: 1 12546 0.56 MB/sec execute 43 sec latency 23508.495 ms oleg412-client.virtnet: 1 12546 0.55 MB/sec execute 44 sec latency 24508.665 ms oleg412-client.virtnet: 1 12546 0.53 MB/sec execute 45 sec latency 25508.820 ms oleg412-client.virtnet: 1 12546 0.52 MB/sec execute 46 sec latency 26509.015 ms oleg412-client.virtnet: 1 12546 0.51 MB/sec execute 47 sec latency 27509.209 ms oleg412-client.virtnet: 1 12546 0.50 MB/sec execute 48 sec latency 28509.349 ms oleg412-client.virtnet: 1 12546 0.49 MB/sec execute 49 sec latency 29509.510 ms oleg412-client.virtnet: 1 12854 0.49 MB/sec execute 50 sec latency 30117.936 ms oleg412-client.virtnet: 1 13660 0.61 MB/sec execute 51 sec latency 30.101 ms oleg412-client.virtnet: 1 14484 0.72 MB/sec execute 52 sec latency 27.048 ms oleg412-client.virtnet: 1 14940 0.72 MB/sec execute 53 sec latency 38.506 ms oleg412-client.virtnet: 1 15455 0.74 MB/sec execute 54 sec latency 29.582 ms oleg412-client.virtnet: 1 16155 0.79 MB/sec execute 55 sec latency 37.152 ms oleg412-client.virtnet: 1 16969 0.87 MB/sec execute 56 sec latency 25.067 ms oleg412-client.virtnet: 1 17521 0.92 MB/sec execute 57 sec latency 532.921 ms oleg412-client.virtnet: 1 17521 0.90 MB/sec execute 58 sec latency 1533.074 ms oleg412-client.virtnet: 1 17521 0.88 MB/sec execute 59 sec latency 2533.284 ms oleg412-client.virtnet: 1 17521 0.87 MB/sec execute 60 sec latency 3533.453 ms oleg412-client.virtnet: 1 17521 0.86 MB/sec execute 61 sec latency 4533.602 ms oleg4Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 40960 3720192 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 14336 3745792 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 55296 7465984 1% /mnt/lustre test_70b fail mds1 5 times Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 12-client.virtnet: 1 17521 0.84 MB/sec execute 62 sec latency 5533.768 ms oleg412-client.virtnet: 1 17521 0.83 MB/sec execute 63 sec latency 6533.920 ms oleg412-client.virtnet: 1 17521 0.82 MB/sec execute 64 sec latency 7534.087 ms oleg412-client.virtnet: 1 17521 0.80 MB/sec execute 65 sec latency 8534.283 ms oleg412-client.virtnet: 1 17521 0.79 MB/sec execute 66 sec latency 9534.454 ms oleg412-client.virtnet: 1 17521 0.78 MB/sec execute 67 sec latency 10534.611 ms oleg412-client.virtnet: 1 17521 0.77 MB/sec execute 68 sec latency 11534.775 ms oleg412-client.virtnet: 1 17521 0.76 MB/sec execute 69 sec latency 12534.931 ms oleg412-client.virtnet: 1 17521 0.75 MB/sec execute 70 sec latency 13535.116 ms oleg412-client.virtnet: 1 17521 0.73 MB/sec execute 71 sec latency 14535.273 ms oleg412-client.virtnet: 1 17521 0.72 MB/sec execute 72 sec latency 15535.440 ms oleg412-client.virtnet: 1 17521 0.71 MB/sec execute 73 sec latency 16535.582 ms oleg412-client.virtnet: 1 17521 0.70 MB/sec execute 74 sec latency 17535.744 ms oleg412-client.virtnet: 1 17521 0.70 MB/sec execute 75 sec latency 18535.917 ms oleg412-client.virtnet: 1 17521 0.69 MB/sec execute 76 sec latency 19536.099 ms oleg412-client.virtnet: 1 17521 0.68 MB/sec execute 77 sec latency 20536.237 ms oleg412-client.virtnet: 1 17521 0.67 MB/sec execute 78 sec latency 21536.431 ms oleg412-client.virtnet: 1 17521 0.66 MB/sec execute 79 sec latency 22536.599 ms oleg412-client.virtnet: 1 17521 0.65 MB/sec execute 80 sec latency 23536.774 ms oleg412-client.virtnet: 1 17521 0.64 MB/sec execute 81 sec latency 24536.953 ms oleg412-client.virtnet: 1 17521 0.64 MB/sec execute 82 sec latency 25537.120 ms oleg412-client.virtnet: 1 17521 0.63 MB/sec execute 83 sec latency 26537.334 ms oleg412-client.virtnet: 1 17521 0.62 MB/sec execute 84 sec latency 27537.480 ms oleg412-client.virtnet: 1 17521 0.61 MB/sec execute 85 sec latency 28537.627 ms oleg412-client.virtnet: 1 17521 0.61 MB/sec execute 86 sec latency 29537.826 ms oleg412-client.virtnet: 1 17825 0.64 MB/sec execute 87 sec latency 30172.223 ms oleg412-client.virtnet: 1 18397 0.65 MB/sec execute 88 sec latency 8.816 ms oleg412-client.virtnet: 1 18836 0.66 MB/sec execute 89 sec latency 36.852 ms oleg412-client.virtnet: 1 19508 0.69 MB/sec execute 90 sec latency 33.834 ms oleg412-client.virtnet: 1 20202 0.70 MB/sec execute 91 sec latency 29.727 ms oleg412-client.virtnet: 1 21158 0.78 MB/sec execute 92 sec latency 32.693 ms oleg412-client.virtnet: 1 21764 0.82 MB/sec execute 93 sec latency 11.162 ms oleg412-client.virtnet: 1 22035 0.81 MB/sec execute 94 sec latency 472.149 ms oleg412-client.virtnet: 1 22035 0.80 MB/sec execute 95 sec latency 1472.319 ms oleg412-client.virtnet: 1 22035 0.79 MB/sec execute 96 sec latency 2472.479 ms oleg412-client.virtnet: 1 22035 0.79 MB/sec execute 97 sec latency 3472.644 ms oleg412-client.virtnet: 1 22035 0.78 MB/sec execute 98 sec latency 4472.829 ms oleg412-client.virtnet: 1 22035 0.77 MB/sec execute 99 sec latency 5472.993 ms oleg412-client.virtnet: 1 22035 0.76 MB/sec execute 100 sec latency 6473.157 ms oleg412-client.virtnet: 1 22035 0.76 MB/sec execute 101 sec latency 7473.314 ms oleg412-client.virtnet: 1 22035 0.75 MB/sec execute 102 sec latency 8473.452 ms oleg412-client.virtnet: 1 22035 0.74 MB/sec execute 103 sec latency 9473.619 ms oleg412-client.virtnet: 1 22035 0.73 MB/sec execute 104 sec latency 10473.798 ms oleg412-client.virtnet: 1 22035 0.73 MB/sec execute 105 sec latency 11473.927 ms oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3712 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 40960 3696640 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 13312 3736576 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 54272 7433216 1% /mnt/lustre test_70b fail mds1 6 times Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3712 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 40960 3702784 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 13312 3742720 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 54272 7445504 1% /mnt/lustre test_70b fail mds1 7 times Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-client.virtnet: 1 22035 0.72 MB/sec execute 106 sec latency 12474.065 ms oleg412-client.virtnet: 1 22035 0.71 MB/sec execute 107 sec latency 13474.233 ms oleg412-client.virtnet: 1 22125 0.71 MB/sec execute 108 sec latency 14242.155 ms oleg412-client.virtnet: 1 22761 0.74 MB/sec execute 109 sec latency 23.305 ms oleg412-client.virtnet: 1 23392 0.74 MB/sec execute 110 sec latency 30.231 ms oleg412-client.virtnet: 1 24292 0.80 MB/sec execute 111 sec latency 27.446 ms oleg412-client.virtnet: 1 25148 0.85 MB/sec execute 112 sec latency 22.246 ms oleg412-client.virtnet: 1 25624 0.85 MB/sec execute 113 sec latency 37.866 ms oleg412-client.virtnet: 1 26275 0.88 MB/sec execute 114 sec latency 27.516 ms oleg412-client.virtnet: 1 26876 0.88 MB/sec execute 115 sec latency 67.585 ms oleg412-client.virtnet: 1 26876 0.87 MB/sec execute 116 sec latency 1067.747 ms oleg412-client.virtnet: 1 26876 0.86 MB/sec execute 117 sec latency 2067.947 ms oleg412-client.virtnet: 1 26876 0.85 MB/sec execute 118 sec latency 3068.102 ms oleg412-client.virtnet: 1 26876 0.85 MB/sec execute 119 sec latency 4068.253 ms oleg412-client.virtnet: 1 26876 0.84 MB/sec execute 120 sec latency 5068.421 ms oleg412-client.virtnet: 1 26876 0.83 MB/sec execute 121 sec latency 6068.594 ms oleg412-client.virtnet: 1 26876 0.83 MB/sec execute 122 sec latency 7068.775 ms oleg412-client.virtnet: 1 26876 0.82 MB/sec execute 123 sec latency 8068.939 ms oleg412-client.virtnet: 1 26876 0.81 MB/sec execute 124 sec latency 9069.090 ms oleg412-client.virtnet: 1 26876 0.81 MB/sec execute 125 sec latency 10069.287 ms oleg412-client.virtnet: 1 26876 0.80 MB/sec execute 126 sec latency 11069.457 ms oleg412-client.virtnet: 1 26876 0.79 MB/sec execute 127 sec latency 12069.564 ms oleg412-client.virtnet: 1 26876 0.79 MB/sec execute 128 sec latency 13069.728 ms oleg412-client.virtnet: 1 26876 0.78 MB/sec execute 129 sec latency 14069.907 ms oleg412-client.virtnet: 1 26876 0.77 MB/sec execute 130 sec latency 15070.053 ms oleg412-client.virtnet: 1 26893 0.77 MB/sec execute 131 sec latency 16024.268 ms oleg412-client.virtnet: 1 27770 0.81 MB/sec execute 132 sec latency 30.348 ms oleg412-client.virtnet: 1 28584 0.85 MB/sec execute 133 sec latency 32.512 ms oleg412-client.virtnet: 1 29100 0.85 MB/sec execute 134 sec latency 38.158 ms oleg412-client.virtnet: 1 29578 0.86 MB/sec execute 135 sec latency 30.592 ms oleg412-client.virtnet: 1 30257 0.88 MB/sec execute 136 sec latency 32.138 ms oleg412-client.virtnet: 1 30983 0.89 MB/sec execute 137 sec latency 26.049 ms oleg412-client.virtnet: 1 31785 0.93 MB/sec execute 138 sec latency 253.109 ms oleg412-client.virtnet: 1 31785 0.93 MB/sec execute 139 sec latency 1253.254 ms oleg412-client.virtnet: 1 31785 0.92 MB/sec execute 140 sec latency 2253.440 ms oleg412-client.virtnet: 1 31785 0.91 MB/sec execute 141 sec latency 3253.613 ms oleg412-client.virtnet: 1 31785 0.91 MB/sec execute 142 sec latency 4253.761 ms oleg412-client.virtnet: 1 31785 0.90 MB/sec execute 143 sec latency 5253.936 ms oleg412-client.virtnet: 1 31785 0.89 MB/sec execute 144 sec latency 6254.099 ms oleg412-client.virtnet: 1 31785 0.89 MB/sec execute 145 sec latency 7254.295 ms oleg412-client.virtnet: 1 31785 0.88 MB/sec execute 146 sec latency 8254.470 ms oleg412-client.virtnet: 1 31785 0.88 MB/sec execute 147 sec latency 9254.590 ms oleg412-client.virtnet: 1 31785 0.87 MB/sec execute 148 sec latency 10254.744 ms oleg412-client.virtnet: 1 31785 0.86 MB/sec execute 149 sec latency 11254.919 ms oleg412-client.virtoleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3712 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 40960 3705856 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 13312 3737600 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 54272 7443456 1% /mnt/lustre test_70b fail mds1 8 times Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec net: 1 31785 0.86 MB/sec execute 150 sec latency 12255.098 ms oleg412-client.virtnet: 1 31785 0.85 MB/sec execute 151 sec latency 13255.253 ms oleg412-client.virtnet: 1 31785 0.85 MB/sec execute 152 sec latency 14255.437 ms oleg412-client.virtnet: 1 31785 0.84 MB/sec execute 153 sec latency 15255.613 ms oleg412-client.virtnet: 1 31785 0.84 MB/sec execute 154 sec latency 16255.721 ms oleg412-client.virtnet: 1 32345 0.86 MB/sec execute 155 sec latency 16419.205 ms oleg412-client.virtnet: 1 32775 0.86 MB/sec execute 156 sec latency 44.269 ms oleg412-client.virtnet: 1 33435 0.88 MB/sec execute 157 sec latency 33.873 ms oleg412-client.virtnet: 1 34070 0.88 MB/sec execute 158 sec latency 37.667 ms oleg412-client.virtnet: 1 34944 0.92 MB/sec execute 159 sec latency 30.181 ms oleg412-client.virtnet: 1 35792 0.95 MB/sec execute 160 sec latency 24.317 ms oleg412-client.virtnet: 1 36250 0.95 MB/sec execute 161 sec latency 40.509 ms oleg412-client.virtnet: 1 36774 0.95 MB/sec execute 162 sec latency 30.998 ms oleg412-client.virtnet: 1 37200 0.97 MB/sec execute 163 sec latency 414.486 ms oleg412-client.virtnet: 1 37200 0.96 MB/sec execute 164 sec latency 1414.650 ms oleg412-client.virtnet: 1 37200 0.96 MB/sec execute 165 sec latency 2414.827 ms oleg412-client.virtnet: 1 37200 0.95 MB/sec execute 166 sec latency 3415.010 ms oleg412-client.virtnet: 1 37200 0.94 MB/sec execute 167 sec latency 4415.165 ms oleg412-client.virtnet: 1 37200 0.94 MB/sec execute 168 sec latency 5415.312 ms oleg412-client.virtnet: 1 37200 0.93 MB/sec execute 169 sec latency 6415.465 ms oleg412-client.virtnet: 1 37200 0.93 MB/sec execute 170 sec latency 7415.615 ms oleg412-client.virtnet: 1 37200 0.92 MB/sec execute 171 sec latency 8415.776 ms oleg412-client.virtnet: 1 37200 0.92 MB/sec execute 172 sec latency 9415.995 ms oleg412-client.virtnet: 1 37200 0.91 MB/sec execute 173 sec latency 10416.169 ms oleg412-client.virtnet: 1 37200 0.91 MB/sec execute 174 sec latency 11416.271 ms oleg412-client.virtnet: 1 37200 0.90 MB/sec execute 175 sec latency 12416.400 ms oleg412-client.virtnet: 1 37200 0.90 MB/sec execute 176 sec latency 13416.572 ms oleg412-client.virtnet: 1 37200 0.89 MB/sec execute 177 sec latency 14416.743 ms oleg412-client.virtnet: 1 37200 0.89 MB/sec execute 178 sec latency 15416.950 ms oleg412-client.virtnet: 1 37200 0.88 MB/sec execute 179 sec latency 16417.120 ms oleg412-client.virtnet: 1 37200 0.88 MB/sec execute 180 sec latency 17417.278 ms oleg412-client.virtnet: 1 37200 0.87 MB/sec execute 181 sec latency 18417.439 ms oleg412-client.virtnet: 1 37200 0.87 MB/sec execute 182 sec latency 19417.623 ms oleg412-client.virtnet: 1 37200 0.86 MB/sec execute 183 sec latency 20417.751 ms oleg412-client.virtnet: 1 37200 0.86 MB/sec execute 184 sec latency 21417.907 ms oleg412-client.virtnet: 1 37200 0.85 MB/sec execute 185 sec latency 22418.040 ms oleg412-client.virtnet: 1 37200 0.85 MB/sec execute 186 sec latency 23418.207 ms oleg412-client.virtnet: 1 37200 0.84 MB/sec execute 187 sec latency 24418.383 ms oleg412-client.virtnet: 1 37200 0.84 MB/sec execute 188 sec latency 25418.561 ms oleg412-client.virtnet: 1 37200 0.83 MB/sec execute 189 sec latency 26418.713 ms oleg412-client.virtnet: 1 37200 0.83 MB/sec execute 190 sec latency 27418.832 ms oleg412-client.virtnet: 1 37200 0.83 MB/sec execute 191 sec latency 28419.035 ms oleg412-client.virtnet: 1 37807 0.83 MB/sec execute 192 sec latency 28595.624 ms oleg412-client.virtnet: 1 38680 0.86 MB/sec execute 193 sec latency 28.841 ms oleg412-clientUUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3712 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 40960 3715072 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 13312 3724288 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 54272 7439360 1% /mnt/lustre test_70b fail mds1 9 times Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3712 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 40960 3714048 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 14336 3734528 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 55296 7448576 1% /mnt/lustre test_70b fail mds1 10 times Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 .virtnet: 1 39381 0.88 MB/sec execute 194 sec latency 28.695 ms oleg412-client.virtnet: 1 39858 0.88 MB/sec execute 195 sec latency 34.685 ms oleg412-client.virtnet: 1 40522 0.90 MB/sec execute 196 sec latency 27.004 ms oleg412-client.virtnet: 1 41229 0.90 MB/sec execute 197 sec latency 31.189 ms oleg412-client.virtnet: 1 42066 0.93 MB/sec execute 198 sec latency 34.255 ms oleg412-client.virtnet: 1 42066 0.93 MB/sec execute 199 sec latency 1034.456 ms oleg412-client.virtnet: 1 42066 0.92 MB/sec execute 200 sec latency 2034.630 ms oleg412-client.virtnet: 1 42066 0.92 MB/sec execute 201 sec latency 3034.760 ms oleg412-client.virtnet: 1 42066 0.91 MB/sec execute 202 sec latency 4034.916 ms oleg412-client.virtnet: 1 42066 0.91 MB/sec execute 203 sec latency 5035.081 ms oleg412-client.virtnet: 1 42066 0.90 MB/sec execute 204 sec latency 6035.229 ms oleg412-client.virtnet: 1 42066 0.90 MB/sec execute 205 sec latency 7035.377 ms oleg412-client.virtnet: 1 42066 0.89 MB/sec execute 206 sec latency 8035.492 ms oleg412-client.virtnet: 1 42066 0.89 MB/sec execute 207 sec latency 9035.628 ms oleg412-client.virtnet: 1 42066 0.89 MB/sec execute 208 sec latency 10035.752 ms oleg412-client.virtnet: 1 42066 0.88 MB/sec execute 209 sec latency 11035.901 ms oleg412-client.virtnet: 1 42066 0.88 MB/sec execute 210 sec latency 12036.034 ms oleg412-client.virtnet: 1 42066 0.87 MB/sec execute 211 sec latency 13036.190 ms oleg412-client.virtnet: 1 42066 0.87 MB/sec execute 212 sec latency 14036.372 ms oleg412-client.virtnet: 1 42066 0.86 MB/sec execute 213 sec latency 15036.489 ms oleg412-client.virtnet: 1 42066 0.86 MB/sec execute 214 sec latency 16036.586 ms oleg412-client.virtnet: 1 42066 0.86 MB/sec execute 215 sec latency 17036.724 ms oleg412-client.virtnet: 1 42066 0.85 MB/sec execute 216 sec latency 18036.870 ms oleg412-client.virtnet: 1 42066 0.85 MB/sec execute 217 sec latency 19037.002 ms oleg412-client.virtnet: 1 42066 0.84 MB/sec execute 218 sec latency 20037.152 ms oleg412-client.virtnet: 1 42066 0.84 MB/sec execute 219 sec latency 21037.324 ms oleg412-client.virtnet: 1 42066 0.84 MB/sec execute 220 sec latency 22037.453 ms oleg412-client.virtnet: 1 42066 0.83 MB/sec execute 221 sec latency 23037.612 ms oleg412-client.virtnet: 1 42066 0.83 MB/sec execute 222 sec latency 24037.763 ms oleg412-client.virtnet: 1 42066 0.83 MB/sec execute 223 sec latency 25037.913 ms oleg412-client.virtnet: 1 42066 0.82 MB/sec execute 224 sec latency 26038.066 ms oleg412-client.virtnet: 1 42066 0.82 MB/sec execute 225 sec latency 27038.233 ms oleg412-client.virtnet: 1 42066 0.82 MB/sec execute 226 sec latency 28038.432 ms oleg412-client.virtnet: 1 42066 0.81 MB/sec execute 227 sec latency 29038.607 ms oleg412-client.virtnet: 1 42066 0.81 MB/sec execute 228 sec latency 30038.723 ms oleg412-client.virtnet: 1 42817 0.83 MB/sec execute 229 sec latency 30259.241 ms oleg412-client.virtnet: 1 43286 0.83 MB/sec execute 230 sec latency 41.456 ms oleg412-client.virtnet: 1 43797 0.83 MB/sec execute 231 sec latency 29.083 ms oleg412-client.virtnet: 1 44452 0.85 MB/sec execute 232 sec latency 31.828 ms oleg412-client.virtnet: 1 45160 0.85 MB/sec execute 233 sec latency 36.042 ms oleg412-client.virtnet: 1 46210 0.89 MB/sec execute 234 sec latency 23.750 ms oleg412-client.virtnet: 1 46767 0.89 MB/sec execute 235 sec latency 12.364 ms oleg412-client.virtnet: 1 46824 0.89 MB/sec execute 236 sec latency 933.580 ms oleg412-client.virtnet: 1 46824 0.89 MB/sec execute 237 sec latency 1933.741 ms oleg412-client.virtFailover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec net: 1 46824 0.88 MB/sec execute 238 sec latency 2933.912 ms oleg412-client.virtnet: 1 46824 0.88 MB/sec execute 239 sec latency 3934.059 ms oleg412-client.virtnet: 1 46824 0.88 MB/sec execute 240 sec latency 4934.216 ms oleg412-client.virtnet: 1 46824 0.87 MB/sec execute 241 sec latency 5934.378 ms oleg412-client.virtnet: 1 46824 0.87 MB/sec execute 242 sec latency 6934.532 ms oleg412-client.virtnet: 1 46824 0.86 MB/sec execute 243 sec latency 7934.690 ms oleg412-client.virtnet: 1 46824 0.86 MB/sec execute 244 sec latency 8934.873 ms oleg412-client.virtnet: 1 46824 0.86 MB/sec execute 245 sec latency 9935.039 ms oleg412-client.virtnet: 1 46824 0.85 MB/sec execute 246 sec latency 10935.218 ms oleg412-client.virtnet: 1 46824 0.85 MB/sec execute 247 sec latency 11935.401 ms oleg412-client.virtnet: 1 46824 0.85 MB/sec execute 248 sec latency 12935.561 ms oleg412-client.virtnet: 1 46824 0.84 MB/sec execute 249 sec latency 13935.747 ms oleg412-client.virtnet: 1 46824 0.84 MB/sec execute 250 sec latency 14935.901 ms oleg412-client.virtnet: 1 46824 0.84 MB/sec execute 251 sec latency 15936.068 ms oleg412-client.virtnet: 1 46824 0.83 MB/sec execute 252 sec latency 16936.279 ms oleg412-client.virtnet: 1 46824 0.83 MB/sec execute 253 sec latency 17936.445 ms oleg412-client.virtnet: 1 46824 0.83 MB/sec execute 254 sec latency 18936.625 ms oleg412-client.virtnet: 1 46824 0.82 MB/sec execute 255 sec latency 19936.810 ms oleg412-client.virtnet: 1 46824 0.82 MB/sec execute 256 sec latency 20936.995 ms oleg412-client.virtnet: 1 46824 0.82 MB/sec execute 257 sec latency 21937.174 ms oleg412-client.virtnet: 1 46824 0.81 MB/sec execute 258 sec latency 22937.324 ms oleg412-client.virtnet: 1 46824 0.81 MB/sec execute 259 sec latency 23937.514 ms oleg412-client.virtnet: 1 46824 0.81 MB/sec execute 260 sec latency 24937.676 ms oleg412-client.virtnet: 1 46824 0.80 MB/sec execute 261 sec latency 25937.837 ms oleg412-client.virtnet: 1 46824 0.80 MB/sec execute 262 sec latency 26938.003 ms oleg412-client.virtnet: 1 46824 0.80 MB/sec execute 263 sec latency 27938.173 ms oleg412-client.virtnet: 1 46824 0.80 MB/sec execute 264 sec latency 28938.340 ms oleg412-client.virtnet: 1 46824 0.79 MB/sec execute 265 sec latency 29938.560 ms oleg412-client.virtnet: 1 47074 0.79 MB/sec execute 266 sec latency 30262.706 ms oleg412-client.virtnet: 1 47733 0.81 MB/sec execute 267 sec latency 29.250 ms oleg412-client.virtnet: 1 48395 0.80 MB/sec execute 268 sec latency 25.973 ms oleg412-client.virtnet: 1 49161 0.83 MB/sec execute 269 sec latency 27.650 ms oleg412-client.virtnet: 1 49999 0.85 MB/sec execute 270 sec latency 21.557 ms oleg412-client.virtnet: 1 50489 0.85 MB/sec execute 271 sec latency 42.002 ms oleg412-client.virtnet: 1 51162 0.86 MB/sec execute 272 sec latency 26.706 ms oleg412-client.virtnet: 1 51864 0.86 MB/sec execute 273 sec latency 28.789 ms oleg412-client.virtnet: 1 52712 0.88 MB/sec execute 274 sec latency 24.720 ms oleg412-client.virtnet: 1 53610 0.90 MB/sec execute 275 sec latency 21.020 ms oleg412-client.virtnet: 1 54120 0.90 MB/sec execute 276 sec latency 34.207 ms oleg412-client.virtnet: 1 54846 0.91 MB/sec execute 277 sec latency 21.166 ms oleg412-client.virtnet: 1 55614 0.92 MB/sec execute 278 sec latency 25.516 ms oleg412-client.virtnet: 1 56596 0.94 MB/sec execute 279 sec latency 30.098 ms oleg412-client.virtnet: 1 57258 0.95 MB/sec execute 280 sec latency 9.780 ms oleg412-client.virtnet: 1 57759 0.95 MB/sec execute 281 sec latency 34.495 ms oleg412-client.virtnet: 1 58466 0.97 MB/sec execute 282 sec latency 26.419 ms oleg412-client.virtnet: 1 59177 0.97 MB/sec execute 283 sec latency 27.713 ms oleg412-client.virtnet: 1 60142 0.99 MB/sec execute 284 sec latency 27.790 ms oleg412-client.virtnet: 1 60794 1.00 MB/sec execute 285 sec latency 9.230 ms oleg412-client.virtnet: 1 61307 1.00 MB/sec execute 286 sec latency 36.790 ms oleg412-client.virtnet: 1 62052 1.02 MB/sec execute 287 sec latency 21.190 ms oleg412-client.virtnet: 1 62784 1.02 MB/sec execute 288 sec latency 29.842 ms oleg412-client.virtnet: 1 63896 1.05 MB/sec execute 289 sec latency 24.646 ms oleg412-client.virtnet: 1 64490 1.05 MB/sec execute 290 sec latency 9.364 ms oleg412-client.virtnet: 1 64980 1.06 MB/sec execute 291 sec latency 45.022 ms oleg412-client.virtnet: 1 65689 1.06 MB/sec execute 292 sec latency 28.952 ms oleg412-client.virtnet: 1 66431 1.07 MB/sec execute 293 sec latency 31.105 ms oleg412-client.virtnet: 1 67490 1.10 MB/sec execute 294 sec latency 26.162 ms oleg412-client.virtnet: 1 68023 1.10 MB/sec execute 295 sec latency 9.732 ms oleg412-client.virtnet: 1 68476 1.10 MB/sec execute 296 sec latency 36.152 ms oleg412-client.virtnet: 1 69199 1.11 MB/sec execute 297 sec latency 24.240 ms oleg412-client.virtnet: 1 69898 1.12 MB/sec execute 298 sec latency 28.863 ms oleg412-client.virtnet: 1 70979 1.14 MB/sec execute 299 sec latency 24.849 ms oleg412-client.virtnet: 1 cleanup 300 sec oleg412-client.virtnet: 0 cleanup 300 sec oleg412-client.virtnet: oleg412-client.virtnet: Operation Count AvgLat MaxLat oleg412-client.virtnet: ---------------------------------------- oleg412-client.virtnet: NTCreateX 11018 5.443 14242.143 oleg412-client.virtnet: Close 8089 6.597 30240.933 oleg412-client.virtnet: Rename 466 131.997 30259.225 oleg412-client.virtnet: Unlink 2221 1.623 3.298 oleg412-client.virtnet: Qpathinfo 9995 2.678 16024.257 oleg412-client.virtnet: Qfileinfo 1741 0.107 1.688 oleg412-client.virtnet: Qfsinfo 1823 0.080 3.192 oleg412-client.virtnet: Sfileinfo 893 2.640 13.343 oleg412-client.virtnet: Find 3857 0.339 12.581 oleg412-client.virtnet: WriteX 5432 0.635 4.603 oleg412-client.virtnet: ReadX 17290 0.011 1.587 oleg412-client.virtnet: LockX 36 0.780 0.918 oleg412-client.virtnet: UnlockX 36 0.888 2.502 oleg412-client.virtnet: Flush 762 136.841 30262.696 oleg412-client.virtnet: oleg412-client.virtnet: Throughput 1.14397 MB/sec 1 clients 1 procs max_latency=30262.706 ms oleg412-client.virtnet: stopping dbench on /mnt/lustre/d70b.replay-single/oleg412-client.virtnet at Wed Apr 17 05:41:23 EDT 2024 with return code 0 oleg412-client.virtnet: clean dbench files on /mnt/lustre/d70b.replay-single/oleg412-client.virtnet oleg412-client.virtnet: /mnt/lustre/d70b.replay-single/oleg412-client.virtnet /mnt/lustre/d70b.replay-single/oleg412-client.virtnet oleg412-client.virtnet: removed directory: 'clients/client0/~dmtmp/WORD' oleg412-client.virtnet: removed directory: 'clients/client0/~dmtmp/WORDPRO' oleg412-client.virtnet: removed directory: 'clients/client0/~dmtmp/EXCEL' oleg412-client.virtnet: removed directory: 'clients/client0/~dmtmp/SEED' oleg412-client.virtnet: removed directory: 'clients/client0/~dmtmp/PWRPNT' oleg412-client.virtnet: removed directory: 'clients/client0/~dmtmp/PARADOX' oleg412-client.virtnet: removed directory: 'clients/client0/~dmtmp/PM' oleg412-client.virtnet: removed directory: 'clients/client0/~dmtmp/ACCESS' oleg412-client.virtnet: removed directory: 'clients/client0/~dmtmp/COREL' oleg412-client.virtnet: removed directory: 'clients/client0/~dmtmp' oleg412-client.virtnet: removed directory: 'clients/client0' oleg412-client.virtnet: removed directory: 'clients' oleg412-client.virtnet: removed 'client.txt' oleg412-client.virtnet: /mnt/lustre/d70b.replay-single/oleg412-client.virtnet oleg412-client.virtnet: dbench successfully finished PASS 70b (363s) == replay-single test 70c: tar 1mdts recovery ============ 05:41:24 (1713346884) Starting client oleg412-client.virtnet: -o user_xattr,flock oleg412-server@tcp:/lustre /mnt/lustre Started clients oleg412-client.virtnet: 192.168.204.112@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt) Started tar 24090 tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 12288 2196352 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 37888 3003392 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 13312 3743744 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 51200 6747136 1% /mnt/lustre test_70c fail mds1 1 times Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 12288 2196352 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 32768 3137536 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 11264 3154944 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 44032 6292480 1% /mnt/lustre test_70c fail mds1 2 times Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 70c (354s) == replay-single test 70d: mkdir/rmdir striped dir 1mdts recovery ========================================================== 05:47:18 (1713347238) SKIP: replay-single test_70d needs >= 2 MDTs SKIP 70d (0s) == replay-single test 70e: rename cross-MDT with random fails ========================================================== 05:47:18 (1713347238) SKIP: replay-single test_70e needs >= 2 MDTs SKIP 70e (1s) == replay-single test 70f: OSS O_DIRECT recovery with 1 clients ========================================================== 05:47:19 (1713347239) mount clients oleg412-client.virtnet ... Starting client oleg412-client.virtnet: -o user_xattr,flock oleg412-server@tcp:/lustre /mnt/lustre Started clients oleg412-client.virtnet: 192.168.204.112@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt) Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3712 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 8192 3721216 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 8192 3721216 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 16384 7442432 1% /mnt/lustre ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... test_70f failing OST 1 times Failing ost1 on oleg412-server Stopping /mnt/lustre-ost1 (opts:) on oleg412-server reboot facets: ost1 Failover ost1 to oleg412-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-OST0000 ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 8192 3715072 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 8192 3715072 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 16384 7430144 1% /mnt/lustre ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... test_70f failing OST 2 times Failing ost1 on oleg412-server Stopping /mnt/lustre-ost1 (opts:) on oleg412-server reboot facets: ost1 Failover ost1 to oleg412-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-OST0000 ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 8192 3731456 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 8192 3731456 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 16384 7462912 1% /mnt/lustre Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... test_70f failing OST 3 times Failing ost1 on oleg412-server Stopping /mnt/lustre-ost1 (opts:) on oleg412-server reboot facets: ost1 Failover ost1 to oleg412-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-OST0000 ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 6144 3750912 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 5120 3750912 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 11264 7501824 1% /mnt/lustre ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... test_70f failing OST 4 times Failing ost1 on oleg412-server Stopping /mnt/lustre-ost1 (opts:) on oleg412-server reboot facets: ost1 Failover ost1 to oleg412-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-OST0000 ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3712 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 7168 3741696 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 6144 3741696 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 13312 7483392 1% /mnt/lustre ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... test_70f failing OST 5 times Failing ost1 on oleg412-server Stopping /mnt/lustre-ost1 (opts:) on oleg412-server reboot facets: ost1 Failover ost1 to oleg412-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-OST0000 ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... ldlm.namespaces.MGC192.168.204.112@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800aaf30800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800aaf30800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg412-client.virtnet' ... PASS 70f (128s) == replay-single test 71a: mkdir/rmdir striped dir with 2 mdts recovery ========================================================== 05:49:27 (1713347367) SKIP: replay-single test_71a needs >= 2 MDTs SKIP 71a (1s) == replay-single test 73a: open(O_CREAT), unlink, replay, reconnect before open replay, close ========================================================== 05:49:28 (1713347368) multiop /mnt/lustre/f73a.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6842 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3712 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 8192 3751936 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 8192 3751936 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 16384 7503872 1% /mnt/lustre fail_loc=0x80000302 Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 73a (40s) == replay-single test 73b: open(O_CREAT), unlink, replay, reconnect at open_replay reply, close ========================================================== 05:50:08 (1713347408) multiop /mnt/lustre/f73b.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.6842 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210560 3584 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 6144 7532544 1% /mnt/lustre fail_loc=0x80000157 Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 73b (38s) == replay-single test 74: Ensure applications don't fail waiting for OST recovery ========================================================== 05:50:46 (1713347446) Stopping clients: oleg412-client.virtnet /mnt/lustre (opts:) Stopping client oleg412-client.virtnet /mnt/lustre opts: Stopping /mnt/lustre-ost1 (opts:) on oleg412-server Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 Starting client oleg412-client.virtnet: -o user_xattr,flock oleg412-server@tcp:/lustre /mnt/lustre Started clients oleg412-client.virtnet: 192.168.204.112@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt) Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-OST0000 PASS 74 (27s) == replay-single test 80a: DNE: create remote dir, drop update rep from MDT0, fail MDT0 ========================================================== 05:51:13 (1713347473) SKIP: replay-single test_80a needs >= 2 MDTs SKIP 80a (1s) == replay-single test 80b: DNE: create remote dir, drop update rep from MDT0, fail MDT1 ========================================================== 05:51:14 (1713347474) SKIP: replay-single test_80b needs >= 2 MDTs SKIP 80b (0s) == replay-single test 80c: DNE: create remote dir, drop update rep from MDT1, fail MDT[0,1] ========================================================== 05:51:14 (1713347474) SKIP: replay-single test_80c needs >= 2 MDTs SKIP 80c (1s) == replay-single test 80d: DNE: create remote dir, drop update rep from MDT1, fail 2 MDTs ========================================================== 05:51:15 (1713347475) SKIP: replay-single test_80d needs >= 2 MDTs SKIP 80d (1s) == replay-single test 80e: DNE: create remote dir, drop MDT1 rep, fail MDT0 ========================================================== 05:51:16 (1713347476) SKIP: replay-single test_80e needs >= 2 MDTs SKIP 80e (1s) == replay-single test 80f: DNE: create remote dir, drop MDT1 rep, fail MDT1 ========================================================== 05:51:17 (1713347477) SKIP: replay-single test_80f needs >= 2 MDTs SKIP 80f (0s) == replay-single test 80g: DNE: create remote dir, drop MDT1 rep, fail MDT0, then MDT1 ========================================================== 05:51:17 (1713347477) SKIP: replay-single test_80g needs >= 2 MDTs SKIP 80g (1s) == replay-single test 80h: DNE: create remote dir, drop MDT1 rep, fail 2 MDTs ========================================================== 05:51:18 (1713347478) SKIP: replay-single test_80h needs >= 2 MDTs SKIP 80h (1s) == replay-single test 81a: DNE: unlink remote dir, drop MDT0 update rep, fail MDT1 ========================================================== 05:51:19 (1713347479) SKIP: replay-single test_81a needs >= 2 MDTs SKIP 81a (1s) == replay-single test 81b: DNE: unlink remote dir, drop MDT0 update reply, fail MDT0 ========================================================== 05:51:20 (1713347480) SKIP: replay-single test_81b needs >= 2 MDTs SKIP 81b (1s) == replay-single test 81c: DNE: unlink remote dir, drop MDT0 update reply, fail MDT0,MDT1 ========================================================== 05:51:21 (1713347481) SKIP: replay-single test_81c needs >= 2 MDTs SKIP 81c (1s) == replay-single test 81d: DNE: unlink remote dir, drop MDT0 update reply, fail 2 MDTs ========================================================== 05:51:22 (1713347482) SKIP: replay-single test_81d needs >= 2 MDTs SKIP 81d (1s) == replay-single test 81e: DNE: unlink remote dir, drop MDT1 req reply, fail MDT0 ========================================================== 05:51:23 (1713347483) SKIP: replay-single test_81e needs >= 2 MDTs SKIP 81e (1s) == replay-single test 81f: DNE: unlink remote dir, drop MDT1 req reply, fail MDT1 ========================================================== 05:51:24 (1713347484) SKIP: replay-single test_81f needs >= 2 MDTs SKIP 81f (0s) == replay-single test 81g: DNE: unlink remote dir, drop req reply, fail M0, then M1 ========================================================== 05:51:24 (1713347484) SKIP: replay-single test_81g needs >= 2 MDTs SKIP 81g (1s) == replay-single test 81h: DNE: unlink remote dir, drop request reply, fail 2 MDTs ========================================================== 05:51:25 (1713347485) SKIP: replay-single test_81h needs >= 2 MDTs SKIP 81h (1s) == replay-single test 84a: stale open during export disconnect ========================================================== 05:51:26 (1713347486) fail_loc=0x80000144 total: 1 open/close in 0.01 seconds: 160.92 ops/second oleg412-client: error: invalid path '/mnt/lustre': Input/output error pdsh@oleg412-client: oleg412-client: ssh exited with exit code 5 PASS 84a (5s) == replay-single test 85a: check the cancellation of unused locks during recovery(IBITS) ========================================================== 05:51:31 (1713347491) before recovery: unused locks count = 301 Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec after recovery: unused locks count = 101 PASS 85a (27s) == replay-single test 85b: check the cancellation of unused locks during recovery(EXTENT) ========================================================== 05:51:58 (1713347518) before recovery: unused locks count = 100 Failing ost1 on oleg412-server Stopping /mnt/lustre-ost1 (opts:) on oleg412-server reboot facets: ost1 Failover ost1 to oleg412-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-OST0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec after recovery: unused locks count = 0 PASS 85b (23s) == replay-single test 86: umount server after clear nid_stats should not hit LBUG ========================================================== 05:52:21 (1713347541) Stopping clients: oleg412-client.virtnet /mnt/lustre (opts:) Stopping client oleg412-client.virtnet /mnt/lustre opts: mdt.lustre-MDT0000.exports.clear=0 Stopping /mnt/lustre-mds1 (opts:) on oleg412-server Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 Starting client oleg412-client.virtnet: -o user_xattr,flock oleg412-server@tcp:/lustre /mnt/lustre Started clients oleg412-client.virtnet: 192.168.204.112@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt) PASS 86 (6s) == replay-single test 87a: write replay ================== 05:52:27 (1713347547) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3712 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 15360 3753984 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 18432 7520256 1% /mnt/lustre 8+0 records in 8+0 records out 8388608 bytes (8.4 MB) copied, 0.101472 s, 82.7 MB/s Failing ost1 on oleg412-server Stopping /mnt/lustre-ost1 (opts:) on oleg412-server reboot facets: ost1 Failover ost1 to oleg412-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-OST0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec 8+0 records in 8+0 records out 8388608 bytes (8.4 MB) copied, 0.0610013 s, 138 MB/s PASS 87a (21s) == replay-single test 87b: write replay with changed data (checksum resend) ========================================================== 05:52:48 (1713347568) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3712 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 11264 3758080 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 14336 7524352 1% /mnt/lustre 8+0 records in 8+0 records out 8388608 bytes (8.4 MB) copied, 0.120525 s, 69.6 MB/s 8+0 records in 8+0 records out 8 bytes (8 B) copied, 0.00184095 s, 4.3 kB/s Failing ost1 on oleg412-server Stopping /mnt/lustre-ost1 (opts:) on oleg412-server reboot facets: ost1 Failover ost1 to oleg412-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-OST0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec 0+1 records in 0+1 records out 72 bytes (72 B) copied, 0.00468224 s, 15.4 kB/s PASS 87b (23s) == replay-single test 88: MDS should not assign same objid to different files ========================================================== 05:53:11 (1713347591) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3712 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 12288 3757056 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 15360 7523328 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3712 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 12288 3757056 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 15360 7523328 1% /mnt/lustre before test: last_id = 18337, next_id = 18306 Creating to objid 18337 on ost lustre-OST0000... total: 33 open/close in 0.07 seconds: 454.30 ops/second total: 8 open/close in 0.02 seconds: 503.90 ops/second before recovery: last_id = 18369, next_id = 18346 Stopping /mnt/lustre-mds1 (opts:) on oleg412-server Stopping /mnt/lustre-ost1 (opts:) on oleg412-server Failover mds1 to oleg412-server Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 Failover ost1 to oleg412-server Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-OST0000 after recovery: last_id = 18377, next_id = 18346 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0224392 s, 23.4 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0214311 s, 24.5 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0206743 s, 25.4 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0215445 s, 24.3 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0232898 s, 22.5 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0301954 s, 17.4 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0209205 s, 25.1 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.021479 s, 24.4 MB/s -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18306 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18307 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18308 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18309 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18310 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18311 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18312 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18313 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18314 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18315 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18316 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18317 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18318 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18319 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18320 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18321 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18322 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18323 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18324 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18325 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18326 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18327 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18328 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18329 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18330 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18331 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18332 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18333 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18334 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18335 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18336 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18337 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18338 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18339 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18340 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18341 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18342 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18343 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18344 -rw-r--r-- 1 root root 0 Apr 17 05:53 /mnt/lustre/d88.replay-single/f-18345 -rw-r--r-- 1 root root 524288 Apr 17 05:54 /mnt/lustre/d88.replay-single/f-18349 -rw-r--r-- 1 root root 524288 Apr 17 05:54 /mnt/lustre/d88.replay-single/f-18350 -rw-r--r-- 1 root root 524288 Apr 17 05:54 /mnt/lustre/d88.replay-single/f-18351 -rw-r--r-- 1 root root 524288 Apr 17 05:54 /mnt/lustre/d88.replay-single/f-18352 -rw-r--r-- 1 root root 524288 Apr 17 05:54 /mnt/lustre/d88.replay-single/f-18353 -rw-r--r-- 1 root root 524288 Apr 17 05:54 /mnt/lustre/d88.replay-single/f-18354 -rw-r--r-- 1 root root 524288 Apr 17 05:54 /mnt/lustre/d88.replay-single/f-18355 -rw-r--r-- 1 root root 524288 Apr 17 05:54 /mnt/lustre/d88.replay-single/f-18356 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.022424 s, 23.4 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0240593 s, 21.8 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0183526 s, 28.6 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0252369 s, 20.8 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0219802 s, 23.9 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0179296 s, 29.2 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0189148 s, 27.7 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0236856 s, 22.1 MB/s PASS 88 (57s) == replay-single test 89: no disk space leak on late ost connection ========================================================== 05:54:08 (1713347648) Waiting for orphan cleanup... osp.lustre-OST0000-osc-MDT0000.old_sync_processed osp.lustre-OST0001-osc-MDT0000.old_sync_processed wait 40 secs maximumly for oleg412-server mds-ost sync done. sleep 5 for ZFS zfs Waiting for MDT destroys to complete 10+0 records in 10+0 records out 41943040 bytes (42 MB) copied, 0.235503 s, 178 MB/s Stopping /mnt/lustre-ost1 (opts:) on oleg412-server Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-OST0000 Starting client: oleg412-client.virtnet: -o user_xattr,flock oleg412-server@tcp:/lustre /mnt/lustre osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 68 sec Waiting for orphan cleanup... osp.lustre-OST0000-osc-MDT0000.old_sync_processed osp.lustre-OST0001-osc-MDT0000.old_sync_processed wait 40 secs maximumly for oleg412-server mds-ost sync done. sleep 5 for ZFS zfs sleep 5 for ZFS zfs Waiting for MDT destroys to complete free_before: 7519232 free_after: 7519232 PASS 89 (126s) == replay-single test 90: lfs find identifies the missing striped file segments ========================================================== 05:56:14 (1713347774) Create the files Fail ost1 lustre-OST0000_UUID, display the list of affected files Stopping /mnt/lustre-ost1 (opts:) on oleg412-server General Query: lfs find /mnt/lustre/d90.replay-single /mnt/lustre/d90.replay-single /mnt/lustre/d90.replay-single/f1 /mnt/lustre/d90.replay-single/all /mnt/lustre/d90.replay-single/f0 Querying files on shutdown ost1: lfs find --obd lustre-OST0000_UUID /mnt/lustre/d90.replay-single/all /mnt/lustre/d90.replay-single/f0 Check getstripe: /home/green/git/lustre-release/lustre/utils/lfs getstripe -r --obd lustre-OST0000_UUID /mnt/lustre/d90.replay-single/all lmm_stripe_count: 2 lmm_stripe_size: 1048576 lmm_pattern: raid0 lmm_layout_gen: 0 lmm_stripe_offset: 0 obdidx objid objid group 0 18378 0x47ca 0 * /mnt/lustre/d90.replay-single/f0 lmm_stripe_count: 1 lmm_stripe_size: 1048576 lmm_pattern: raid0 lmm_layout_gen: 0 lmm_stripe_offset: 0 obdidx objid objid group 0 18379 0x47cb 0 * /mnt/lustre/d90.replay-single/all /mnt/lustre/d90.replay-single/f0 Failover ost1 to oleg412-server Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-OST0000 PASS 90 (18s) == replay-single test 93a: replay + reconnect ============ 05:56:32 (1713347792) 1+0 records in 1+0 records out 1024 bytes (1.0 kB) copied, 0.00131928 s, 776 kB/s fail_val=40 fail_loc=0x715 Failing ost1 on oleg412-server Stopping /mnt/lustre-ost1 (opts:) on oleg412-server reboot facets: ost1 Failover ost1 to oleg412-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-OST0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec PASS 93a (59s) == replay-single test 93b: replay + reconnect on mds ===== 05:57:31 (1713347851) total: 20 open/close in 0.06 seconds: 344.14 ops/second fail_val=80 fail_loc=0x715 Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 93b (113s) == replay-single test 100a: DNE: create striped dir, drop update rep from MDT1, fail MDT1 ========================================================== 05:59:24 (1713347964) SKIP: replay-single test_100a needs >= 2 MDTs SKIP 100a (1s) == replay-single test 100b: DNE: create striped dir, fail MDT0 ========================================================== 05:59:25 (1713347965) SKIP: replay-single test_100b needs >= 2 MDTs SKIP 100b (1s) == replay-single test 100c: DNE: create striped dir, fail MDT0 ========================================================== 05:59:26 (1713347966) SKIP: replay-single test_100c needs >= 2 MDTs SKIP 100c (0s) == replay-single test 101: Shouldn't reassign precreated objs to other files after recovery ========================================================== 05:59:26 (1713347966) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3712 2204928 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 20480 3748864 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 23552 7515136 1% /mnt/lustre Stopping /mnt/lustre-mds1 (opts:) on oleg412-server Failover mds1 to oleg412-server Starting mds1: -o localrecov -o abort_recovery lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 PASS 101 (22s) == replay-single test 102a: check resend (request lost) with multiple modify RPCs in flight ========================================================== 05:59:48 (1713347988) creating 7 files ... fail_loc=0x159 launch 7 chmod in parallel (05:59:48) ... fail_loc=0 done (05:59:55) /mnt/lustre/d102a.replay-single/file-1 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-2 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-3 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-4 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-5 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-6 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-7 has perms 0600 OK PASS 102a (8s) == replay-single test 102b: check resend (reply lost) with multiple modify RPCs in flight ========================================================== 05:59:56 (1713347996) creating 7 files ... fail_loc=0x15a launch 7 chmod in parallel (05:59:57) ... fail_loc=0 done (06:00:04) /mnt/lustre/d102b.replay-single/file-1 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-2 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-3 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-4 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-5 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-6 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-7 has perms 0600 OK PASS 102b (10s) == replay-single test 102c: check replay w/o reconstruction with multiple mod RPCs in flight ========================================================== 06:00:06 (1713348006) creating 7 files ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 20480 3334144 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3351552 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 23552 6685696 1% /mnt/lustre fail_loc=0x15a launch 7 chmod in parallel (06:00:08) ... fail_loc=0 Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec done (06:00:35) /mnt/lustre/d102c.replay-single/file-1 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-2 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-3 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-4 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-5 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-6 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-7 has perms 0600 OK PASS 102c (30s) == replay-single test 102d: check replay & reconstruction with multiple mod RPCs in flight ========================================================== 06:00:36 (1713348036) creating 7 files ... fail_loc=0x15a launch 7 chmod in parallel (06:00:37) ... fail_loc=0 Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec done (06:01:04) /mnt/lustre/d102d.replay-single/file-1 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-2 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-3 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-4 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-5 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-6 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-7 has perms 0600 OK PASS 102d (29s) == replay-single test 103: Check otr_next_id overflow ==== 06:01:05 (1713348065) fail_loc=0x80000162 total: 30 open/close in 0.06 seconds: 534.96 ops/second Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 103 (28s) == replay-single test 110a: DNE: create striped dir, fail MDT1 ========================================================== 06:01:33 (1713348093) SKIP: replay-single test_110a needs >= 2 MDTs SKIP 110a (2s) == replay-single test 110b: DNE: create striped dir, fail MDT1 and client ========================================================== 06:01:35 (1713348095) SKIP: replay-single test_110b needs >= 2 MDTs SKIP 110b (1s) == replay-single test 110c: DNE: create striped dir, fail MDT2 ========================================================== 06:01:36 (1713348096) SKIP: replay-single test_110c needs >= 2 MDTs SKIP 110c (1s) == replay-single test 110d: DNE: create striped dir, fail MDT2 and client ========================================================== 06:01:37 (1713348097) SKIP: replay-single test_110d needs >= 2 MDTs SKIP 110d (1s) == replay-single test 110e: DNE: create striped dir, uncommit on MDT2, fail client/MDT1/MDT2 ========================================================== 06:01:38 (1713348098) SKIP: replay-single test_110e needs >= 2 MDTs SKIP 110e (1s) SKIP: replay-single test_110f skipping excluded test 110f == replay-single test 110g: DNE: create striped dir, uncommit on MDT1, fail client/MDT1/MDT2 ========================================================== 06:01:40 (1713348100) SKIP: replay-single test_110g needs >= 2 MDTs SKIP 110g (0s) == replay-single test 111a: DNE: unlink striped dir, fail MDT1 ========================================================== 06:01:40 (1713348100) SKIP: replay-single test_111a needs >= 2 MDTs SKIP 111a (1s) == replay-single test 111b: DNE: unlink striped dir, fail MDT2 ========================================================== 06:01:41 (1713348101) SKIP: replay-single test_111b needs >= 2 MDTs SKIP 111b (1s) == replay-single test 111c: DNE: unlink striped dir, uncommit on MDT1, fail client/MDT1/MDT2 ========================================================== 06:01:42 (1713348102) SKIP: replay-single test_111c needs >= 2 MDTs SKIP 111c (1s) == replay-single test 111d: DNE: unlink striped dir, uncommit on MDT2, fail client/MDT1/MDT2 ========================================================== 06:01:43 (1713348103) SKIP: replay-single test_111d needs >= 2 MDTs SKIP 111d (0s) == replay-single test 111e: DNE: unlink striped dir, uncommit on MDT2, fail MDT1/MDT2 ========================================================== 06:01:43 (1713348103) SKIP: replay-single test_111e needs >= 2 MDTs SKIP 111e (1s) == replay-single test 111f: DNE: unlink striped dir, uncommit on MDT1, fail MDT1/MDT2 ========================================================== 06:01:44 (1713348104) SKIP: replay-single test_111f needs >= 2 MDTs SKIP 111f (1s) == replay-single test 111g: DNE: unlink striped dir, fail MDT1/MDT2 ========================================================== 06:01:45 (1713348105) SKIP: replay-single test_111g needs >= 2 MDTs SKIP 111g (1s) == replay-single test 112a: DNE: cross MDT rename, fail MDT1 ========================================================== 06:01:46 (1713348106) SKIP: replay-single test_112a needs >= 4 MDTs SKIP 112a (1s) == replay-single test 112b: DNE: cross MDT rename, fail MDT2 ========================================================== 06:01:47 (1713348107) SKIP: replay-single test_112b needs >= 4 MDTs SKIP 112b (1s) == replay-single test 112c: DNE: cross MDT rename, fail MDT3 ========================================================== 06:01:48 (1713348108) SKIP: replay-single test_112c needs >= 4 MDTs SKIP 112c (1s) == replay-single test 112d: DNE: cross MDT rename, fail MDT4 ========================================================== 06:01:49 (1713348109) SKIP: replay-single test_112d needs >= 4 MDTs SKIP 112d (1s) == replay-single test 112e: DNE: cross MDT rename, fail MDT1 and MDT2 ========================================================== 06:01:50 (1713348110) SKIP: replay-single test_112e needs >= 4 MDTs SKIP 112e (1s) == replay-single test 112f: DNE: cross MDT rename, fail MDT1 and MDT3 ========================================================== 06:01:51 (1713348111) SKIP: replay-single test_112f needs >= 4 MDTs SKIP 112f (0s) == replay-single test 112g: DNE: cross MDT rename, fail MDT1 and MDT4 ========================================================== 06:01:51 (1713348111) SKIP: replay-single test_112g needs >= 4 MDTs SKIP 112g (1s) == replay-single test 112h: DNE: cross MDT rename, fail MDT2 and MDT3 ========================================================== 06:01:52 (1713348112) SKIP: replay-single test_112h needs >= 4 MDTs SKIP 112h (1s) == replay-single test 112i: DNE: cross MDT rename, fail MDT2 and MDT4 ========================================================== 06:01:53 (1713348113) SKIP: replay-single test_112i needs >= 4 MDTs SKIP 112i (1s) == replay-single test 112j: DNE: cross MDT rename, fail MDT3 and MDT4 ========================================================== 06:01:54 (1713348114) SKIP: replay-single test_112j needs >= 4 MDTs SKIP 112j (1s) == replay-single test 112k: DNE: cross MDT rename, fail MDT1,MDT2,MDT3 ========================================================== 06:01:55 (1713348115) SKIP: replay-single test_112k needs >= 4 MDTs SKIP 112k (1s) == replay-single test 112l: DNE: cross MDT rename, fail MDT1,MDT2,MDT4 ========================================================== 06:01:56 (1713348116) SKIP: replay-single test_112l needs >= 4 MDTs SKIP 112l (0s) == replay-single test 112m: DNE: cross MDT rename, fail MDT1,MDT3,MDT4 ========================================================== 06:01:56 (1713348116) SKIP: replay-single test_112m needs >= 4 MDTs SKIP 112m (1s) == replay-single test 112n: DNE: cross MDT rename, fail MDT2,MDT3,MDT4 ========================================================== 06:01:57 (1713348117) SKIP: replay-single test_112n needs >= 4 MDTs SKIP 112n (1s) == replay-single test 115: failover for create/unlink striped directory ========================================================== 06:01:58 (1713348118) SKIP: replay-single test_115 needs >= 2 MDTs SKIP 115 (1s) == replay-single test 116a: large update log master MDT recovery ========================================================== 06:01:59 (1713348119) SKIP: replay-single test_116a needs >= 2 MDTs SKIP 116a (1s) == replay-single test 116b: large update log slave MDT recovery ========================================================== 06:02:00 (1713348120) SKIP: replay-single test_116b needs >= 2 MDTs SKIP 116b (1s) == replay-single test 117: DNE: cross MDT unlink, fail MDT1 and MDT2 ========================================================== 06:02:01 (1713348121) SKIP: replay-single test_117 needs >= 4 MDTs SKIP 117 (1s) == replay-single test 118: invalidate osp update will not cause update log corruption ========================================================== 06:02:02 (1713348122) SKIP: replay-single test_118 needs >= 2 MDTs SKIP 118 (0s) == replay-single test 119: timeout of normal replay does not cause DNE replay fails ========================================================== 06:02:02 (1713348122) SKIP: replay-single test_119 needs >= 2 MDTs SKIP 119 (1s) == replay-single test 120: DNE fail abort should stop both normal and DNE replay ========================================================== 06:02:03 (1713348123) SKIP: replay-single test_120 needs >= 2 MDTs SKIP 120 (1s) == replay-single test 121: lock replay timed out and race ========================================================== 06:02:04 (1713348124) multiop /mnt/lustre/f121.replay-single vs_s TMPPIPE=/tmp/multiop_open_wait_pipe.6842 Stopping /mnt/lustre-mds1 (opts:) on oleg412-server Failover mds1 to oleg412-server fail_loc=0x721 fail_val=0 at_max=0 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 fail_loc=0x0 at_max=600 PASS 121 (15s) == replay-single test 130a: DoM file create (setstripe) replay ========================================================== 06:02:19 (1713348139) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 20480 3334144 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 23552 7100416 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 130a (33s) == replay-single test 130b: DoM file create (inherited) replay ========================================================== 06:02:53 (1713348173) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 20480 3334144 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 23552 7100416 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 130b (26s) == replay-single test 131a: DoM file write lock replay === 06:03:18 (1713348198) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 20480 3334144 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 23552 7100416 1% /mnt/lustre 1+0 records in 1+0 records out 8 bytes (8 B) copied, 0.00108928 s, 7.3 kB/s Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 131a (31s) SKIP: replay-single test_131b skipping excluded test 131b == replay-single test 132a: PFL new component instantiate replay ========================================================== 06:03:49 (1713348229) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204032 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 20480 3334144 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 3072 3766272 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 23552 7100416 1% /mnt/lustre 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0162903 s, 64.4 MB/s /mnt/lustre/f132a.replay-single lcm_layout_gen: 3 lcm_mirror_count: 1 lcm_entry_count: 2 lcme_id: 1 lcme_mirror_id: 0 lcme_flags: init lcme_extent.e_start: 0 lcme_extent.e_end: 1048576 lmm_stripe_count: 1 lmm_stripe_size: 1048576 lmm_pattern: raid0 lmm_layout_gen: 0 lmm_stripe_offset: 1 lmm_objects: - 0: { l_ost_idx: 1, l_fid: [0x100010000:0x2ca2:0x0] } lcme_id: 2 lcme_mirror_id: 0 lcme_flags: init lcme_extent.e_start: 1048576 lcme_extent.e_end: EOF lmm_stripe_count: 2 lmm_stripe_size: 1048576 lmm_pattern: raid0 lmm_layout_gen: 0 lmm_stripe_offset: 0 lmm_objects: - 0: { l_ost_idx: 0, l_fid: [0x100000000:0x4d2a:0x0] } - 1: { l_ost_idx: 1, l_fid: [0x100010000:0x2ca3:0x0] } Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f132a.replay-single lcm_layout_gen: 1 lcm_mirror_count: 1 lcm_entry_count: 2 lcme_id: 1 lcme_mirror_id: 0 lcme_flags: init lcme_extent.e_start: 0 lcme_extent.e_end: 1048576 lmm_stripe_count: 1 lmm_stripe_size: 1048576 lmm_pattern: raid0 lmm_layout_gen: 65535 lmm_stripe_offset: 1 lmm_objects: - 0: { l_ost_idx: 1, l_fid: [0x100010000:0x2ca2:0x0] } lcme_id: 2 lcme_mirror_id: 0 lcme_flags: init lcme_extent.e_start: 1048576 lcme_extent.e_end: EOF lmm_stripe_count: 2 lmm_stripe_size: 1048576 lmm_pattern: raid0 lmm_layout_gen: 65535 lmm_stripe_offset: 0 lmm_objects: - 0: { l_ost_idx: 0, l_fid: [0x100000000:0x4d2a:0x0] } - 1: { l_ost_idx: 1, l_fid: [0x100010000:0x2ca3:0x0] } PASS 132a (25s) == replay-single test 133: check resend of ongoing requests for lwp during failover ========================================================== 06:04:15 (1713348255) SKIP: replay-single test_133 needs >= 2 MDTs SKIP 133 (1s) == replay-single test 134: replay creation of a file created in a pool ========================================================== 06:04:15 (1713348255) Creating new pool oleg412-server: Pool lustre.pool_134 created Adding targets to pool oleg412-server: OST lustre-OST0001_UUID added to pool lustre.pool_134 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 20480 3334144 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 24576 7099392 1% /mnt/lustre Failing mds1 on oleg412-server Stopping /mnt/lustre-mds1 (opts:) on oleg412-server reboot facets: mds1 Failover mds1 to oleg412-server mount facets: mds1 Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Destroy the created pools: pool_134 lustre.pool_134 oleg412-server: OST lustre-OST0001_UUID removed from pool lustre.pool_134 oleg412-server: Pool lustre.pool_134 destroyed Waiting 90s for 'foo' PASS 134 (41s) == replay-single test 135: Server failure in lock replay phase ========================================================== 06:04:56 (1713348296) Failing ost1 on oleg412-server Stopping /mnt/lustre-ost1 (opts:) on oleg412-server reboot facets: ost1 Failover ost1 to oleg412-server mount facets: ost1 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-OST0000 oleg412-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 2210688 3840 2204800 1% /mnt/lustre[MDT:0] lustre-OST0000_UUID 3771392 20480 3748864 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3771392 4096 3765248 1% /mnt/lustre[OST:1] filesystem_summary: 7542784 24576 7514112 1% /mnt/lustre Stopping /mnt/lustre-ost1 (opts:) on oleg412-server Failover ost1 to oleg412-server oleg412-server: oleg412-server.virtnet: executing load_module ../libcfs/libcfs/libcfs fail_val=20 fail_loc=0x32d Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-OST0000 Stopping /mnt/lustre-ost1 (opts:) on oleg412-server Failover ost1 to oleg412-server oleg412-server: oleg412-server.virtnet: executing load_module ../libcfs/libcfs/libcfs fail_loc=0 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-OST0000 Stopping /mnt/lustre-ost1 (opts:-f) on oleg412-server Stopping /mnt/lustre-ost2 (opts:-f) on oleg412-server Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-OST0000 Starting ost2: -o localrecov lustre-ost2/ost2 /mnt/lustre-ost2 oleg412-server: oleg412-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all 8 pdsh@oleg412-client: oleg412-server: ssh exited with exit code 1 Started lustre-OST0001 pdsh@oleg412-client: oleg412-client: ssh exited with exit code 5 PASS 135 (98s) == replay-single test complete, duration 4968 sec ======== 06:06:34 (1713348394)