-----============= acceptance-small: replay-single ============----- Thu Apr 18 03:48:11 EDT 2024 excepting tests: 110f 131b 59 36 === replay-single: start setup 03:48:15 (1713426495) === oleg328-client.virtnet: executing check_config_client /mnt/lustre oleg328-client.virtnet: Checking config lustre mounted on /mnt/lustre Checking servers environments Checking clients oleg328-client.virtnet environments Using TIMEOUT=20 osc.lustre-OST0000-osc-ffff8800b5898000.idle_timeout=debug osc.lustre-OST0001-osc-ffff8800b5898000.idle_timeout=debug disable quota as required oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all osd-ldiskfs.track_declares_assert=1 === replay-single: finish setup 03:48:23 (1713426503) === debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 0a: empty replay =================== 03:48:24 (1713426504) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1772 1285916 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1612 1286076 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1524 3605496 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1524 3605496 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3048 7210992 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 03:48:27 (1713426507) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 03:48:42 (1713426522) targets are mounted 03:48:42 (1713426522) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 0a (25s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 0b: ensure object created after recover exists. (3284) ========================================================== 03:48:50 (1713426530) Failing ost1 on oleg328-server Stopping /mnt/lustre-ost1 (opts:) on oleg328-server 03:48:51 (1713426531) shut down Failover ost1 to oleg328-server mount facets: ost1 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-OST0000 03:49:05 (1713426545) targets are mounted 03:49:05 (1713426545) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec total: 20 open/close in 0.07 seconds: 278.87 ops/second - unlinked 0 (time 1713426548 ; total 0 ; last 0) total: 20 unlinks in 0 seconds: inf unlinks/second PASS 0b (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 0c: check replay-barrier =========== 03:49:11 (1713426551) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1768 1285920 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1612 1286076 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1540 3605480 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1524 3605496 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3064 7210976 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 03:49:14 (1713426554) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 03:49:28 (1713426568) targets are mounted 03:49:28 (1713426568) facet_failover done Starting client: oleg328-client.virtnet: -o user_xattr,flock oleg328-server@tcp:/lustre /mnt/lustre rm: cannot remove '/mnt/lustre/f0c.replay-single': No such file or directory PASS 0c (90s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 0d: expired recovery with no clients ========================================================== 03:50:43 (1713426643) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1768 1285920 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1612 1286076 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1540 3605480 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1524 3605496 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3064 7210976 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 03:50:46 (1713426646) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 03:51:00 (1713426660) targets are mounted 03:51:00 (1713426660) facet_failover done Starting client: oleg328-client.virtnet: -o user_xattr,flock oleg328-server@tcp:/lustre /mnt/lustre PASS 0d (90s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 1: simple create =================== 03:52:14 (1713426734) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1768 1285920 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1612 1286076 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1540 3605480 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1524 3605496 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3064 7210976 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 03:52:17 (1713426737) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 03:52:31 (1713426751) targets are mounted 03:52:31 (1713426751) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f1.replay-single has type file OK PASS 1 (24s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 2a: touch ========================== 03:52:40 (1713426760) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1768 1285920 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1612 1286076 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1540 3605480 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1524 3605496 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3064 7210976 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 03:52:43 (1713426763) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 03:52:57 (1713426777) targets are mounted 03:52:57 (1713426777) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f2a.replay-single has type file OK PASS 2a (24s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 2b: touch ========================== 03:53:06 (1713426786) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1768 1285920 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1612 1286076 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1540 3605480 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1524 3605496 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3064 7210976 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 03:53:09 (1713426789) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 03:53:24 (1713426804) targets are mounted 03:53:24 (1713426804) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f2b.replay-single has type file OK PASS 2b (25s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 2c: setstripe replay =============== 03:53:33 (1713426813) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1768 1285920 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1612 1286076 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1540 3605480 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1524 3605496 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3064 7210976 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 03:53:36 (1713426816) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 03:53:51 (1713426831) targets are mounted 03:53:51 (1713426831) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f2c.replay-single has type file OK PASS 2c (25s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 2d: setdirstripe replay ============ 03:54:00 (1713426840) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1768 1285920 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1612 1286076 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1540 3605480 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1524 3605496 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3064 7210976 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 03:54:04 (1713426844) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 03:54:18 (1713426858) targets are mounted 03:54:18 (1713426858) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d2d.replay-single has type dir OK PASS 2d (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 2e: O_CREAT|O_EXCL create replay === 03:54:28 (1713426868) fail_loc=0x8000013b UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1852 1285836 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1540 3605480 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1524 3605496 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3064 7210976 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 03:54:33 (1713426873) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 03:54:48 (1713426888) targets are mounted 03:54:48 (1713426888) facet_failover done Succeed in opening file "/mnt/lustre/f2e.replay-single"(flags=O_CREAT) oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f2e.replay-single has type file OK PASS 2e (27s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 3a: replay failed open(O_DIRECTORY) ========================================================== 03:54:57 (1713426897) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1852 1285836 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1540 3605480 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1524 3605496 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3064 7210976 1% /mnt/lustre Error in opening file "/mnt/lustre/f3a.replay-single"(flags=O_DIRECTORY) 20: Not a directory Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 03:55:02 (1713426902) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 03:55:16 (1713426916) targets are mounted 03:55:16 (1713426916) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f3a.replay-single has type file OK PASS 3a (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 3b: replay failed open -ENOMEM ===== 03:55:25 (1713426925) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1852 1285836 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1540 3605480 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1524 3605496 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3064 7210976 1% /mnt/lustre fail_loc=0x80000114 touch: cannot touch '/mnt/lustre/f3b.replay-single': Cannot allocate memory fail_loc=0 Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 03:55:29 (1713426929) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 03:55:43 (1713426943) targets are mounted 03:55:43 (1713426943) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/f3b.replay-single: No such file or directory PASS 3b (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 3c: replay failed open -ENOMEM ===== 03:55:53 (1713426953) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1852 1285836 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1540 3605480 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1524 3605496 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3064 7210976 1% /mnt/lustre fail_loc=0x80000128 touch: cannot touch '/mnt/lustre/f3c.replay-single': Cannot allocate memory fail_loc=0 Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 03:55:57 (1713426957) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 03:56:12 (1713426972) targets are mounted 03:56:12 (1713426972) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/f3c.replay-single: No such file or directory PASS 3c (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 4a: |x| 10 open(O_CREAT)s ========== 03:56:20 (1713426980) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1852 1285836 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1540 3605480 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1524 3605496 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3064 7210976 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 03:56:25 (1713426985) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 03:56:40 (1713427000) targets are mounted 03:56:40 (1713427000) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 4a (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 4b: |x| rm 10 files ================ 03:56:50 (1713427010) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1852 1285836 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1560 3605460 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1544 3605476 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3104 7210936 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 03:56:55 (1713427015) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 03:57:09 (1713427029) targets are mounted 03:57:09 (1713427029) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/f4b.replay-single-*: No such file or directory PASS 4b (27s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 5: |x| 220 open(O_CREAT) =========== 03:57:19 (1713427039) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1852 1285836 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1560 3605460 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1544 3605476 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3104 7210936 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 03:57:25 (1713427045) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 03:57:40 (1713427060) targets are mounted 03:57:40 (1713427060) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 5 (36s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 6a: mkdir + contained create ======= 03:57:57 (1713427077) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1884 1285804 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1560 3605460 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1544 3605476 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3104 7210936 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 03:58:00 (1713427080) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 03:58:14 (1713427094) targets are mounted 03:58:14 (1713427094) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d6a.replay-single has type dir OK /mnt/lustre/d6a.replay-single/f6a.replay-single has type file OK PASS 6a (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 6b: |X| rmdir ====================== 03:58:24 (1713427104) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1880 1285808 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1560 3605460 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1544 3605476 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3104 7210936 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 03:58:27 (1713427107) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 03:58:41 (1713427121) targets are mounted 03:58:41 (1713427121) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/d6b.replay-single: No such file or directory PASS 6b (24s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 7: mkdir |X| contained create ====== 03:58:49 (1713427129) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1880 1285808 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1560 3605460 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1544 3605476 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3104 7210936 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 03:58:52 (1713427132) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 03:59:06 (1713427146) targets are mounted 03:59:06 (1713427146) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d7.replay-single has type dir OK /mnt/lustre/d7.replay-single/f7.replay-single has type file OK PASS 7 (24s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 8: creat open |X| close ============ 03:59:14 (1713427154) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1560 3605460 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1544 3605476 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3104 7210936 1% /mnt/lustre multiop /mnt/lustre/f8.replay-single vmo_c TMPPIPE=/tmp/multiop_open_wait_pipe.7515 Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 03:59:18 (1713427158) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 03:59:32 (1713427172) targets are mounted 03:59:32 (1713427172) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f8.replay-single /mnt/lustre/f8.replay-single has type file OK PASS 8 (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 9: |X| create (same inum/gen) ====== 03:59:42 (1713427182) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1560 3605460 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1544 3605476 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3104 7210936 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 03:59:47 (1713427187) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:00:02 (1713427202) targets are mounted 04:00:02 (1713427202) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec old_inum == 144115305935798546, new_inum == 144115305935798546 old_inum and new_inum match PASS 9 (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 10: create |X| rename unlink ======= 04:00:12 (1713427212) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1560 3605460 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1544 3605476 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3104 7210936 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:00:16 (1713427216) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:00:32 (1713427232) targets are mounted 04:00:32 (1713427232) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/f10.replay-single: No such file or directory PASS 10 (27s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 11: create open write rename |X| create-old-name read ========================================================== 04:00:42 (1713427242) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1560 3605460 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1544 3605476 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3104 7210936 1% /mnt/lustre new old Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:00:46 (1713427246) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:01:02 (1713427262) targets are mounted 04:01:02 (1713427262) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec new old PASS 11 (27s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 12: open, unlink |X| close ========= 04:01:11 (1713427271) multiop /mnt/lustre/f12.replay-single vo_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7515 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1548 3605472 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3112 7210928 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:01:16 (1713427276) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:01:31 (1713427291) targets are mounted 04:01:31 (1713427291) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 12 (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 13: open chmod 0 |x| write close === 04:01:41 (1713427301) multiop /mnt/lustre/f13.replay-single vO_wc TMPPIPE=/tmp/multiop_open_wait_pipe.7515 /mnt/lustre/f13.replay-single has perms 00 OK UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1548 3605472 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3112 7210928 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:01:46 (1713427306) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:02:01 (1713427321) targets are mounted 04:02:01 (1713427321) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f13.replay-single has perms 00 OK /mnt/lustre/f13.replay-single has size 1 OK PASS 13 (27s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 14: open(O_CREAT), unlink |X| close ========================================================== 04:02:11 (1713427331) multiop /mnt/lustre/f14.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7515 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1548 3605472 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3112 7210928 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:02:15 (1713427335) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:02:31 (1713427351) targets are mounted 04:02:31 (1713427351) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 14 (27s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 15: open(O_CREAT), unlink |X| touch new, close ========================================================== 04:02:41 (1713427361) multiop /mnt/lustre/f15.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7515 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1548 3605472 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3112 7210928 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:02:45 (1713427365) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:03:01 (1713427381) targets are mounted 04:03:01 (1713427381) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 15 (27s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 16: |X| open(O_CREAT), unlink, touch new, unlink new ========================================================== 04:03:10 (1713427390) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1548 3605472 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3112 7210928 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:03:15 (1713427395) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:03:31 (1713427411) targets are mounted 04:03:31 (1713427411) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 16 (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 17: |X| open(O_CREAT), |replay| close ========================================================== 04:03:40 (1713427420) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1548 3605472 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3112 7210928 1% /mnt/lustre multiop /mnt/lustre/f17.replay-single vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.7515 Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:03:45 (1713427425) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:04:00 (1713427440) targets are mounted 04:04:00 (1713427440) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f17.replay-single has type file OK PASS 17 (27s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 18: open(O_CREAT), unlink, touch new, close, touch, unlink ========================================================== 04:04:10 (1713427450) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1548 3605472 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3112 7210928 1% /mnt/lustre multiop /mnt/lustre/f18.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7515 pid: 24964 will close Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:04:14 (1713427454) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:04:30 (1713427470) targets are mounted 04:04:30 (1713427470) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 18 (27s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 19: mcreate, open, write, rename === 04:04:39 (1713427479) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1548 3605472 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3112 7210928 1% /mnt/lustre old Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:04:44 (1713427484) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:04:59 (1713427499) targets are mounted 04:04:59 (1713427499) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec old PASS 19 (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 20a: |X| open(O_CREAT), unlink, replay, close (test mds_cleanup_orphans) ========================================================== 04:05:09 (1713427509) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1552 3605468 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3116 7210924 1% /mnt/lustre multiop /mnt/lustre/f20a.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7515 Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:05:13 (1713427513) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:05:28 (1713427528) targets are mounted 04:05:28 (1713427528) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 20a (27s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 20b: write, unlink, eviction, replay (test mds_cleanup_orphans) ========================================================== 04:05:38 (1713427538) /mnt/lustre/f20b.replay-single lmm_stripe_count: 1 lmm_stripe_size: 4194304 lmm_pattern: raid0 lmm_layout_gen: 0 lmm_stripe_offset: 0 obdidx objid objid group 0 1090 0x442 0x280000401 10000+0 records in 10000+0 records out 40960000 bytes (41 MB) copied, 1.49307 s, 27.4 MB/s Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:05:42 (1713427542) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:05:55 (1713427555) targets are mounted 04:05:55 (1713427555) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec affected facets: mds1 oleg328-server: oleg328-server.virtnet: executing _wait_recovery_complete *.lustre-MDT0000.recovery_status 1475 oleg328-server: *.lustre-MDT0000.recovery_status status: COMPLETE Waiting for MDT destroys to complete before 3116, after 3116 PASS 20b (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 20c: check that client eviction does not affect file content ========================================================== 04:06:08 (1713427568) multiop /mnt/lustre/f20c.replay-single vOw_c TMPPIPE=/tmp/multiop_open_wait_pipe.7515 -rw-r--r-- 1 root root 1 Apr 18 04:06 /mnt/lustre/f20c.replay-single pdsh@oleg328-client: oleg328-client: ssh exited with exit code 5 PASS 20c (5s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 21: |X| open(O_CREAT), unlink touch new, replay, close (test mds_cleanup_orphans) ========================================================== 04:06:15 (1713427575) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1552 3605468 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3116 7210924 1% /mnt/lustre multiop /mnt/lustre/f21.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7515 Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:06:20 (1713427580) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:06:35 (1713427595) targets are mounted 04:06:35 (1713427595) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 21 (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 22: open(O_CREAT), |X| unlink, replay, close (test mds_cleanup_orphans) ========================================================== 04:06:45 (1713427605) multiop /mnt/lustre/f22.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7515 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3120 7210920 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:06:50 (1713427610) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:07:05 (1713427625) targets are mounted 04:07:05 (1713427625) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 22 (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 23: open(O_CREAT), |X| unlink touch new, replay, close (test mds_cleanup_orphans) ========================================================== 04:07:15 (1713427635) multiop /mnt/lustre/f23.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7515 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3120 7210920 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:07:20 (1713427640) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:07:35 (1713427655) targets are mounted 04:07:35 (1713427655) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 23 (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 24: open(O_CREAT), replay, unlink, close (test mds_cleanup_orphans) ========================================================== 04:07:45 (1713427665) multiop /mnt/lustre/f24.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7515 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3120 7210920 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:07:50 (1713427670) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:08:05 (1713427685) targets are mounted 04:08:05 (1713427685) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 24 (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 25: open(O_CREAT), unlink, replay, close (test mds_cleanup_orphans) ========================================================== 04:08:15 (1713427695) multiop /mnt/lustre/f25.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7515 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3120 7210920 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:08:20 (1713427700) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:08:35 (1713427715) targets are mounted 04:08:35 (1713427715) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 25 (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 26: |X| open(O_CREAT), unlink two, close one, replay, close one (test mds_cleanup_orphans) ========================================================== 04:08:45 (1713427725) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3120 7210920 1% /mnt/lustre multiop /mnt/lustre/f26.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7515 multiop /mnt/lustre/f26.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7515 Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:08:50 (1713427730) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:09:05 (1713427745) targets are mounted 04:09:05 (1713427745) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 26 (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 27: |X| open(O_CREAT), unlink two, replay, close two (test mds_cleanup_orphans) ========================================================== 04:09:15 (1713427755) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3120 7210920 1% /mnt/lustre multiop /mnt/lustre/f27.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7515 multiop /mnt/lustre/f27.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7515 Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:09:20 (1713427760) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:09:35 (1713427775) targets are mounted 04:09:35 (1713427775) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 27 (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 28: open(O_CREAT), |X| unlink two, close one, replay, close one (test mds_cleanup_orphans) ========================================================== 04:09:45 (1713427785) multiop /mnt/lustre/f28.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7515 multiop /mnt/lustre/f28.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7515 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3120 7210920 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:09:50 (1713427790) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:10:05 (1713427805) targets are mounted 04:10:05 (1713427805) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 28 (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 29: open(O_CREAT), |X| unlink two, replay, close two (test mds_cleanup_orphans) ========================================================== 04:10:15 (1713427815) multiop /mnt/lustre/f29.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7515 multiop /mnt/lustre/f29.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7515 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3120 7210920 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:10:20 (1713427820) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:10:35 (1713427835) targets are mounted 04:10:35 (1713427835) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 29 (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 30: open(O_CREAT) two, unlink two, replay, close two (test mds_cleanup_orphans) ========================================================== 04:10:45 (1713427845) multiop /mnt/lustre/f30.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7515 multiop /mnt/lustre/f30.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7515 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3120 7210920 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:10:50 (1713427850) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:11:05 (1713427865) targets are mounted 04:11:05 (1713427865) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 30 (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 31: open(O_CREAT) two, unlink one, |X| unlink one, close two (test mds_cleanup_orphans) ========================================================== 04:11:15 (1713427875) multiop /mnt/lustre/f31.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7515 multiop /mnt/lustre/f31.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7515 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3120 7210920 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:11:19 (1713427879) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:11:34 (1713427894) targets are mounted 04:11:34 (1713427894) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 31 (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 32: close() notices client eviction; close() after client eviction ========================================================== 04:11:43 (1713427903) multiop /mnt/lustre/f32.replay-single vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.7515 multiop /mnt/lustre/f32.replay-single vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.7515 pdsh@oleg328-client: oleg328-client: ssh exited with exit code 5 PASS 32 (4s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 33a: fid seq shouldn't be reused after abort recovery ========================================================== 04:11:49 (1713427909) total: 10 open/close in 0.05 seconds: 181.83 ops/second Replay barrier on lustre-MDT0000 Stopping /mnt/lustre-mds1 (opts:) on oleg328-server Failover mds1 to oleg328-server Starting mds1: -o localrecov -o abort_recovery /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 total: 10 open/close in 0.05 seconds: 217.27 ops/second PASS 33a (18s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 33b: test fid seq allocation ======= 04:12:09 (1713427929) fail_loc=0x1311 total: 10 open/close in 0.05 seconds: 203.83 ops/second Replay barrier on lustre-MDT0000 Stopping /mnt/lustre-mds1 (opts:) on oleg328-server Failover mds1 to oleg328-server Starting mds1: -o localrecov -o abort_recovery /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 total: 10 open/close in 0.07 seconds: 144.24 ops/second PASS 33b (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 34: abort recovery before client does replay (test mds_cleanup_orphans) ========================================================== 04:12:30 (1713427950) multiop /mnt/lustre/f34.replay-single vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.7515 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1908 1285780 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1720 1285968 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3120 7210920 1% /mnt/lustre Stopping /mnt/lustre-mds1 (opts:) on oleg328-server Failover mds1 to oleg328-server Starting mds1: -o localrecov -o abort_recovery /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg328-client: error: invalid path '/mnt/lustre': Input/output error pdsh@oleg328-client: oleg328-client: ssh exited with exit code 5 first stat failed: 5 PASS 34 (21s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 35: test recovery from llog for unlink op ========================================================== 04:12:53 (1713427973) fail_loc=0x80000119 Stopping /mnt/lustre-mds1 (opts:) on oleg328-server Failover mds1 to oleg328-server Starting mds1: -o localrecov -o abort_recovery /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 rm: cannot remove '/mnt/lustre/f35.replay-single': Input/output error Can't lstat /mnt/lustre/f35.replay-single: No such file or directory PASS 35 (19s) debug_raw_pointers=0 debug_raw_pointers=0 SKIP: replay-single test_36 skipping ALWAYS excluded test 36 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 37: abort recovery before client does replay (test mds_cleanup_orphans for directories) ========================================================== 04:13:15 (1713427995) multiop /mnt/lustre/d37.replay-single/f37.replay-single vdD_c TMPPIPE=/tmp/multiop_open_wait_pipe.7515 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1980 1285708 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1784 1285904 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3120 7210920 1% /mnt/lustre Stopping /mnt/lustre-mds1 (opts:) on oleg328-server Failover mds1 to oleg328-server Starting mds1: -o localrecov -o abort_recovery /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 PASS 37 (22s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 38: test recovery from unlink llog (test llog_gen_rec) ========================================================== 04:13:39 (1713428019) total: 800 open/close in 4.68 seconds: 171.12 ops/second - unlinked 0 (time 1713428026 ; total 0 ; last 0) total: 400 unlinks in 2 seconds: 200.000000 unlinks/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2108 1285580 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1816 1285872 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3120 7210920 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:13:52 (1713428032) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:14:08 (1713428048) targets are mounted 04:14:08 (1713428048) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec - unlinked 0 (time 1713428055 ; total 0 ; last 0) total: 400 unlinks in 2 seconds: 200.000000 unlinks/second Can't lstat /mnt/lustre/f38.replay-single-*: No such file or directory PASS 38 (42s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 39: test recovery from unlink llog (test llog_gen_rec) ========================================================== 04:14:23 (1713428063) total: 800 open/close in 3.38 seconds: 236.47 ops/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2108 1285580 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1816 1285872 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3120 7210920 1% /mnt/lustre - unlinked 0 (time 1713428071 ; total 0 ; last 0) total: 400 unlinks in 2 seconds: 200.000000 unlinks/second Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:14:34 (1713428074) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:14:49 (1713428089) targets are mounted 04:14:49 (1713428089) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec - unlinked 0 (time 1713428096 ; total 0 ; last 0) total: 400 unlinks in 2 seconds: 200.000000 unlinks/second Can't lstat /mnt/lustre/f39.replay-single-*: No such file or directory PASS 39 (39s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 41: read from a valid osc while other oscs are invalid ========================================================== 04:15:03 (1713428103) 1+0 records in 1+0 records out 4096 bytes (4.1 kB) copied, 0.00210513 s, 1.9 MB/s 1+0 records in 1+0 records out 4096 bytes (4.1 kB) copied, 0.00446835 s, 917 kB/s PASS 41 (3s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 42: recovery after ost failure ===== 04:15:07 (1713428107) total: 800 open/close in 3.20 seconds: 250.25 ops/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2132 1285556 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1816 1285872 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre - unlinked 0 (time 1713428115 ; total 0 ; last 0) total: 400 unlinks in 1 seconds: 400.000000 unlinks/second debug=-1 Failing ost1 on oleg328-server Stopping /mnt/lustre-ost1 (opts:) on oleg328-server 04:15:18 (1713428118) shut down Failover ost1 to oleg328-server mount facets: ost1 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-OST0000 04:15:33 (1713428133) targets are mounted 04:15:33 (1713428133) facet_failover done wait for MDS to timeout and recover debug=super ioctl neterror warning dlmtrace error emerg ha rpctrace vfstrace config console lfsck debug=super ioctl neterror warning dlmtrace error emerg ha rpctrace vfstrace config console lfsck - unlinked 0 (time 1713428174 ; total 0 ; last 0) total: 400 unlinks in 1 seconds: 400.000000 unlinks/second Can't lstat /mnt/lustre/f42.replay-single-*: No such file or directory PASS 42 (69s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 43: mds osc import failure during recovery; don't LBUG ========================================================== 04:16:18 (1713428178) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2188 1285500 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1816 1285872 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre fail_loc=0x80000204 Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:16:22 (1713428182) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:16:37 (1713428197) targets are mounted 04:16:37 (1713428197) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 43 (36s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 44a: race in target handle connect ========================================================== 04:16:56 (1713428216) at_max=40 1 of 10 (1713428217) service : cur 5 worst 5 (at 1713426456, 1762s ago) 4 4 4 4 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2084 1285604 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1816 1285872 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre 2 of 10 (1713428223) service : cur 5 worst 5 (at 1713426456, 1767s ago) 5 4 4 4 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2084 1285604 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1816 1285872 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre 3 of 10 (1713428228) service : cur 5 worst 5 (at 1713426456, 1773s ago) 5 4 4 4 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2084 1285604 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1816 1285872 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre 4 of 10 (1713428234) service : cur 5 worst 5 (at 1713426456, 1778s ago) 5 4 4 4 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2084 1285604 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1816 1285872 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre 5 of 10 (1713428239) service : cur 5 worst 5 (at 1713426456, 1784s ago) 5 4 4 4 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2084 1285604 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1816 1285872 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre 6 of 10 (1713428245) service : cur 5 worst 5 (at 1713426456, 1790s ago) 5 4 4 4 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2084 1285604 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1816 1285872 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre 7 of 10 (1713428251) service : cur 5 worst 5 (at 1713426456, 1795s ago) 5 4 4 4 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2084 1285604 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1816 1285872 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre 8 of 10 (1713428256) service : cur 5 worst 5 (at 1713426456, 1801s ago) 5 5 4 4 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2084 1285604 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1816 1285872 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre 9 of 10 (1713428262) service : cur 5 worst 5 (at 1713426456, 1807s ago) 5 5 4 4 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2084 1285604 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1816 1285872 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre 10 of 10 (1713428267) service : cur 5 worst 5 (at 1713426456, 1812s ago) 5 5 4 4 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2084 1285604 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1816 1285872 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre fail_loc=0 at_max=600 PASS 44a (59s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 44b: race in target handle connect ========================================================== 04:17:57 (1713428277) 1 of 10 (1713428278) service : cur 5 worst 5 (at 1713426456, 1822s ago) 5 5 4 4 fail_loc=0x80000704 error: recover: Connection timed out Filesystem 1K-blocks Used Available Use% Mounted on 192.168.203.128@tcp:/lustre 7666232 3124 7210916 1% /mnt/lustre 2 of 10 (1713428318) service : cur 40 worst 40 (at 1713428319, 0s ago) 40 5 4 4 fail_loc=0x80000704 error: recover: Connection timed out Filesystem 1K-blocks Used Available Use% Mounted on 192.168.203.128@tcp:/lustre 7666232 3124 7210916 1% /mnt/lustre 3 of 10 (1713428339) service : cur 40 worst 40 (at 1713428319, 21s ago) 40 5 4 4 fail_loc=0x80000704 error: recover: Connection timed out Filesystem 1K-blocks Used Available Use% Mounted on 192.168.203.128@tcp:/lustre 7666232 3124 7210916 1% /mnt/lustre 4 of 10 (1713428360) service : cur 40 worst 40 (at 1713428319, 42s ago) 40 5 4 4 fail_loc=0x80000704 error: recover: Connection timed out Filesystem 1K-blocks Used Available Use% Mounted on 192.168.203.128@tcp:/lustre 7666232 3124 7210916 1% /mnt/lustre 5 of 10 (1713428380) service : cur 40 worst 40 (at 1713428319, 62s ago) 40 5 4 4 fail_loc=0x80000704 error: recover: Connection timed out Filesystem 1K-blocks Used Available Use% Mounted on 192.168.203.128@tcp:/lustre 7666232 3124 7210916 1% /mnt/lustre 6 of 10 (1713428401) service : cur 40 worst 40 (at 1713428319, 83s ago) 40 5 4 4 fail_loc=0x80000704 error: recover: Connection timed out Filesystem 1K-blocks Used Available Use% Mounted on 192.168.203.128@tcp:/lustre 7666232 3124 7210916 1% /mnt/lustre 7 of 10 (1713428421) service : cur 40 worst 40 (at 1713428319, 103s ago) 1 40 5 4 fail_loc=0x80000704 error: recover: Connection timed out Filesystem 1K-blocks Used Available Use% Mounted on 192.168.203.128@tcp:/lustre 7666232 3124 7210916 1% /mnt/lustre 8 of 10 (1713428442) service : cur 40 worst 40 (at 1713428319, 124s ago) 40 40 5 4 fail_loc=0x80000704 error: recover: Connection timed out Filesystem 1K-blocks Used Available Use% Mounted on 192.168.203.128@tcp:/lustre 7666232 3124 7210916 1% /mnt/lustre 9 of 10 (1713428463) service : cur 40 worst 40 (at 1713428319, 144s ago) 40 40 5 4 fail_loc=0x80000704 error: recover: Connection timed out Filesystem 1K-blocks Used Available Use% Mounted on 192.168.203.128@tcp:/lustre 7666232 3124 7210916 1% /mnt/lustre 10 of 10 (1713428483) service : cur 40 worst 40 (at 1713428319, 165s ago) 40 40 5 4 fail_loc=0x80000704 error: recover: Connection timed out Filesystem 1K-blocks Used Available Use% Mounted on 192.168.203.128@tcp:/lustre 7666232 3124 7210916 1% /mnt/lustre PASS 44b (229s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 44c: race in target handle connect ========================================================== 04:21:47 (1713428507) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2084 1285604 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1816 1285872 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre total: 100 create in 0.28 seconds: 352.11 ops/second fail_loc=0x80000712 Stopping /mnt/lustre-mds1 (opts:) on oleg328-server Failover mds1 to oleg328-server Starting mds1: -o localrecov -o abort_recovery /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg328-client: error: invalid path '/mnt/lustre': Input/output error pdsh@oleg328-client: oleg328-client: ssh exited with exit code 5 first stat failed: 5 unlink(/mnt/lustre/f44c.replay-single-0) error: No such file or directory total: 0 unlinks in 0 seconds: -nan unlinks/second Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:22:08 (1713428528) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:22:21 (1713428541) targets are mounted 04:22:21 (1713428541) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec unlink(/mnt/lustre/f44c.replay-single-0) error: No such file or directory total: 0 unlinks in 0 seconds: -nan unlinks/second PASS 44c (42s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 45: Handle failed close ============ 04:22:30 (1713428550) multiop /mnt/lustre/f45.replay-single vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.7515 /mnt/lustre/f45.replay-single has type file OK PASS 45 (3s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 46: Don't leak file handle after open resend (3325) ========================================================== 04:22:35 (1713428555) fail_loc=0x122 fail_loc=0 Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:22:53 (1713428573) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:23:06 (1713428586) targets are mounted 04:23:06 (1713428586) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec lfs path2fid: cannot get fid for 'f46.replay-single': No such file or directory PASS 46 (39s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 47: MDS->OSC failure during precreate cleanup (2824) ========================================================== 04:23:16 (1713428596) total: 20 open/close in 0.13 seconds: 152.80 ops/second Failing ost1 on oleg328-server Stopping /mnt/lustre-ost1 (opts:) on oleg328-server 04:23:18 (1713428598) shut down Failover ost1 to oleg328-server mount facets: ost1 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-OST0000 04:23:32 (1713428612) targets are mounted 04:23:32 (1713428612) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec fail_loc=0x80000204 total: 20 open/close in 0.09 seconds: 213.21 ops/second - unlinked 0 (time 1713428677 ; total 0 ; last 0) total: 20 unlinks in 0 seconds: inf unlinks/second PASS 47 (82s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 48: MDS->OSC failure during precreate cleanup (2824) ========================================================== 04:24:40 (1713428680) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2116 1285572 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1848 1285840 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre total: 20 open/close in 0.12 seconds: 170.44 ops/second Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:24:45 (1713428685) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:24:59 (1713428699) targets are mounted 04:24:59 (1713428699) facet_failover done fail_loc=0x80000216 total: 20 open/close in 0.08 seconds: 240.36 ops/second - unlinked 0 (time 1713428763 ; total 0 ; last 0) total: 40 unlinks in 0 seconds: inf unlinks/second PASS 48 (84s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 50: Double OSC recovery, don't LASSERT (3812) ========================================================== 04:26:06 (1713428766) PASS 50 (7s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 52: time out lock replay (3764) ==== 04:26:15 (1713428775) multiop /mnt/lustre/f52.replay-single vs_s TMPPIPE=/tmp/multiop_open_wait_pipe.7515 fail_loc=0x80000157 Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:26:16 (1713428776) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:26:29 (1713428789) targets are mounted 04:26:29 (1713428789) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec fail_loc=0x0 fail_loc=0x0 PASS 52 (80s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53a: |X| close request while two MDC requests in flight ========================================================== 04:27:37 (1713428857) fail_loc=0x80000115 fail_loc=0 Replay barrier on lustre-MDT0000 Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:27:43 (1713428863) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:27:57 (1713428877) targets are mounted 04:27:57 (1713428877) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d53a.replay-single-1/f has type file OK /mnt/lustre/d53a.replay-single-2/f has type file OK PASS 53a (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53b: |X| open request while two MDC requests in flight ========================================================== 04:28:07 (1713428887) multiop /mnt/lustre/d53b.replay-single-1/f vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.7515 fail_loc=0x80000107 fail_loc=0 Replay barrier on lustre-MDT0000 Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:28:12 (1713428892) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:28:26 (1713428906) targets are mounted 04:28:26 (1713428906) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d53b.replay-single-1/f has type file OK /mnt/lustre/d53b.replay-single-2/f has type file OK PASS 53b (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53c: |X| open request and close request while two MDC requests in flight ========================================================== 04:28:35 (1713428915) fail_loc=0x80000107 fail_loc=0x80000115 Replay barrier on lustre-MDT0000 Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:28:40 (1713428920) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:28:54 (1713428934) targets are mounted 04:28:54 (1713428934) facet_failover done fail_loc=0 /mnt/lustre/d53c.replay-single-1/f has type file OK /mnt/lustre/d53c.replay-single-2/f has type file OK PASS 53c (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53d: close reply while two MDC requests in flight ========================================================== 04:29:03 (1713428943) fail_loc=0x8000013b fail_loc=0 Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:29:06 (1713428946) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:29:19 (1713428959) targets are mounted 04:29:19 (1713428959) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d53d.replay-single-1/f has type file OK /mnt/lustre/d53d.replay-single-2/f has type file OK PASS 53d (24s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53e: |X| open reply while two MDC requests in flight ========================================================== 04:29:29 (1713428969) fail_loc=0x119 fail_loc=0 Replay barrier on lustre-MDT0000 Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:29:34 (1713428974) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:29:48 (1713428988) targets are mounted 04:29:48 (1713428988) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d53e.replay-single-1/f has type file OK /mnt/lustre/d53e.replay-single-2/f has type file OK PASS 53e (27s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53f: |X| open reply and close reply while two MDC requests in flight ========================================================== 04:29:57 (1713428997) fail_loc=0x119 fail_loc=0x8000013b Replay barrier on lustre-MDT0000 Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:30:02 (1713429002) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:30:16 (1713429016) targets are mounted 04:30:16 (1713429016) facet_failover done fail_loc=0 /mnt/lustre/d53f.replay-single-1/f has type file OK /mnt/lustre/d53f.replay-single-2/f has type file OK PASS 53f (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53g: |X| drop open reply and close request while close and open are both in flight ========================================================== 04:30:25 (1713429025) fail_loc=0x119 fail_loc=0x80000115 fail_loc=0 Replay barrier on lustre-MDT0000 Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:30:30 (1713429030) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:30:44 (1713429044) targets are mounted 04:30:44 (1713429044) facet_failover done /mnt/lustre/d53g.replay-single-1/f has type file OK /mnt/lustre/d53g.replay-single-2/f has type file OK PASS 53g (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53h: open request and close reply while two MDC requests in flight ========================================================== 04:30:52 (1713429052) fail_loc=0x80000107 fail_loc=0x8000013b Replay barrier on lustre-MDT0000 Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:30:58 (1713429058) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:31:12 (1713429072) targets are mounted 04:31:12 (1713429072) facet_failover done fail_loc=0 /mnt/lustre/d53h.replay-single-1/f has type file OK /mnt/lustre/d53h.replay-single-2/f has type file OK PASS 53h (29s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 55: let MDS_CHECK_RESENT return the original return code instead of 0 ========================================================== 04:31:23 (1713429083) fail_loc=0x8000012b fail_loc=0x0 touch: cannot touch '/mnt/lustre/f55.replay-single'rm: cannot remove '/mnt/lustre/f55.replay-single': No such file or directory : No such file or directory PASS 55 (62s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 56: don't replay a symlink open request (3440) ========================================================== 04:32:26 (1713429146) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2124 1285564 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1848 1285840 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:32:30 (1713429150) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:32:44 (1713429164) targets are mounted 04:32:44 (1713429164) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 56 (35s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 57: test recovery from llog for setattr op ========================================================== 04:33:03 (1713429183) fail_loc=0x8000012c UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2124 1285564 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1848 1285840 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:33:07 (1713429187) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:33:21 (1713429201) targets are mounted 04:33:21 (1713429201) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec affected facets: mds1 oleg328-server: oleg328-server.virtnet: executing _wait_recovery_complete *.lustre-MDT0000.recovery_status 1475 oleg328-server: *.lustre-MDT0000.recovery_status status: COMPLETE Waiting for orphan cleanup... osp.lustre-OST0000-osc-MDT0000.old_sync_processed osp.lustre-OST0000-osc-MDT0001.old_sync_processed osp.lustre-OST0001-osc-MDT0000.old_sync_processed osp.lustre-OST0001-osc-MDT0001.old_sync_processed wait 40 secs maximumly for oleg328-server mds-ost sync done. /mnt/lustre/f57.replay-single has type file OK fail_loc=0x0 PASS 57 (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 58a: test recovery from llog for setattr op (test llog_gen_rec) ========================================================== 04:33:33 (1713429213) fail_loc=0x8000012c total: 2500 open/close in 5.66 seconds: 441.44 ops/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2428 1285260 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1848 1285840 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:33:43 (1713429223) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:33:57 (1713429237) targets are mounted 04:33:57 (1713429237) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec fail_loc=0x0 - unlinked 0 (time 1713429251 ; total 0 ; last 0) total: 2500 unlinks in 3 seconds: 833.333313 unlinks/second PASS 58a (42s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 58b: test replay of setxattr op ==== 04:34:17 (1713429257) Starting client: oleg328-client.virtnet: -o user_xattr,flock oleg328-server@tcp:/lustre /mnt/lustre2 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2400 1285288 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1848 1285840 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:34:21 (1713429261) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:34:34 (1713429274) targets are mounted 04:34:34 (1713429274) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Stopping client oleg328-client.virtnet /mnt/lustre2 (opts:) oleg328-client.virtnet: executing wait_import_state_mount FULL mgc.*.mgs_server_uuid mgc.*.mgs_server_uuid in FULL state after 0 sec PASS 58b (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 58c: resend/reconstruct setxattr op ========================================================== 04:34:45 (1713429285) Starting client: oleg328-client.virtnet: -o user_xattr,flock oleg328-server@tcp:/lustre /mnt/lustre2 fail_val=0 fail_loc=0x123 fail_loc=0 fail_loc=0x119 fail_loc=0 Stopping client oleg328-client.virtnet /mnt/lustre2 (opts:) PASS 58c (128s) debug_raw_pointers=0 debug_raw_pointers=0 SKIP: replay-single test_59 skipping ALWAYS excluded test 59 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 60: test llog post recovery init vs llog unlink ========================================================== 04:36:55 (1713429415) total: 200 open/close in 0.54 seconds: 373.37 ops/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2260 1285428 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1848 1285840 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre - unlinked 0 (time 1713429419 ; total 0 ; last 0) total: 100 unlinks in 0 seconds: inf unlinks/second Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:37:00 (1713429420) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:37:14 (1713429434) targets are mounted 04:37:14 (1713429434) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec - unlinked 0 (time 1713429440 ; total 0 ; last 0) total: 100 unlinks in 0 seconds: inf unlinks/second PASS 60 (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 61a: test race llog recovery vs llog cleanup ========================================================== 04:37:23 (1713429443) total: 800 open/close in 1.78 seconds: 450.46 ops/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2344 1285344 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2028 1285660 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre - unlinked 0 (time 1713429448 ; total 0 ; last 0) total: 800 unlinks in 1 seconds: 800.000000 unlinks/second fail_val=0 fail_loc=0x80000221 Failing ost1 on oleg328-server Stopping /mnt/lustre-ost1 (opts:) on oleg328-server 04:37:31 (1713429451) shut down Failover ost1 to oleg328-server mount facets: ost1 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-OST0000 04:37:45 (1713429465) targets are mounted 04:37:45 (1713429465) facet_failover done Failing ost1 on oleg328-server Stopping /mnt/lustre-ost1 (opts:) on oleg328-server 04:37:56 (1713429476) shut down Failover ost1 to oleg328-server mount facets: ost1 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-OST0000 04:38:09 (1713429489) targets are mounted 04:38:09 (1713429489) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec fail_val=0 fail_loc=0x0 Can't lstat /mnt/lustre/d61a.replay-single/f61a.replay-single-*: No such file or directory PASS 61a (82s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 61b: test race mds llog sync vs llog cleanup ========================================================== 04:38:49 (1713429529) fail_loc=0x8000013a Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:38:53 (1713429533) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:39:06 (1713429546) targets are mounted 04:39:06 (1713429546) facet_failover done Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:39:18 (1713429558) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:39:31 (1713429571) targets are mounted 04:39:31 (1713429571) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec 1+0 records in 1+0 records out 4096 bytes (4.1 kB) copied, 0.00984523 s, 416 kB/s PASS 61b (50s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 61c: test race mds llog sync vs llog cleanup ========================================================== 04:39:41 (1713429581) fail_val=0 fail_loc=0x80000222 Failing ost1 on oleg328-server Stopping /mnt/lustre-ost1 (opts:) on oleg328-server 04:39:53 (1713429593) shut down Failover ost1 to oleg328-server mount facets: ost1 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-OST0000 04:40:07 (1713429607) targets are mounted 04:40:07 (1713429607) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec fail_val=0 fail_loc=0x0 PASS 61c (31s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 61d: error in llog_setup should cleanup the llog context correctly ========================================================== 04:40:14 (1713429614) Stopping /mnt/lustre-mds1 (opts:) on oleg328-server fail_loc=0x80000605 Starting mgs: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: mount.lustre: mount /dev/mapper/mds1_flakey at /mnt/lustre-mds1 failed: Operation not supported pdsh@oleg328-client: oleg328-server: ssh exited with exit code 95 Start of /dev/mapper/mds1_flakey on mgs failed 95 fail_loc=0 Starting mgs: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 PASS 61d (10s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 62: don't mis-drop resent replay === 04:40:25 (1713429625) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2332 1285356 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2004 1285684 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3605448 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3128 7210912 1% /mnt/lustre total: 25 open/close in 0.16 seconds: 153.92 ops/second fail_loc=0x80000707 Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:40:30 (1713429630) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:40:45 (1713429645) targets are mounted 04:40:45 (1713429645) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec fail_loc=0 - unlinked 0 (time 1713429710 ; total 0 ; last 0) total: 25 unlinks in 0 seconds: inf unlinks/second PASS 62 (87s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 65a: AT: verify early replies ====== 04:41:54 (1713429714) at_history=8 at_history=8 debug=other fail_val=11000 fail_loc=0x8000050a 00000100:00001000:3.0:1713429744.615464:0:14681:0:(client.c:537:ptlrpc_at_recv_early_reply()) @@@ Early reply #1, new deadline in 35s (25s) req@ffff8800a91d8700 x1796657825694912/t0(0) o101->lustre-MDT0000-mdc-ffff8800b589b800@192.168.203.128@tcp:12/10 lens 664/66320 e 1 to 0 dl 1713429779 ref 2 fl Rpc:PQr/200/ffffffff rc 0/-1 job:'createmany.0' uid:0 gid:0 portal 12 : cur 36 worst 40 (at 1713428318, 1432s ago) 36 40 40 40 portal 29 : cur 5 worst 5 (at 1713426730, 3020s ago) 5 0 0 0 portal 23 : cur 5 worst 5 (at 1713426730, 3020s ago) 5 5 5 5 portal 30 : cur 5 worst 5 (at 1713426735, 3015s ago) 5 0 0 0 portal 17 : cur 5 worst 5 (at 1713426815, 2935s ago) 5 5 5 0 portal 13 : cur 5 worst 5 (at 1713427215, 2535s ago) 5 5 0 0 portal 12 : cur 11 worst 40 (at 1713428318, 1441s ago) 36 40 40 40 portal 29 : cur 5 worst 5 (at 1713426730, 3029s ago) 5 0 0 0 portal 23 : cur 5 worst 5 (at 1713426730, 3029s ago) 5 5 5 5 portal 30 : cur 5 worst 5 (at 1713426735, 3024s ago) 5 0 0 0 portal 17 : cur 5 worst 5 (at 1713426815, 2944s ago) 5 5 5 0 portal 13 : cur 5 worst 5 (at 1713427215, 2544s ago) 5 5 0 0 PASS 65a (46s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 65b: AT: verify early replies on packed reply / bulk ========================================================== 04:42:42 (1713429762) at_history=8 at_history=8 debug=other trace fail_val=11 fail_loc=0x224 fail_loc=0 00000100:00001000:0.0:1713429792.023403:0:1958:0:(client.c:537:ptlrpc_at_recv_early_reply()) @@@ Early reply #1, new deadline in 35s (25s) req@ffff8801301a0700 x1796657825705472/t0(0) o4->lustre-OST0000-osc-ffff8800b589b800@192.168.203.128@tcp:6/4 lens 4584/448 e 1 to 0 dl 1713429827 ref 2 fl Rpc:Qr/200/ffffffff rc 0/-1 job:'multiop.0' uid:0 gid:0 debug=super ioctl neterror warning dlmtrace error emerg ha rpctrace vfstrace config console lfsck debug=super ioctl neterror warning dlmtrace error emerg ha rpctrace vfstrace config console lfsck portal 28 : cur 5 worst 5 (at 1713426730, 3067s ago) 5 5 5 5 portal 7 : cur 5 worst 5 (at 1713426731, 3066s ago) 5 5 5 5 portal 17 : cur 5 worst 5 (at 1713426786, 3011s ago) 5 0 0 5 portal 6 : cur 36 worst 36 (at 1713429797, 0s ago) 36 5 0 0 PASS 65b (37s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 66a: AT: verify MDT service time adjusts with no early replies ========================================================== 04:43:21 (1713429801) at_history=8 at_history=8 portal 12 : cur 5 worst 40 (at 1713428318, 1505s ago) 36 40 40 40 fail_val=5000 fail_loc=0x8000050a portal 12 : cur 5 worst 40 (at 1713428318, 1511s ago) 36 40 40 40 fail_val=10000 fail_loc=0x8000050a portal 12 : cur 36 worst 40 (at 1713428318, 1522s ago) 36 40 40 40 fail_loc=0 portal 12 : cur 5 worst 40 (at 1713428318, 1531s ago) 36 40 40 40 Current MDT timeout 5, worst 40 PASS 66a (50s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 66b: AT: verify net latency adjusts ========================================================== 04:44:12 (1713429852) at_history=8 at_history=8 fail_val=10 fail_loc=0x50c fail_loc=0 network timeout orig 5, cur 10, worst 10 PASS 66b (84s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 67a: AT: verify slow request processing doesn't induce reconnects ========================================================== 04:45:38 (1713429938) at_history=8 at_history=8 fail_val=400 fail_loc=0x50a fail_loc=0 0 osc reconnect attempts on gradual slow PASS 67a (72s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 67b: AT: verify instant slowdown doesn't induce reconnects ========================================================== 04:46:52 (1713430012) at_history=8 at_history=8 Creating to objid 5057 on ost lustre-OST0000... fail_val=20000 fail_loc=0x80000223 total: 17 open/close in 0.04 seconds: 401.01 ops/second Connected clients: oleg328-client.virtnet oleg328-client.virtnet service : cur 5 worst 5 (at 1713426465, 3571s ago) 1 1 1 0 phase 2 0 osc reconnect attempts on instant slow fail_loc=0x80000223 fail_loc=0 Connected clients: oleg328-client.virtnet oleg328-client.virtnet service : cur 5 worst 5 (at 1713426465, 3572s ago) 1 1 1 1 0 osc reconnect attempts on 2nd slow PASS 67b (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 68: AT: verify slowing locks ======= 04:47:20 (1713430040) at_history=8 at_history=8 /home/green/git/lustre-release/lustre/tests/replay-single.sh: line 1968: $ldlm_enqueue_min: ambiguous redirect fail_val=19 fail_loc=0x80000312 fail_val=25 fail_loc=0x80000312 fail_loc=0 /home/green/git/lustre-release/lustre/tests/replay-single.sh: line 1983: $ldlm_enqueue_min: ambiguous redirect PASS 68 (68s) debug_raw_pointers=0 debug_raw_pointers=0 Cleaning up AT ... debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 70a: check multi client t-f ======== 04:48:30 (1713430110) SKIP: replay-single test_70a Need two or more clients, have 1 SKIP 70a (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 70b: dbench 2mdts recovery; 1 clients ========================================================== 04:48:33 (1713430113) Starting client oleg328-client.virtnet: -o user_xattr,flock oleg328-server@tcp:/lustre /mnt/lustre Started clients oleg328-client.virtnet: 192.168.203.128@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) striped dir -i0 -c2 -H crush2 /mnt/lustre/d70b.replay-single + MISSING_DBENCH_OK= + PATH=/opt/iozone/bin:/opt/iozone/bin:/home/green/git/lustre-release/lustre/tests/mpi:/home/green/git/lustre-release/lustre/tests/racer:/home/green/git/lustre-release/lustre/../lustre-iokit/sgpdd-survey:/home/green/git/lustre-release/lustre/tests:/home/green/git/lustre-release/lustre/utils/gss:/home/green/git/lustre-release/lustre/utils:/opt/iozone/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin::/home/green/git/lustre-release/lustre/utils:/home/green/git/lustre-release/lustre/tests:/sbin:/usr/sbin:/home/green/git/lustre-release/lustre/utils:/home/green/git/lustre-release/lustre/tests/: + DBENCH_LIB= + TESTSUITE=replay-single + TESTNAME=test_70b + MOUNT=/mnt/lustre ++ hostname + DIR=/mnt/lustre/d70b.replay-single/oleg328-client.virtnet + LCTL=/home/green/git/lustre-release/lustre/utils/lctl + rundbench 1 -t 300 dbench: no process found Started rundbench load pid=21792 ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2460 1285228 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2160 1285528 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1976 3592960 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 26132 3556168 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 28108 7149128 1% /mnt/lustre test_70b fail mds1 1 times Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:48:41 (1713430121) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:48:55 (1713430135) targets are mounted 04:48:55 (1713430135) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2420 1285268 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2128 1285560 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1864 3592916 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 26396 3569800 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 28260 7162716 1% /mnt/lustre test_70b fail mds2 2 times Failing mds2 on oleg328-server Stopping /mnt/lustre-mds2 (opts:) on oleg328-server 04:49:07 (1713430147) shut down oleg328-client.virtnet: looking for dbench program oleg328-client.virtnet: /usr/bin/dbench oleg328-client.virtnet: creating output directory /mnt/lustre/d70b.replay-single/oleg328-client.virtnet oleg328-client.virtnet: mkdir: created directory '/mnt/lustre/d70b.replay-single/oleg328-client.virtnet' oleg328-client.virtnet: found dbench client file /usr/share/dbench/client.txt oleg328-client.virtnet: '/usr/share/dbench/client.txt' -> 'client.txt' oleg328-client.virtnet: running 'dbench 1 -t 300' on /mnt/lustre/d70b.replay-single/oleg328-client.virtnet at Thu Apr 18 04:48:34 EDT 2024 oleg328-client.virtnet: waiting for dbench pid 21816 oleg328-client.virtnet: dbench version 4.00 - Copyright Andrew Tridgell 1999-2004 oleg328-client.virtnet: oleg328-client.virtnet: Running for 300 seconds with load 'client.txt' and minimum warmup 60 secs oleg328-client.virtnet: failed to create barrier semaphore oleg328-client.virtnet: 0 of 1 processes prepared for launch 0 sec oleg328-client.virtnet: 1 of 1 processes prepared for launch 0 sec oleg328-client.virtnet: releasing clients oleg328-client.virtnet: 1 470 16.43 MB/sec warmup 1 sec latency 13.542 ms oleg328-client.virtnet: 1 959 10.91 MB/sec warmup 2 sec latency 11.621 ms oleg328-client.virtnet: 1 1558 8.70 MB/sec warmup 3 sec latency 12.201 ms oleg328-client.virtnet: 1 2054 6.65 MB/sec warmup 4 sec latency 11.289 ms oleg328-client.virtnet: 1 2500 5.74 MB/sec warmup 5 sec latency 18.758 ms oleg328-client.virtnet: 1 3210 5.71 MB/sec warmup 6 sec latency 15.543 ms oleg328-client.virtnet: 1 3767 5.68 MB/sec warmup 7 sec latency 21.307 ms oleg328-client.virtnet: 1 4017 5.01 MB/sec warmup 8 sec latency 19.353 ms oleg328-client.virtnet: 1 4083 4.45 MB/sec warmup 9 sec latency 784.158 ms oleg328-client.virtnet: 1 4083 4.01 MB/sec warmup 10 sec latency 1784.344 ms oleg328-client.virtnet: 1 4083 3.64 MB/sec warmup 11 sec latency 2784.731 ms oleg328-client.virtnet: 1 4083 3.34 MB/sec warmup 12 sec latency 3784.909 ms oleg328-client.virtnet: 1 4083 3.08 MB/sec warmup 13 sec latency 4785.098 ms oleg328-client.virtnet: 1 4083 2.86 MB/sec warmup 14 sec latency 5785.743 ms oleg328-client.virtnet: 1 4083 2.67 MB/sec warmup 15 sec latency 6785.903 ms oleg328-client.virtnet: 1 4083 2.51 MB/sec warmup 16 sec latency 7786.239 ms oleg328-client.virtnet: 1 4083 2.36 MB/sec warmup 17 sec latency 8786.821 ms oleg328-client.virtnet: 1 4083 2.23 MB/sec warmup 18 sec latency 9786.992 ms oleg328-client.virtnet: 1 4083 2.11 MB/sec warmup 19 sec latency 10787.189 ms oleg328-client.virtnet: 1 4083 2.00 MB/sec warmup 20 sec latency 11787.382 ms oleg328-client.virtnet: 1 4083 1.91 MB/sec warmup 21 sec latency 12787.553 ms oleg328-client.virtnet: 1 4083 1.82 MB/sec warmup 22 sec latency 13788.345 ms oleg328-client.virtnet: 1 4083 1.74 MB/sec warmup 23 sec latency 14788.624 ms oleg328-client.virtnet: 1 4083 1.67 MB/sec warmup 24 sec latency 15788.797 ms oleg328-client.virtnet: 1 4174 1.60 MB/sec warmup 25 sec latency 16665.289 ms oleg328-client.virtnet: 1 4631 1.62 MB/sec warmup 26 sec latency 12.326 ms oleg328-client.virtnet: 1 5246 1.68 MB/sec warmup 27 sec latency 11.650 ms oleg328-client.virtnet: 1 5804 1.64 MB/sec warmup 28 sec latency 10.571 ms oleg328-client.virtnet: 1 6449 1.79 MB/sec warmup 29 sec latency 12.834 ms oleg328-client.virtnet: 1 7230 1.93 MB/sec warmup 30 sec latency 14.023 ms oleg328-client.virtnet: 1 7665 1.91 MB/sec warmup 31 sec latency 13.594 ms oleg328-client.virtnet: 1 8063 1.88 MB/sec warmup 32 sec latency 20.651 ms oleg328-client.virtnet: 1 8227 1.85 MB/sec warmup 33 sec latency 645.786 msFailover mds2 to oleg328-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0001 04:49:21 (1713430161) targets are mounted 04:49:21 (1713430161) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2380 1285308 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2052 1285636 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 13100 3593044 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 36384 3569972 2% /mnt/lustre[OST:1] filesystem_summary: 7666232 49484 7163016 1% /mnt/lustre test_70b fail mds1 3 times Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:49:33 (1713430173) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:49:47 (1713430187) targets are mounted 04:49:47 (1713430187) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec oleg328-client.virtnet: 1 8227 1.80 MB/sec warmup 34 sec latency 1646.507 ms oleg328-client.virtnet: 1 8227 1.75 MB/sec warmup 35 sec latency 2646.690 ms oleg328-client.virtnet: 1 8227 1.70 MB/sec warmup 36 sec latency 3646.942 ms oleg328-client.virtnet: 1 8227 1.65 MB/sec warmup 37 sec latency 4647.094 ms oleg328-client.virtnet: 1 8227 1.61 MB/sec warmup 38 sec latency 5647.449 ms oleg328-client.virtnet: 1 8227 1.57 MB/sec warmup 39 sec latency 6647.705 ms oleg328-client.virtnet: 1 8227 1.53 MB/sec warmup 40 sec latency 7647.850 ms oleg328-client.virtnet: 1 8227 1.49 MB/sec warmup 41 sec latency 8648.004 ms oleg328-client.virtnet: 1 8227 1.46 MB/sec warmup 42 sec latency 9648.197 ms oleg328-client.virtnet: 1 8227 1.42 MB/sec warmup 43 sec latency 10648.400 ms oleg328-client.virtnet: 1 8227 1.39 MB/sec warmup 44 sec latency 11648.645 ms oleg328-client.virtnet: 1 8227 1.36 MB/sec warmup 45 sec latency 12648.787 ms oleg328-client.virtnet: 1 8227 1.33 MB/sec warmup 46 sec latency 13648.898 ms oleg328-client.virtnet: 1 8227 1.30 MB/sec warmup 47 sec latency 14649.275 ms oleg328-client.virtnet: 1 8227 1.27 MB/sec warmup 48 sec latency 15649.439 ms oleg328-client.virtnet: 1 8227 1.25 MB/sec warmup 49 sec latency 16649.641 ms oleg328-client.virtnet: 1 8227 1.22 MB/sec warmup 50 sec latency 17649.787 ms oleg328-client.virtnet: 1 8227 1.20 MB/sec warmup 51 sec latency 18649.933 ms oleg328-client.virtnet: 1 8568 1.24 MB/sec warmup 52 sec latency 19029.294 ms oleg328-client.virtnet: 1 9093 1.22 MB/sec warmup 53 sec latency 11.249 ms oleg328-client.virtnet: 1 9689 1.24 MB/sec warmup 54 sec latency 10.473 ms oleg328-client.virtnet: 1 10676 1.39 MB/sec warmup 55 sec latency 12.304 ms oleg328-client.virtnet: 1 11119 1.40 MB/sec warmup 56 sec latency 14.229 ms oleg328-client.virtnet: 1 11572 1.39 MB/sec warmup 57 sec latency 13.089 ms oleg328-client.virtnet: 1 12179 1.44 MB/sec warmup 58 sec latency 9.942 ms oleg328-client.virtnet: 1 12772 1.42 MB/sec warmup 59 sec latency 11.284 ms oleg328-client.virtnet: 1 14258 6.63 MB/sec execute 1 sec latency 14.652 ms oleg328-client.virtnet: 1 14606 4.05 MB/sec execute 2 sec latency 16.349 ms oleg328-client.virtnet: 1 15010 2.88 MB/sec execute 3 sec latency 13.723 ms oleg328-client.virtnet: 1 15455 2.56 MB/sec execute 4 sec latency 13.243 ms oleg328-client.virtnet: 1 16011 2.70 MB/sec execute 5 sec latency 13.775 ms oleg328-client.virtnet: 1 16649 2.54 MB/sec execute 6 sec latency 9.128 ms oleg328-client.virtnet: 1 17750 3.65 MB/sec execute 7 sec latency 8.713 ms oleg328-client.virtnet: 1 18209 3.40 MB/sec execute 8 sec latency 10.263 ms oleg328-client.virtnet: 1 18630 3.10 MB/sec execute 9 sec latency 14.956 ms oleg328-client.virtnet: 1 19232 3.23 MB/sec execute 10 sec latency 11.591 ms oleg328-client.virtnet: 1 19633 2.97 MB/sec execute 11 sec latency 15.346 ms oleg328-client.virtnet: 1 20169 2.86 MB/sec execute 12 sec latency 11.167 ms oleg328-client.virtnet: 1 20989 3.12 MB/sec execute 13 sec latency 14.334 ms oleg328-client.virtnet: 1 21566 3.29 MB/sec execute 14 sec latency 12.526 ms oleg328-client.virtnet: 1 21878 3.09 MB/sec execute 15 sec latency 13.415 ms oleg328-client.virtnet: 1 22304 2.95 MB/sec execute 16 sec latency 20.766 ms oleg328-client.virtnet: 1 22828 3.03 MB/sec execute 17 sec latency 14.180 ms oleg328-client.virtnet: 1 23190 2.88 MB/sec execute 18 sec latency 19.299 ms oleg328-client.virtnet: 1 23755 2.82 MB/sec execute 19 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2380 1285308 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2084 1285604 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 12996 3592948 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 36660 3568944 2% /mnt/lustre[OST:1] filesystem_summary: 7666232 49656 7161892 1% /mnt/lustre test_70b fail mds2 4 times Failing mds2 on oleg328-server Stopping /mnt/lustre-mds2 (opts:) on oleg328-server 04:49:59 (1713430199) shut down Failover mds2 to oleg328-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0001 04:50:13 (1713430213) targets are mounted 04:50:13 (1713430213) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2380 1285308 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2052 1285636 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 13032 3591148 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 36660 3569532 2% /mnt/lustre[OST:1] filesystem_summary: 7666232 49692 7160680 1% /mnt/lustre test_70b fail mds1 5 times Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:50:25 (1713430225) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 latency 11.453 ms oleg328-client.virtnet: 1 24684 3.04 MB/sec execute 20 sec latency 10.569 ms oleg328-client.virtnet: 1 25235 3.12 MB/sec execute 21 sec latency 15.151 ms oleg328-client.virtnet: 1 25607 2.99 MB/sec execute 22 sec latency 91.589 ms oleg328-client.virtnet: 1 25872 2.89 MB/sec execute 23 sec latency 442.948 ms oleg328-client.virtnet: 1 26449 2.95 MB/sec execute 24 sec latency 13.838 ms oleg328-client.virtnet: 1 26730 2.84 MB/sec execute 25 sec latency 447.429 ms oleg328-client.virtnet: 1 26730 2.73 MB/sec execute 26 sec latency 1447.600 ms oleg328-client.virtnet: 1 26730 2.63 MB/sec execute 27 sec latency 2447.787 ms oleg328-client.virtnet: 1 26730 2.53 MB/sec execute 28 sec latency 3448.036 ms oleg328-client.virtnet: 1 26730 2.45 MB/sec execute 29 sec latency 4448.208 ms oleg328-client.virtnet: 1 26730 2.36 MB/sec execute 30 sec latency 5448.411 ms oleg328-client.virtnet: 1 26730 2.29 MB/sec execute 31 sec latency 6448.565 ms oleg328-client.virtnet: 1 26730 2.22 MB/sec execute 32 sec latency 7448.691 ms oleg328-client.virtnet: 1 26730 2.15 MB/sec execute 33 sec latency 8449.562 ms oleg328-client.virtnet: 1 26730 2.09 MB/sec execute 34 sec latency 9449.804 ms oleg328-client.virtnet: 1 26730 2.03 MB/sec execute 35 sec latency 10449.930 ms oleg328-client.virtnet: 1 26730 1.97 MB/sec execute 36 sec latency 11450.098 ms oleg328-client.virtnet: 1 26730 1.92 MB/sec execute 37 sec latency 12450.250 ms oleg328-client.virtnet: 1 26730 1.87 MB/sec execute 38 sec latency 13450.462 ms oleg328-client.virtnet: 1 26730 1.82 MB/sec execute 39 sec latency 14450.590 ms oleg328-client.virtnet: 1 26730 1.77 MB/sec execute 40 sec latency 15450.692 ms oleg328-client.virtnet: 1 26730 1.73 MB/sec execute 41 sec latency 16450.871 ms oleg328-client.virtnet: 1 26730 1.69 MB/sec execute 42 sec latency 17451.020 ms oleg328-client.virtnet: 1 26738 1.65 MB/sec execute 43 sec latency 18424.007 ms oleg328-client.virtnet: 1 27343 1.66 MB/sec execute 44 sec latency 12.088 ms oleg328-client.virtnet: 1 28375 1.84 MB/sec execute 45 sec latency 11.048 ms oleg328-client.virtnet: 1 28834 1.84 MB/sec execute 46 sec latency 10.336 ms oleg328-client.virtnet: 1 29336 1.82 MB/sec execute 47 sec latency 11.321 ms oleg328-client.virtnet: 1 30003 1.87 MB/sec execute 48 sec latency 9.293 ms oleg328-client.virtnet: 1 30684 1.85 MB/sec execute 49 sec latency 8.134 ms oleg328-client.virtnet: 1 31604 1.96 MB/sec execute 50 sec latency 10.099 ms oleg328-client.virtnet: 1 32311 2.03 MB/sec execute 51 sec latency 26.282 ms oleg328-client.virtnet: 1 32885 2.01 MB/sec execute 52 sec latency 10.593 ms oleg328-client.virtnet: 1 33602 2.06 MB/sec execute 53 sec latency 9.224 ms oleg328-client.virtnet: 1 34361 2.05 MB/sec execute 54 sec latency 8.861 ms oleg328-client.virtnet: 1 35340 2.15 MB/sec execute 55 sec latency 9.918 ms oleg328-client.virtnet: 1 35985 2.19 MB/sec execute 56 sec latency 9.129 ms oleg328-client.virtnet: 1 36503 2.17 MB/sec execute 57 sec latency 13.581 ms oleg328-client.virtnet: 1 37190 2.21 MB/sec execute 58 sec latency 11.358 ms oleg328-client.virtnet: 1 37858 2.20 MB/sec execute 59 sec latency 9.723 ms oleg328-client.virtnet: 1 38782 2.29 MB/sec execute 60 sec latency 10.153 ms oleg328-client.virtnet: 1 39396 2.33 MB/sec execute 61 sec latency 10.571 ms oleg328-client.virtnet: 1 39793 2.30 MB/sec execute 62 sec latency 14.023 ms oleg328-client.virtnet: 1 40138 2.28 MB/sec execute 63 sec latency 23.845 ms oleg328-client.virtnet: 1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:50:39 (1713430239) targets are mounted 04:50:39 (1713430239) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2380 1285308 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2092 1285596 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 13100 3592892 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 36384 3570124 2% /mnt/lustre[OST:1] filesystem_summary: 7666232 49484 7163016 1% /mnt/lustre test_70b fail mds2 6 times Failing mds2 on oleg328-server Stopping /mnt/lustre-mds2 (opts:) on oleg328-server 04:50:51 (1713430251) shut down Failover mds2 to oleg328-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0001 04:51:05 (1713430265) targets are mounted 04:51:05 (1713430265) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2380 1285308 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2052 1285636 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 13184 3592148 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 36536 3569976 2% /mnt/lustre[OST:1] filesystem_summary: 7666232 49720 7162124 1% /mnt/lustre test_70b fail mds1 7 times Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:51:17 (1713430277) shut down 40644 2.30 MB/sec execute 64 sec latency 14.363 ms oleg328-client.virtnet: 1 41047 2.27 MB/sec execute 65 sec latency 16.093 ms oleg328-client.virtnet: 1 41487 2.26 MB/sec execute 66 sec latency 12.869 ms oleg328-client.virtnet: 1 42375 2.33 MB/sec execute 67 sec latency 13.130 ms oleg328-client.virtnet: 1 42922 2.36 MB/sec execute 68 sec latency 11.119 ms oleg328-client.virtnet: 1 43332 2.34 MB/sec execute 69 sec latency 16.272 ms oleg328-client.virtnet: 1 43739 2.33 MB/sec execute 70 sec latency 14.295 ms oleg328-client.virtnet: 1 44218 2.34 MB/sec execute 71 sec latency 11.616 ms oleg328-client.virtnet: 1 44680 2.32 MB/sec execute 72 sec latency 12.283 ms oleg328-client.virtnet: 1 45193 2.32 MB/sec execute 73 sec latency 16.516 ms oleg328-client.virtnet: 1 45805 2.36 MB/sec execute 74 sec latency 444.823 ms oleg328-client.virtnet: 1 46428 2.40 MB/sec execute 75 sec latency 10.149 ms oleg328-client.virtnet: 1 46856 2.37 MB/sec execute 76 sec latency 15.671 ms oleg328-client.virtnet: 1 46862 2.34 MB/sec execute 77 sec latency 984.997 ms oleg328-client.virtnet: 1 46862 2.31 MB/sec execute 78 sec latency 1985.153 ms oleg328-client.virtnet: 1 46862 2.28 MB/sec execute 79 sec latency 2985.426 ms oleg328-client.virtnet: 1 46862 2.26 MB/sec execute 80 sec latency 3985.670 ms oleg328-client.virtnet: 1 46862 2.23 MB/sec execute 81 sec latency 4986.174 ms oleg328-client.virtnet: 1 46862 2.20 MB/sec execute 82 sec latency 5986.338 ms oleg328-client.virtnet: 1 46862 2.17 MB/sec execute 83 sec latency 6986.664 ms oleg328-client.virtnet: 1 46862 2.15 MB/sec execute 84 sec latency 7986.826 ms oleg328-client.virtnet: 1 46862 2.12 MB/sec execute 85 sec latency 8987.193 ms oleg328-client.virtnet: 1 46862 2.10 MB/sec execute 86 sec latency 9987.383 ms oleg328-client.virtnet: 1 46862 2.07 MB/sec execute 87 sec latency 10987.528 ms oleg328-client.virtnet: 1 46862 2.05 MB/sec execute 88 sec latency 11987.692 ms oleg328-client.virtnet: 1 46862 2.03 MB/sec execute 89 sec latency 12987.905 ms oleg328-client.virtnet: 1 46862 2.01 MB/sec execute 90 sec latency 13988.399 ms oleg328-client.virtnet: 1 46862 1.98 MB/sec execute 91 sec latency 14988.593 ms oleg328-client.virtnet: 1 46862 1.96 MB/sec execute 92 sec latency 15988.998 ms oleg328-client.virtnet: 1 46862 1.94 MB/sec execute 93 sec latency 16989.177 ms oleg328-client.virtnet: 1 46862 1.92 MB/sec execute 94 sec latency 17989.354 ms oleg328-client.virtnet: 1 47060 1.90 MB/sec execute 95 sec latency 18530.621 ms oleg328-client.virtnet: 1 47599 1.93 MB/sec execute 96 sec latency 11.007 ms oleg328-client.virtnet: 1 48131 1.91 MB/sec execute 97 sec latency 14.649 ms oleg328-client.virtnet: 1 48711 1.92 MB/sec execute 98 sec latency 14.577 ms oleg328-client.virtnet: 1 49646 2.00 MB/sec execute 99 sec latency 12.738 ms oleg328-client.virtnet: 1 50003 1.99 MB/sec execute 100 sec latency 19.162 ms oleg328-client.virtnet: 1 50408 1.98 MB/sec execute 101 sec latency 14.673 ms oleg328-client.virtnet: 1 50876 1.97 MB/sec execute 102 sec latency 11.872 ms oleg328-client.virtnet: 1 51380 1.99 MB/sec execute 103 sec latency 15.694 ms oleg328-client.virtnet: 1 51842 1.97 MB/sec execute 104 sec latency 17.513 ms oleg328-client.virtnet: 1 52287 1.97 MB/sec execute 105 sec latency 18.812 ms oleg328-client.virtnet: 1 53180 2.04 MB/sec execute 106 sec latency 17.072 ms oleg328-client.virtnet: 1 53538 2.04 MB/sec execute 107 sec latency 16.978 ms oleg328-client.virtnet: 1 53958 2.02 MB/sec execute 108 sec latenFailover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:51:31 (1713430291) targets are mounted 04:51:31 (1713430291) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2380 1285308 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2084 1285604 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 12948 3593100 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 36616 3569724 2% /mnt/lustre[OST:1] filesystem_summary: 7666232 49564 7162824 1% /mnt/lustre test_70b fail mds2 8 times Failing mds2 on oleg328-server Stopping /mnt/lustre-mds2 (opts:) on oleg328-server 04:51:43 (1713430303) shut down Failover mds2 to oleg328-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0001 04:51:56 (1713430316) targets are mounted 04:51:56 (1713430316) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2380 1285308 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2052 1285636 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 13028 3593060 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 36696 3568940 2% /mnt/lustre[OST:1] filesystem_summary: 7666232 49724 7162000 1% /mnt/lustre test_70b fail mds1 9 times cy 14.666 ms oleg328-client.virtnet: 1 54417 2.02 MB/sec execute 109 sec latency 10.794 ms oleg328-client.virtnet: 1 54963 2.03 MB/sec execute 110 sec latency 12.418 ms oleg328-client.virtnet: 1 55580 2.03 MB/sec execute 111 sec latency 10.015 ms oleg328-client.virtnet: 1 56444 2.07 MB/sec execute 112 sec latency 9.927 ms oleg328-client.virtnet: 1 57121 2.10 MB/sec execute 113 sec latency 12.775 ms oleg328-client.virtnet: 1 57640 2.09 MB/sec execute 114 sec latency 11.044 ms oleg328-client.virtnet: 1 58183 2.11 MB/sec execute 115 sec latency 13.239 ms oleg328-client.virtnet: 1 58616 2.09 MB/sec execute 116 sec latency 13.728 ms oleg328-client.virtnet: 1 59170 2.09 MB/sec execute 117 sec latency 10.677 ms oleg328-client.virtnet: 1 60081 2.14 MB/sec execute 118 sec latency 10.647 ms oleg328-client.virtnet: 1 60665 2.16 MB/sec execute 119 sec latency 13.911 ms oleg328-client.virtnet: 1 61061 2.14 MB/sec execute 120 sec latency 12.833 ms oleg328-client.virtnet: 1 61509 2.14 MB/sec execute 121 sec latency 15.021 ms oleg328-client.virtnet: 1 62024 2.15 MB/sec execute 122 sec latency 15.316 ms oleg328-client.virtnet: 1 62508 2.13 MB/sec execute 123 sec latency 16.180 ms oleg328-client.virtnet: 1 63106 2.16 MB/sec execute 124 sec latency 13.056 ms oleg328-client.virtnet: 1 63960 2.19 MB/sec execute 125 sec latency 10.725 ms oleg328-client.virtnet: 1 64306 2.19 MB/sec execute 126 sec latency 15.177 ms oleg328-client.virtnet: 1 64728 2.18 MB/sec execute 127 sec latency 16.403 ms oleg328-client.virtnet: 1 65139 2.17 MB/sec execute 128 sec latency 19.244 ms oleg328-client.virtnet: 1 65139 2.15 MB/sec execute 129 sec latency 1002.557 ms oleg328-client.virtnet: 1 65139 2.14 MB/sec execute 130 sec latency 2002.842 ms oleg328-client.virtnet: 1 65139 2.12 MB/sec execute 131 sec latency 3003.088 ms oleg328-client.virtnet: 1 65139 2.11 MB/sec execute 132 sec latency 4003.254 ms oleg328-client.virtnet: 1 65139 2.09 MB/sec execute 133 sec latency 5003.525 ms oleg328-client.virtnet: 1 65139 2.07 MB/sec execute 134 sec latency 6003.721 ms oleg328-client.virtnet: 1 65139 2.06 MB/sec execute 135 sec latency 7003.881 ms oleg328-client.virtnet: 1 65139 2.04 MB/sec execute 136 sec latency 8004.048 ms oleg328-client.virtnet: 1 65139 2.03 MB/sec execute 137 sec latency 9004.202 ms oleg328-client.virtnet: 1 65139 2.01 MB/sec execute 138 sec latency 10004.357 ms oleg328-client.virtnet: 1 65139 2.00 MB/sec execute 139 sec latency 11004.501 ms oleg328-client.virtnet: 1 65139 1.99 MB/sec execute 140 sec latency 12004.676 ms oleg328-client.virtnet: 1 65139 1.97 MB/sec execute 141 sec latency 13005.089 ms oleg328-client.virtnet: 1 65139 1.96 MB/sec execute 142 sec latency 14005.444 ms oleg328-client.virtnet: 1 65139 1.94 MB/sec execute 143 sec latency 15005.618 ms oleg328-client.virtnet: 1 65139 1.93 MB/sec execute 144 sec latency 16005.784 ms oleg328-client.virtnet: 1 65139 1.92 MB/sec execute 145 sec latency 17005.965 ms oleg328-client.virtnet: 1 65139 1.90 MB/sec execute 146 sec latency 18006.386 ms oleg328-client.virtnet: 1 65464 1.91 MB/sec execute 147 sec latency 18522.260 ms oleg328-client.virtnet: 1 66071 1.90 MB/sec execute 148 sec latency 11.211 ms oleg328-client.virtnet: 1 66864 1.94 MB/sec execute 149 sec latency 10.100 ms oleg328-client.virtnet: 1 67697 1.96 MB/sec execute 150 sec latency 11.260 ms oleg328-client.virtnet: 1 68175 1.96 MB/sec execute 151 sec latency 18.998 ms oleg328-client.virtnet: 1 68617 1.95 MB/sec execute 152 sec latency 14.974 ms oleg328-client.virtnet: 1 Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:52:08 (1713430328) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:52:22 (1713430342) targets are mounted 04:52:22 (1713430342) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2380 1285308 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2092 1285596 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 12940 3593116 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 36592 3569668 2% /mnt/lustre[OST:1] filesystem_summary: 7666232 49532 7162784 1% /mnt/lustre test_70b fail mds2 10 times Failing mds2 on oleg328-server Stopping /mnt/lustre-mds2 (opts:) on oleg328-server 04:52:34 (1713430354) shut down Failover mds2 to oleg328-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0001 04:52:48 (1713430368) targets are mounted 04:52:48 (1713430368) facet_failover done 69197 1.96 MB/sec execute 153 sec latency 11.674 ms oleg328-client.virtnet: 1 69809 1.96 MB/sec execute 154 sec latency 9.155 ms oleg328-client.virtnet: 1 70627 1.99 MB/sec execute 155 sec latency 12.298 ms oleg328-client.virtnet: 1 71304 2.01 MB/sec execute 156 sec latency 10.273 ms oleg328-client.virtnet: 1 71798 2.00 MB/sec execute 157 sec latency 14.748 ms oleg328-client.virtnet: 1 72451 2.02 MB/sec execute 158 sec latency 11.184 ms oleg328-client.virtnet: 1 72970 2.01 MB/sec execute 159 sec latency 10.842 ms oleg328-client.virtnet: 1 73555 2.01 MB/sec execute 160 sec latency 10.432 ms oleg328-client.virtnet: 1 74573 2.06 MB/sec execute 161 sec latency 9.409 ms oleg328-client.virtnet: 1 75057 2.06 MB/sec execute 162 sec latency 12.004 ms oleg328-client.virtnet: 1 75664 2.06 MB/sec execute 163 sec latency 12.042 ms oleg328-client.virtnet: 1 76297 2.06 MB/sec execute 164 sec latency 10.282 ms oleg328-client.virtnet: 1 76947 2.06 MB/sec execute 165 sec latency 10.290 ms oleg328-client.virtnet: 1 77893 2.09 MB/sec execute 166 sec latency 10.969 ms oleg328-client.virtnet: 1 78570 2.11 MB/sec execute 167 sec latency 8.099 ms oleg328-client.virtnet: 1 79212 2.11 MB/sec execute 168 sec latency 9.902 ms oleg328-client.virtnet: 1 79948 2.12 MB/sec execute 169 sec latency 7.988 ms oleg328-client.virtnet: 1 80802 2.14 MB/sec execute 170 sec latency 9.839 ms oleg328-client.virtnet: 1 81859 2.17 MB/sec execute 171 sec latency 8.813 ms oleg328-client.virtnet: 1 82430 2.16 MB/sec execute 172 sec latency 10.138 ms oleg328-client.virtnet: 1 83123 2.18 MB/sec execute 173 sec latency 12.614 ms oleg328-client.virtnet: 1 83973 2.17 MB/sec execute 174 sec latency 13.263 ms oleg328-client.virtnet: 1 85169 2.22 MB/sec execute 175 sec latency 9.652 ms oleg328-client.virtnet: 1 85680 2.22 MB/sec execute 176 sec latency 9.747 ms oleg328-client.virtnet: 1 86060 2.21 MB/sec execute 177 sec latency 334.442 ms oleg328-client.virtnet: 1 86669 2.22 MB/sec execute 178 sec latency 416.056 ms oleg328-client.virtnet: 1 87274 2.21 MB/sec execute 179 sec latency 10.895 ms oleg328-client.virtnet: 1 87528 2.21 MB/sec execute 180 sec latency 708.225 ms oleg328-client.virtnet: 1 87528 2.20 MB/sec execute 181 sec latency 1708.397 ms oleg328-client.virtnet: 1 87528 2.18 MB/sec execute 182 sec latency 2708.566 ms oleg328-client.virtnet: 1 87528 2.17 MB/sec execute 183 sec latency 3708.729 ms oleg328-client.virtnet: 1 87528 2.16 MB/sec execute 184 sec latency 4708.930 ms oleg328-client.virtnet: 1 87528 2.15 MB/sec execute 185 sec latency 5709.066 ms oleg328-client.virtnet: 1 87528 2.14 MB/sec execute 186 sec latency 6709.263 ms oleg328-client.virtnet: 1 87528 2.13 MB/sec execute 187 sec latency 7709.437 ms oleg328-client.virtnet: 1 87528 2.11 MB/sec execute 188 sec latency 8709.608 ms oleg328-client.virtnet: 1 87528 2.10 MB/sec execute 189 sec latency 9709.825 ms oleg328-client.virtnet: 1 87528 2.09 MB/sec execute 190 sec latency 10709.966 ms oleg328-client.virtnet: 1 87528 2.08 MB/sec execute 191 sec latency 11710.094 ms oleg328-client.virtnet: 1 87528 2.07 MB/sec execute 192 sec latency 12710.271 ms oleg328-client.virtnet: 1 87528 2.06 MB/sec execute 193 sec latency 13710.508 ms oleg328-client.virtnet: 1 87528 2.05 MB/sec execute 194 sec latency 14710.728 ms oleg328-client.virtnet: 1 87528 2.04 MB/sec execute 195 sec latency 15710.914 ms oleg328-client.virtnet: 1 87528 2.03 MB/sec execute 196 sec latency 16711.120 ms oleg328-client.virtnet: 1 87528 2.02 MB/sec execute 197 sec latency 17711.400 oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2380 1285308 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2060 1285628 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 12940 3593304 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 36612 3570016 2% /mnt/lustre[OST:1] filesystem_summary: 7666232 49552 7163320 1% /mnt/lustre test_70b fail mds1 11 times Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:53:00 (1713430380) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:53:13 (1713430393) targets are mounted 04:53:13 (1713430393) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2380 1285308 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2092 1285596 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 12944 3593188 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 36772 3569596 2% /mnt/lustre[OST:1] filesystem_summary: 7666232 49716 7162784 1% /mnt/lustre test_70b fail mds2 12 times Failing mds2 on oleg328-server Stopping /mnt/lustre-mds2 (opts:) on oleg328-server 04:53:25 (1713430405) shut down Failover mds2 to oleg328-server mount facets: mds2 ms oleg328-client.virtnet: 1 87772 2.01 MB/sec execute 198 sec latency 18308.490 ms oleg328-client.virtnet: 1 88830 2.05 MB/sec execute 199 sec latency 13.336 ms oleg328-client.virtnet: 1 89334 2.05 MB/sec execute 200 sec latency 9.679 ms oleg328-client.virtnet: 1 89881 2.05 MB/sec execute 201 sec latency 12.671 ms oleg328-client.virtnet: 1 90525 2.05 MB/sec execute 202 sec latency 10.215 ms oleg328-client.virtnet: 1 91182 2.05 MB/sec execute 203 sec latency 9.586 ms oleg328-client.virtnet: 1 92338 2.09 MB/sec execute 204 sec latency 8.068 ms oleg328-client.virtnet: 1 92787 2.09 MB/sec execute 205 sec latency 10.737 ms oleg328-client.virtnet: 1 93433 2.09 MB/sec execute 206 sec latency 9.674 ms oleg328-client.virtnet: 1 94096 2.10 MB/sec execute 207 sec latency 9.949 ms oleg328-client.virtnet: 1 94861 2.10 MB/sec execute 208 sec latency 7.958 ms oleg328-client.virtnet: 1 95974 2.14 MB/sec execute 209 sec latency 7.654 ms oleg328-client.virtnet: 1 96541 2.13 MB/sec execute 210 sec latency 8.572 ms oleg328-client.virtnet: 1 97216 2.15 MB/sec execute 211 sec latency 8.981 ms oleg328-client.virtnet: 1 97879 2.14 MB/sec execute 212 sec latency 8.469 ms oleg328-client.virtnet: 1 98812 2.16 MB/sec execute 213 sec latency 8.387 ms oleg328-client.virtnet: 1 99690 2.18 MB/sec execute 214 sec latency 8.950 ms oleg328-client.virtnet: 1 100282 2.18 MB/sec execute 215 sec latency 11.351 ms oleg328-client.virtnet: 1 100982 2.19 MB/sec execute 216 sec latency 8.438 ms oleg328-client.virtnet: 1 101726 2.18 MB/sec execute 217 sec latency 9.566 ms oleg328-client.virtnet: 1 102873 2.22 MB/sec execute 218 sec latency 7.937 ms oleg328-client.virtnet: 1 103418 2.22 MB/sec execute 219 sec latency 8.602 ms oleg328-client.virtnet: 1 104097 2.22 MB/sec execute 220 sec latency 12.409 ms oleg328-client.virtnet: 1 104793 2.22 MB/sec execute 221 sec latency 8.296 ms oleg328-client.virtnet: 1 105676 2.24 MB/sec execute 222 sec latency 9.041 ms oleg328-client.virtnet: 1 106687 2.26 MB/sec execute 223 sec latency 9.058 ms oleg328-client.virtnet: 1 107238 2.26 MB/sec execute 224 sec latency 11.123 ms oleg328-client.virtnet: 1 107938 2.27 MB/sec execute 225 sec latency 9.088 ms oleg328-client.virtnet: 1 108621 2.26 MB/sec execute 226 sec latency 10.145 ms oleg328-client.virtnet: 1 109659 2.28 MB/sec execute 227 sec latency 9.126 ms oleg328-client.virtnet: 1 110400 2.30 MB/sec execute 228 sec latency 9.122 ms oleg328-client.virtnet: 1 110979 2.29 MB/sec execute 229 sec latency 11.489 ms oleg328-client.virtnet: 1 111650 2.30 MB/sec execute 230 sec latency 9.526 ms oleg328-client.virtnet: 1 111756 2.29 MB/sec execute 231 sec latency 804.643 ms oleg328-client.virtnet: 1 111756 2.28 MB/sec execute 232 sec latency 1804.853 ms oleg328-client.virtnet: 1 111756 2.27 MB/sec execute 233 sec latency 2805.039 ms oleg328-client.virtnet: 1 111756 2.26 MB/sec execute 234 sec latency 3805.226 ms oleg328-client.virtnet: 1 111756 2.25 MB/sec execute 235 sec latency 4805.408 ms oleg328-client.virtnet: 1 111756 2.24 MB/sec execute 236 sec latency 5805.639 ms oleg328-client.virtnet: 1 111756 2.24 MB/sec execute 237 sec latency 6805.873 ms oleg328-client.virtnet: 1 111756 2.23 MB/sec execute 238 sec latency 7806.081 ms oleg328-client.virtnet: 1 111756 2.22 MB/sec execute 239 sec latency 8806.284 ms oleg328-client.virtnet: 1 111756 2.21 MB/sec execute 240 sec latency 9806.494 ms oleg328-client.virtnet: 1 111756 2.20 MB/sec execute 241 sec latency 10806.759 ms oleg328-client.virtnet: 1 111756 2.19 MB/sec execute 242 sec latency 11806.Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0001 04:53:39 (1713430419) targets are mounted 04:53:39 (1713430419) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec 954 ms oleg328-client.virtnet: 1 111756 2.18 MB/sec execute 243 sec latency 12807.084 ms oleg328-client.virtnet: 1 111756 2.17 MB/sec execute 244 sec latency 13807.230 ms oleg328-client.virtnet: 1 111756 2.16 MB/sec execute 245 sec latency 14807.415 ms oleg328-client.virtnet: 1 111756 2.15 MB/sec execute 246 sec latency 15807.548 ms oleg328-client.virtnet: 1 111756 2.14 MB/sec execute 247 sec latency 16807.707 ms oleg328-client.virtnet: 1 111756 2.14 MB/sec execute 248 sec latency 17807.891 ms oleg328-client.virtnet: 1 111756 2.13 MB/sec execute 249 sec latency 18808.028 ms oleg328-client.virtnet: 1 112271 2.12 MB/sec execute 250 sec latency 18875.071 ms oleg328-client.virtnet: 1 113195 2.14 MB/sec execute 251 sec latency 8.904 ms oleg328-client.virtnet: 1 113887 2.16 MB/sec execute 252 sec latency 10.577 ms oleg328-client.virtnet: 1 114427 2.15 MB/sec execute 253 sec latency 11.185 ms oleg328-client.virtnet: 1 115101 2.16 MB/sec execute 254 sec latency 9.121 ms oleg328-client.virtnet: 1 115711 2.15 MB/sec execute 255 sec latency 9.523 ms oleg328-client.virtnet: 1 116553 2.17 MB/sec execute 256 sec latency 8.985 ms oleg328-client.virtnet: 1 117415 2.19 MB/sec execute 257 sec latency 9.462 ms oleg328-client.virtnet: 1 117988 2.18 MB/sec execute 258 sec latency 10.748 ms oleg328-client.virtnet: 1 118650 2.19 MB/sec execute 259 sec latency 10.883 ms oleg328-client.virtnet: 1 119253 2.19 MB/sec execute 260 sec latency 11.229 ms oleg328-client.virtnet: 1 120078 2.20 MB/sec execute 261 sec latency 8.892 ms oleg328-client.virtnet: 1 120921 2.22 MB/sec execute 262 sec latency 10.447 ms oleg328-client.virtnet: 1 121491 2.21 MB/sec execute 263 sec latency 10.842 ms oleg328-client.virtnet: 1 122192 2.22 MB/sec execute 264 sec latency 8.804 ms oleg328-client.virtnet: 1 122929 2.22 MB/sec execute 265 sec latency 8.404 ms oleg328-client.virtnet: 1 123989 2.24 MB/sec execute 266 sec latency 7.802 ms oleg328-client.virtnet: 1 124621 2.25 MB/sec execute 267 sec latency 10.704 ms oleg328-client.virtnet: 1 125220 2.25 MB/sec execute 268 sec latency 11.420 ms oleg328-client.virtnet: 1 125905 2.25 MB/sec execute 269 sec latency 11.256 ms oleg328-client.virtnet: 1 126633 2.25 MB/sec execute 270 sec latency 7.968 ms oleg328-client.virtnet: 1 127798 2.28 MB/sec execute 271 sec latency 12.279 ms oleg328-client.virtnet: 1 128379 2.28 MB/sec execute 272 sec latency 8.384 ms oleg328-client.virtnet: 1 128965 2.28 MB/sec execute 273 sec latency 10.327 ms oleg328-client.virtnet: 1 129590 2.28 MB/sec execute 274 sec latency 10.085 ms oleg328-client.virtnet: 1 130293 2.28 MB/sec execute 275 sec latency 10.565 ms oleg328-client.virtnet: 1 131438 2.31 MB/sec execute 276 sec latency 8.173 ms oleg328-client.virtnet: 1 132009 2.31 MB/sec execute 277 sec latency 12.466 ms oleg328-client.virtnet: 1 132679 2.32 MB/sec execute 278 sec latency 9.358 ms oleg328-client.virtnet: 1 133346 2.31 MB/sec execute 279 sec latency 9.437 ms oleg328-client.virtnet: 1 134258 2.33 MB/sec execute 280 sec latency 8.531 ms oleg328-client.virtnet: 1 135125 2.34 MB/sec execute 281 sec latency 9.770 ms oleg328-client.virtnet: 1 135731 2.34 MB/sec execute 282 sec latency 9.643 ms oleg328-client.virtnet: 1 136441 2.34 MB/sec execute 283 sec latency 9.461 ms oleg328-client.virtnet: 1 137172 2.34 MB/sec execute 284 sec latency 9.264 ms oleg328-client.virtnet: 1 138216 2.36 MB/sec execute 285 sec latency 7.938 ms oleg328-client.virtnet: 1 138791 2.37 MB/sec execute 286 sec latency 9.389 ms oleg328-client.virtnet: 1 139364 2.36 MB/sec execute 287 sec latency 10.702 ms oleg328-client.virtnet: 1 140072 2.37 MB/sec execute 288 sec latency 9.500 ms oleg328-client.virtnet: 1 140816 2.37 MB/sec execute 289 sec latency 8.218 ms oleg328-client.virtnet: 1 141974 2.40 MB/sec execute 290 sec latency 8.937 ms oleg328-client.virtnet: 1 142561 2.39 MB/sec execute 291 sec latency 8.888 ms oleg328-client.virtnet: 1 143163 2.39 MB/sec execute 292 sec latency 10.957 ms oleg328-client.virtnet: 1 143881 2.40 MB/sec execute 293 sec latency 9.554 ms oleg328-client.virtnet: 1 144682 2.41 MB/sec execute 294 sec latency 9.912 ms oleg328-client.virtnet: 1 145704 2.42 MB/sec execute 295 sec latency 8.539 ms oleg328-client.virtnet: 1 146286 2.42 MB/sec execute 296 sec latency 12.131 ms oleg328-client.virtnet: 1 147053 2.43 MB/sec execute 297 sec latency 8.363 ms oleg328-client.virtnet: 1 147833 2.42 MB/sec execute 298 sec latency 8.558 ms oleg328-client.virtnet: 1 148834 2.44 MB/sec execute 299 sec latency 9.109 ms oleg328-client.virtnet: 1 cleanup 300 sec oleg328-client.virtnet: 0 cleanup 300 sec oleg328-client.virtnet: oleg328-client.virtnet: Operation Count AvgLat MaxLat oleg328-client.virtnet: ---------------------------------------- oleg328-client.virtnet: NTCreateX 23512 5.663 18423.993 oleg328-client.virtnet: Close 17244 3.160 18875.059 oleg328-client.virtnet: Rename 996 6.662 18.792 oleg328-client.virtnet: Unlink 4758 1.884 11.959 oleg328-client.virtnet: Qpathinfo 21286 1.252 19.286 oleg328-client.virtnet: Qfileinfo 3713 0.152 2.230 oleg328-client.virtnet: Qfsinfo 3923 0.218 10.575 oleg328-client.virtnet: Sfileinfo 1900 22.957 18522.249 oleg328-client.virtnet: Find 8241 0.394 11.098 oleg328-client.virtnet: WriteX 11616 0.861 10.036 oleg328-client.virtnet: ReadX 36998 0.020 3.783 oleg328-client.virtnet: LockX 78 0.964 1.832 oleg328-client.virtnet: UnlockX 78 0.943 1.868 oleg328-client.virtnet: Flush 1642 6.012 444.810 oleg328-client.virtnet: oleg328-client.virtnet: Throughput 2.44186 MB/sec 1 clients 1 procs max_latency=18875.071 ms oleg328-client.virtnet: stopping dbench on /mnt/lustre/d70b.replay-single/oleg328-client.virtnet at Thu Apr 18 04:54:34 EDT 2024 with return code 0 oleg328-client.virtnet: clean dbench files on /mnt/lustre/d70b.replay-single/oleg328-client.virtnet oleg328-client.virtnet: /mnt/lustre/d70b.replay-single/oleg328-client.virtnet /mnt/lustre/d70b.replay-single/oleg328-client.virtnet oleg328-client.virtnet: removed directory: 'clients/client0/~dmtmp/ACCESS' oleg328-client.virtnet: removed directory: 'clients/client0/~dmtmp/SEED' oleg328-client.virtnet: removed directory: 'clients/client0/~dmtmp/WORDPRO' oleg328-client.virtnet: removed directory: 'clients/client0/~dmtmp/WORD' oleg328-client.virtnet: removed directory: 'clients/client0/~dmtmp/PARADOX' oleg328-client.virtnet: removed directory: 'clients/client0/~dmtmp/PM' oleg328-client.virtnet: removed directory: 'clients/client0/~dmtmp/COREL' oleg328-client.virtnet: removed directory: 'clients/client0/~dmtmp/PWRPNT' oleg328-client.virtnet: removed directory: 'clients/client0/~dmtmp/EXCEL' oleg328-client.virtnet: removed directory: 'clients/client0/~dmtmp' oleg328-client.virtnet: removed directory: 'clients/client0' oleg328-client.virtnet: removed directory: 'clients' oleg328-client.virtnet: removed 'client.txt' oleg328-client.virtnet: /mnt/lustre/d70b.replay-single/oleg328-client.virtnet oleg328-client.virtnet: dbench successfully finished PASS 70b (363s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 70c: tar 2mdts recovery ============ 04:54:37 (1713430477) Starting client oleg328-client.virtnet: -o user_xattr,flock oleg328-server@tcp:/lustre /mnt/lustre Started clients oleg328-client.virtnet: 192.168.203.128@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Started tar 3479 striped dir -i0 -c2 -H fnv_1a_64 /mnt/lustre/d70c.replay-single tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 7068 1280620 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 4684 1283004 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 10348 3560732 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 10488 3558636 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 20836 7119368 1% /mnt/lustre tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets test_70c fail mds1 1 times Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:56:52 (1713430612) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:57:06 (1713430626) targets are mounted 04:57:06 (1713430626) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 7296 1280392 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 4800 1282888 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 6340 3575692 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 10120 3565212 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 16460 7140904 1% /mnt/lustre tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets test_70c fail mds1 2 times Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 04:59:27 (1713430767) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 04:59:41 (1713430781) targets are mounted 04:59:41 (1713430781) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 70c (353s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 70d: mkdir/rmdir striped dir 2mdts recovery ========================================================== 05:00:32 (1713430832) Starting client oleg328-client.virtnet: -o user_xattr,flock oleg328-server@tcp:/lustre /mnt/lustre Started clients oleg328-client.virtnet: 192.168.203.128@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Started 6854 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3768 1283920 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 5336 1282352 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3600756 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3597840 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7198596 1% /mnt/lustre test_70d fail mds2 1 times Failing mds2 on oleg328-server Stopping /mnt/lustre-mds2 (opts:) on oleg328-server 05:02:46 (1713430966) shut down Failover mds2 to oleg328-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0001 05:03:00 (1713430980) targets are mounted 05:03:00 (1713430980) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 5900 1281788 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 3716 1283972 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3600756 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3597840 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7198596 1% /mnt/lustre test_70d fail mds1 2 times Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 05:05:25 (1713431125) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 05:05:39 (1713431139) targets are mounted 05:05:39 (1713431139) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /home/green/git/lustre-release/lustre/tests/test-framework.sh: line 4697: 6854 Killed ( while true; do $LFS mkdir -i0 -c2 $DIR/$tdir/test || { echo "mkdir fails"; break; }; $LFS mkdir -i1 -c2 $DIR/$tdir/test1 || { echo "mkdir fails"; break; }; touch $DIR/$tdir/test/a || { echo "touch fails"; break; }; mkdir $DIR/$tdir/test/b || { echo "mkdir fails"; break; }; rm -rf $DIR/$tdir/test || { echo "rmdir fails"; ls -lR $DIR/$tdir; break; }; touch $DIR/$tdir/test1/a || { echo "touch fails"; break; }; mkdir $DIR/$tdir/test1/b || { echo "mkdir fails"; break; }; rm -rf $DIR/$tdir/test1 || { echo "rmdir fails"; ls -lR $DIR/$tdir/test1; break; }; done ) (wd: ~) PASS 70d (315s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 70e: rename cross-MDT with random fails ========================================================== 05:05:49 (1713431149) debug=+ha Starting client oleg328-client.virtnet: -o user_xattr,flock oleg328-server@tcp:/lustre /mnt/lustre Started clients oleg328-client.virtnet: 192.168.203.128@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Started PID=2943 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 4592 1283096 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 3652 1284036 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3600756 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3597840 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7198596 1% /mnt/lustre test_70e fail mds1 1 times Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 05:08:03 (1713431283) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 05:08:17 (1713431297) targets are mounted 05:08:17 (1713431297) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 4200 1283488 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 3556 1284132 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3600756 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3597840 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7198596 1% /mnt/lustre test_70e fail mds1 2 times Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 05:10:39 (1713431439) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 05:10:53 (1713431453) targets are mounted 05:10:53 (1713431453) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 70e (315s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 70f: OSS O_DIRECT recovery with 1 clients ========================================================== 05:11:05 (1713431465) mount clients oleg328-client.virtnet ... Starting client oleg328-client.virtnet: -o user_xattr,flock oleg328-server@tcp:/lustre /mnt/lustre Started clients oleg328-client.virtnet: 192.168.203.128@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3928 1283760 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 3560 1284128 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3555520 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3572216 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7127736 1% /mnt/lustre ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... test_70f failing OST 1 times Failing ost1 on oleg328-server Stopping /mnt/lustre-ost1 (opts:) on oleg328-server 05:11:13 (1713431473) shut down Failover ost1 to oleg328-server mount facets: ost1 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-OST0000 05:11:27 (1713431487) targets are mounted 05:11:27 (1713431487) facet_failover done ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3928 1283760 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 3560 1284128 1% /mnt/lustre[MDT:1] Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... lustre-OST0000_UUID 3833116 1580 3574288 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 5652 3580624 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 7232 7154912 1% /mnt/lustre ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... test_70f failing OST 2 times Failing ost1 on oleg328-server Stopping /mnt/lustre-ost1 (opts:) on oleg328-server 05:11:40 (1713431500) shut down Failover ost1 to oleg328-server mount facets: ost1 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-OST0000 05:11:54 (1713431514) targets are mounted 05:11:54 (1713431514) facet_failover done ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3928 1283760 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 3560 1284128 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3571144 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3584720 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7155864 1% /mnt/lustre ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... test_70f failing OST 3 times Failing ost1 on oleg328-server Stopping /mnt/lustre-ost1 (opts:) on oleg328-server 05:12:07 (1713431527) shut down Failover ost1 to oleg328-server mount facets: ost1 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... Started lustre-OST0000 05:12:22 (1713431542) targets are mounted 05:12:22 (1713431542) facet_failover done ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3928 1283760 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 3560 1284128 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3574288 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3601344 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7175632 1% /mnt/lustre Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... test_70f failing OST 4 times Failing ost1 on oleg328-server Stopping /mnt/lustre-ost1 (opts:) on oleg328-server 05:12:34 (1713431554) shut down Failover ost1 to oleg328-server mount facets: ost1 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Started lustre-OST0000 05:12:48 (1713431568) targets are mounted 05:12:48 (1713431568) facet_failover done Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3928 1283760 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 3560 1284128 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3576360 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3601344 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7177704 1% /mnt/lustre ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... test_70f failing OST 5 times Failing ost1 on oleg328-server Stopping /mnt/lustre-ost1 (opts:) on oleg328-server 05:13:01 (1713431581) shut down Failover ost1 to oleg328-server mount facets: ost1 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Started lustre-OST0000 05:13:15 (1713431595) targets are mounted 05:13:15 (1713431595) facet_failover done Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg328-client.virtnet' ... ldlm.namespaces.MGC192.168.203.128@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b589b800.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b589b800.lru_size=clear PASS 70f (137s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 71a: mkdir/rmdir striped dir with 2 mdts recovery ========================================================== 05:13:24 (1713431604) Starting client oleg328-client.virtnet: -o user_xattr,flock oleg328-server@tcp:/lustre /mnt/lustre Started clients oleg328-client.virtnet: 192.168.203.128@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Started 423 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 4168 1283520 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 5216 1282472 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3605440 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 4392 1283296 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 5468 1282220 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3605440 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre fail mds1 mds2 1 times Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server Failing mds2 on oleg328-server Stopping /mnt/lustre-mds2 (opts:) on oleg328-server 05:15:47 (1713431747) shut down Failover mds1 to oleg328-server mount facets: mds1 Failover mds2 to oleg328-server mount facets: mds2 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0001 Started lustre-MDT0000 05:16:11 (1713431771) targets are mounted 05:16:11 (1713431771) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid,mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2536 1285152 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 3364 1284324 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3605440 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2700 1284988 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 3548 1284140 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3605440 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre fail mds2 mds1 2 times Failing mds2 on oleg328-server Stopping /mnt/lustre-mds2 (opts:) on oleg328-server Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 05:18:45 (1713431925) shut down Failover mds1 to oleg328-server Failover mds2 to oleg328-server mount facets: mds2 mount facets: mds1 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0001 Started lustre-MDT0000 05:19:00 (1713431940) targets are mounted 05:19:00 (1713431940) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid,mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec /home/green/git/lustre-release/lustre/tests/test-framework.sh: line 4697: 423 Killed ( while true; do $LFS mkdir -i0 -c2 $DIR/$tdir/test; rmdir $DIR/$tdir/test; done ) (wd: ~) PASS 71a (351s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 73a: open(O_CREAT), unlink, replay, reconnect before open replay, close ========================================================== 05:19:16 (1713431956) multiop /mnt/lustre/f73a.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7515 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 4180 1283508 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 4020 1283668 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3605440 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre fail_loc=0x80000302 Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 05:19:20 (1713431960) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 05:19:34 (1713431974) targets are mounted 05:19:34 (1713431974) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 73a (39s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 73b: open(O_CREAT), unlink, replay, reconnect at open_replay reply, close ========================================================== 05:19:57 (1713431997) multiop /mnt/lustre/f73b.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7515 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2436 1285252 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2040 1285648 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3605440 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre fail_loc=0x80000157 Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 05:20:00 (1713432000) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 05:20:14 (1713432014) targets are mounted 05:20:14 (1713432014) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 73b (38s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 74: Ensure applications don't fail waiting for OST recovery ========================================================== 05:20:37 (1713432037) Stopping clients: oleg328-client.virtnet /mnt/lustre (opts:) Stopping client oleg328-client.virtnet /mnt/lustre opts: Stopping /mnt/lustre-ost1 (opts:) on oleg328-server Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 05:20:39 (1713432039) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 05:20:52 (1713432052) targets are mounted 05:20:52 (1713432052) facet_failover done Starting client oleg328-client.virtnet: -o user_xattr,flock oleg328-server@tcp:/lustre /mnt/lustre Started clients oleg328-client.virtnet: 192.168.203.128@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-OST0000 PASS 74 (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80a: DNE: create remote dir, drop update rep from MDT0, fail MDT0 ========================================================== 05:21:04 (1713432064) fail_loc=0x1701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2440 1285248 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2080 1285608 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3605440 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 05:21:08 (1713432068) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 05:21:21 (1713432081) targets are mounted 05:21:21 (1713432081) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec total: 20 open/close in 0.05 seconds: 402.94 ops/second PASS 80a (24s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80b: DNE: create remote dir, drop update rep from MDT0, fail MDT1 ========================================================== 05:21:30 (1713432090) fail_loc=0x1701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2480 1285208 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2080 1285608 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3605440 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2480 1285208 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2080 1285608 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3605440 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre Failing mds2 on oleg328-server Stopping /mnt/lustre-mds2 (opts:) on oleg328-server 05:21:40 (1713432100) shut down Failover mds2 to oleg328-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0001 05:21:54 (1713432114) targets are mounted 05:21:54 (1713432114) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec total: 20 open/close in 0.05 seconds: 421.66 ops/second PASS 80b (31s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80c: DNE: create remote dir, drop update rep from MDT1, fail MDT[0,1] ========================================================== 05:22:02 (1713432122) fail_loc=0x1701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2512 1285176 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2112 1285576 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3605440 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2516 1285172 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2116 1285572 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3605440 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 05:22:07 (1713432127) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 05:22:21 (1713432141) targets are mounted 05:22:21 (1713432141) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Failing mds2 on oleg328-server Stopping /mnt/lustre-mds2 (opts:) on oleg328-server 05:22:28 (1713432148) shut down Failover mds2 to oleg328-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0001 05:22:41 (1713432161) targets are mounted 05:22:41 (1713432161) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec total: 20 open/close in 0.05 seconds: 437.11 ops/second PASS 80c (47s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80d: DNE: create remote dir, drop update rep from MDT1, fail 2 MDTs ========================================================== 05:22:50 (1713432170) fail_loc=0x1701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2552 1285136 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2080 1285608 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3605440 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2552 1285136 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2080 1285608 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3605440 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server Failing mds2 on oleg328-server Stopping /mnt/lustre-mds2 (opts:) on oleg328-server 05:23:04 (1713432184) shut down Failover mds1 to oleg328-server mount facets: mds1 Failover mds2 to oleg328-server mount facets: mds2 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0001 Started lustre-MDT0000 05:23:25 (1713432205) targets are mounted 05:23:25 (1713432205) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid,mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec total: 20 open/close in 0.09 seconds: 219.52 ops/second PASS 80d (43s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80e: DNE: create remote dir, drop MDT1 rep, fail MDT0 ========================================================== 05:23:34 (1713432214) fail_loc=0x119 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2480 1285208 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2080 1285608 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3605440 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 05:23:42 (1713432222) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 05:23:56 (1713432236) targets are mounted 05:23:56 (1713432236) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec total: 20 open/close in 0.05 seconds: 399.99 ops/second PASS 80e (29s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80f: DNE: create remote dir, drop MDT1 rep, fail MDT1 ========================================================== 05:24:05 (1713432245) fail_loc=0x119 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2480 1285208 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2080 1285608 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3605440 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre Failing mds2 on oleg328-server Stopping /mnt/lustre-mds2 (opts:) on oleg328-server 05:24:08 (1713432248) shut down Failover mds2 to oleg328-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0001 05:24:23 (1713432263) targets are mounted 05:24:23 (1713432263) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec total: 20 open/close in 0.10 seconds: 206.60 ops/second PASS 80f (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80g: DNE: create remote dir, drop MDT1 rep, fail MDT0, then MDT1 ========================================================== 05:24:33 (1713432273) fail_loc=0x119 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2480 1285208 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2080 1285608 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3605440 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2480 1285208 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2080 1285608 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3605440 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 05:24:42 (1713432282) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 05:24:57 (1713432297) targets are mounted 05:24:57 (1713432297) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Failing mds2 on oleg328-server Stopping /mnt/lustre-mds2 (opts:) on oleg328-server 05:25:04 (1713432304) shut down Failover mds2 to oleg328-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0001 05:25:19 (1713432319) targets are mounted 05:25:19 (1713432319) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec total: 20 open/close in 0.05 seconds: 409.95 ops/second PASS 80g (53s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80h: DNE: create remote dir, drop MDT1 rep, fail 2 MDTs ========================================================== 05:25:28 (1713432328) fail_loc=0x119 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2480 1285208 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2080 1285608 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3605440 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2480 1285208 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2080 1285608 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3605440 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server Failing mds2 on oleg328-server Stopping /mnt/lustre-mds2 (opts:) on oleg328-server 05:25:39 (1713432339) shut down Failover mds1 to oleg328-server mount facets: mds1 Failover mds2 to oleg328-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 Started lustre-MDT0001 05:26:05 (1713432365) targets are mounted 05:26:05 (1713432365) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid,mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec total: 20 open/close in 0.06 seconds: 334.10 ops/second PASS 80h (53s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81a: DNE: unlink remote dir, drop MDT0 update rep, fail MDT1 ========================================================== 05:26:23 (1713432383) fail_loc=0x1701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2476 1285212 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2080 1285608 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3605440 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre Failing mds2 on oleg328-server Stopping /mnt/lustre-mds2 (opts:) on oleg328-server 05:26:32 (1713432392) shut down Failover mds2 to oleg328-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0001 05:26:46 (1713432406) targets are mounted 05:26:46 (1713432406) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 81a (30s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81b: DNE: unlink remote dir, drop MDT0 update reply, fail MDT0 ========================================================== 05:26:55 (1713432415) fail_loc=0x1701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2512 1285176 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2084 1285604 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3605440 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 05:26:58 (1713432418) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 05:27:12 (1713432432) targets are mounted 05:27:12 (1713432432) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 81b (24s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81c: DNE: unlink remote dir, drop MDT0 update reply, fail MDT0,MDT1 ========================================================== 05:27:20 (1713432440) fail_loc=0x1701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2476 1285212 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2084 1285604 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3605440 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2476 1285212 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2084 1285604 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3605440 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 05:27:27 (1713432447) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 05:27:41 (1713432461) targets are mounted 05:27:41 (1713432461) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Failing mds2 on oleg328-server Stopping /mnt/lustre-mds2 (opts:) on oleg328-server 05:27:48 (1713432468) shut down Failover mds2 to oleg328-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0001 05:28:03 (1713432483) targets are mounted 05:28:03 (1713432483) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 81c (50s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81d: DNE: unlink remote dir, drop MDT0 update reply, fail 2 MDTs ========================================================== 05:28:12 (1713432492) fail_loc=0x1701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2512 1285176 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2088 1285600 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3605440 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2512 1285176 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2088 1285600 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3605440 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server Failing mds2 on oleg328-server Stopping /mnt/lustre-mds2 (opts:) on oleg328-server 05:28:24 (1713432504) shut down Failover mds1 to oleg328-server mount facets: mds1 Failover mds2 to oleg328-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 Started lustre-MDT0001 05:28:45 (1713432525) targets are mounted 05:28:45 (1713432525) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid,mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 81d (46s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81e: DNE: unlink remote dir, drop MDT1 req reply, fail MDT0 ========================================================== 05:28:59 (1713432539) fail_loc=0x119 fail_loc=0 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2476 1285212 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2088 1285600 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3605440 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 05:29:03 (1713432543) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 05:29:17 (1713432557) targets are mounted 05:29:17 (1713432557) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 81e (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81f: DNE: unlink remote dir, drop MDT1 req reply, fail MDT1 ========================================================== 05:29:27 (1713432567) fail_loc=0x119 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2476 1285212 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2088 1285600 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3605440 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre Failing mds2 on oleg328-server Stopping /mnt/lustre-mds2 (opts:) on oleg328-server 05:29:30 (1713432570) shut down Failover mds2 to oleg328-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0001 05:29:44 (1713432584) targets are mounted 05:29:44 (1713432584) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 81f (24s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81g: DNE: unlink remote dir, drop req reply, fail M0, then M1 ========================================================== 05:29:53 (1713432593) fail_loc=0x119 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2476 1285212 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2088 1285600 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3605440 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2476 1285212 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2088 1285600 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3605440 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 05:29:58 (1713432598) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 05:30:12 (1713432612) targets are mounted 05:30:12 (1713432612) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Failing mds2 on oleg328-server Stopping /mnt/lustre-mds2 (opts:) on oleg328-server 05:30:19 (1713432619) shut down Failover mds2 to oleg328-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0001 05:30:33 (1713432633) targets are mounted 05:30:33 (1713432633) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 81g (47s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81h: DNE: unlink remote dir, drop request reply, fail 2 MDTs ========================================================== 05:30:41 (1713432641) fail_loc=0x119 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2436 1285252 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2052 1285636 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3605440 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2476 1285212 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2088 1285600 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1580 3605440 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server Failing mds2 on oleg328-server Stopping /mnt/lustre-mds2 (opts:) on oleg328-server 05:30:48 (1713432648) shut down Failover mds1 to oleg328-server mount facets: mds1 Failover mds2 to oleg328-server mount facets: mds2 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 Started lustre-MDT0001 05:31:14 (1713432674) targets are mounted 05:31:14 (1713432674) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid,mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 81h (49s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 84a: stale open during export disconnect ========================================================== 05:31:32 (1713432692) fail_loc=0x80000144 total: 1 open/close in 0.00 seconds: 240.33 ops/second pdsh@oleg328-client: oleg328-client: ssh exited with exit code 5 PASS 84a (4s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 85a: check the cancellation of unused locks during recovery(IBITS) ========================================================== 05:31:37 (1713432697) before recovery: unused locks count = 201 Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 05:31:40 (1713432700) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 05:31:53 (1713432713) targets are mounted 05:31:53 (1713432713) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec after recovery: unused locks count = 101 PASS 85a (23s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 85b: check the cancellation of unused locks during recovery(EXTENT) ========================================================== 05:32:01 (1713432721) before recovery: unused locks count = 100 Failing ost1 on oleg328-server Stopping /mnt/lustre-ost1 (opts:) on oleg328-server 05:32:05 (1713432725) shut down Failover ost1 to oleg328-server mount facets: ost1 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-OST0000 05:32:18 (1713432738) targets are mounted 05:32:18 (1713432738) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec after recovery: unused locks count = 0 PASS 85b (22s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 86: umount server after clear nid_stats should not hit LBUG ========================================================== 05:32:24 (1713432744) Stopping clients: oleg328-client.virtnet /mnt/lustre (opts:) Stopping client oleg328-client.virtnet /mnt/lustre opts: mdt.lustre-MDT0000.exports.clear=0 mdt.lustre-MDT0001.exports.clear=0 Stopping /mnt/lustre-mds1 (opts:) on oleg328-server Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 Starting client oleg328-client.virtnet: -o user_xattr,flock oleg328-server@tcp:/lustre /mnt/lustre Started clients oleg328-client.virtnet: 192.168.203.128@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) PASS 86 (11s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 87a: write replay ================== 05:32:36 (1713432756) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2472 1285216 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2092 1285596 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1780 3605240 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1756 3605264 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3536 7210504 1% /mnt/lustre 8+0 records in 8+0 records out 8388608 bytes (8.4 MB) copied, 0.104325 s, 80.4 MB/s Failing ost1 on oleg328-server Stopping /mnt/lustre-ost1 (opts:) on oleg328-server 05:32:40 (1713432760) shut down Failover ost1 to oleg328-server mount facets: ost1 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-OST0000 05:32:54 (1713432774) targets are mounted 05:32:54 (1713432774) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec 8+0 records in 8+0 records out 8388608 bytes (8.4 MB) copied, 0.0342541 s, 245 MB/s PASS 87a (23s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 87b: write replay with changed data (checksum resend) ========================================================== 05:33:01 (1713432781) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2472 1285216 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2092 1285596 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 9972 3597048 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1756 3605264 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 11728 7202312 1% /mnt/lustre 8+0 records in 8+0 records out 8388608 bytes (8.4 MB) copied, 0.125197 s, 67.0 MB/s 8+0 records in 8+0 records out 8 bytes (8 B) copied, 0.00159896 s, 5.0 kB/s Failing ost1 on oleg328-server Stopping /mnt/lustre-ost1 (opts:) on oleg328-server 05:33:06 (1713432786) shut down Failover ost1 to oleg328-server mount facets: ost1 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-OST0000 05:33:21 (1713432801) targets are mounted 05:33:21 (1713432801) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec 0+1 records in 0+1 records out 72 bytes (72 B) copied, 0.00395972 s, 18.2 kB/s PASS 87b (24s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 88: MDS should not assign same objid to different files ========================================================== 05:33:27 (1713432807) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2476 1285212 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2092 1285596 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 9976 3597044 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1756 3605264 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 11732 7202308 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2476 1285212 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2092 1285596 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 9976 3597044 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1756 3605264 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 11732 7202308 1% /mnt/lustre before test: last_id = 17345, next_id = 17316 Creating to objid 17345 on ost lustre-OST0000... total: 31 open/close in 0.07 seconds: 427.42 ops/second total: 8 open/close in 0.02 seconds: 426.01 ops/second before recovery: last_id = 17377, next_id = 17354 Stopping /mnt/lustre-mds1 (opts:) on oleg328-server Stopping /mnt/lustre-ost1 (opts:) on oleg328-server Failover mds1 to oleg328-server Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 Failover ost1 to oleg328-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-OST0000 after recovery: last_id = 17385, next_id = 17354 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0258692 s, 20.3 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0243844 s, 21.5 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0256004 s, 20.5 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0279429 s, 18.8 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0257503 s, 20.4 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0246959 s, 21.2 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.028809 s, 18.2 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0333588 s, 15.7 MB/s -rw-r--r-- 1 root root 0 Apr 18 05:33 /mnt/lustre/d88.replay-single/f-17316 -rw-r--r-- 1 root root 0 Apr 18 05:33 /mnt/lustre/d88.replay-single/f-17317 -rw-r--r-- 1 root root 0 Apr 18 05:33 /mnt/lustre/d88.replay-single/f-17318 -rw-r--r-- 1 root root 0 Apr 18 05:33 /mnt/lustre/d88.replay-single/f-17319 -rw-r--r-- 1 root root 0 Apr 18 05:33 /mnt/lustre/d88.replay-single/f-17320 -rw-r--r-- 1 root root 0 Apr 18 05:33 /mnt/lustre/d88.replay-single/f-17321 -rw-r--r-- 1 root root 0 Apr 18 05:33 /mnt/lustre/d88.replay-single/f-17322 -rw-r--r-- 1 root root 0 Apr 18 05:33 /mnt/lustre/d88.replay-single/f-17323 -rw-r--r-- 1 root root 0 Apr 18 05:33 /mnt/lustre/d88.replay-single/f-17324 -rw-r--r-- 1 root root 0 Apr 18 05:33 /mnt/lustre/d88.replay-single/f-17325 -rw-r--r-- 1 root root 0 Apr 18 05:33 /mnt/lustre/d88.replay-single/f-17326 -rw-r--r-- 1 root root 0 Apr 18 05:33 /mnt/lustre/d88.replay-single/f-17327 -rw-r--r-- 1 root root 0 Apr 18 05:33 /mnt/lustre/d88.replay-single/f-17328 -rw-r--r-- 1 root root 0 Apr 18 05:33 /mnt/lustre/d88.replay-single/f-17329 -rw-r--r-- 1 root root 0 Apr 18 05:33 /mnt/lustre/d88.replay-single/f-17330 -rw-r--r-- 1 root root 0 Apr 18 05:33 /mnt/lustre/d88.replay-single/f-17331 -rw-r--r-- 1 root root 0 Apr 18 05:33 /mnt/lustre/d88.replay-single/f-17332 -rw-r--r-- 1 root root 0 Apr 18 05:33 /mnt/lustre/d88.replay-single/f-17333 -rw-r--r-- 1 root root 0 Apr 18 05:33 /mnt/lustre/d88.replay-single/f-17334 -rw-r--r-- 1 root root 0 Apr 18 05:33 /mnt/lustre/d88.replay-single/f-17335 -rw-r--r-- 1 root root 0 Apr 18 05:33 /mnt/lustre/d88.replay-single/f-17336 -rw-r--r-- 1 root root 0 Apr 18 05:33 /mnt/lustre/d88.replay-single/f-17337 -rw-r--r-- 1 root root 0 Apr 18 05:33 /mnt/lustre/d88.replay-single/f-17338 -rw-r--r-- 1 root root 0 Apr 18 05:33 /mnt/lustre/d88.replay-single/f-17339 -rw-r--r-- 1 root root 0 Apr 18 05:33 /mnt/lustre/d88.replay-single/f-17340 -rw-r--r-- 1 root root 0 Apr 18 05:33 /mnt/lustre/d88.replay-single/f-17341 -rw-r--r-- 1 root root 0 Apr 18 05:33 /mnt/lustre/d88.replay-single/f-17342 -rw-r--r-- 1 root root 0 Apr 18 05:33 /mnt/lustre/d88.replay-single/f-17343 -rw-r--r-- 1 root root 0 Apr 18 05:33 /mnt/lustre/d88.replay-single/f-17344 -rw-r--r-- 1 root root 0 Apr 18 05:33 /mnt/lustre/d88.replay-single/f-17345 -rw-r--r-- 1 root root 0 Apr 18 05:33 /mnt/lustre/d88.replay-single/f-17346 -rw-r--r-- 1 root root 0 Apr 18 05:33 /mnt/lustre/d88.replay-single/f-17347 -rw-r--r-- 1 root root 0 Apr 18 05:33 /mnt/lustre/d88.replay-single/f-17348 -rw-r--r-- 1 root root 0 Apr 18 05:33 /mnt/lustre/d88.replay-single/f-17349 -rw-r--r-- 1 root root 0 Apr 18 05:33 /mnt/lustre/d88.replay-single/f-17350 -rw-r--r-- 1 root root 0 Apr 18 05:33 /mnt/lustre/d88.replay-single/f-17351 -rw-r--r-- 1 root root 0 Apr 18 05:33 /mnt/lustre/d88.replay-single/f-17352 -rw-r--r-- 1 root root 0 Apr 18 05:33 /mnt/lustre/d88.replay-single/f-17353 -rw-r--r-- 1 root root 524288 Apr 18 05:34 /mnt/lustre/d88.replay-single/f-17357 -rw-r--r-- 1 root root 524288 Apr 18 05:34 /mnt/lustre/d88.replay-single/f-17358 -rw-r--r-- 1 root root 524288 Apr 18 05:34 /mnt/lustre/d88.replay-single/f-17359 -rw-r--r-- 1 root root 524288 Apr 18 05:34 /mnt/lustre/d88.replay-single/f-17360 -rw-r--r-- 1 root root 524288 Apr 18 05:34 /mnt/lustre/d88.replay-single/f-17361 -rw-r--r-- 1 root root 524288 Apr 18 05:34 /mnt/lustre/d88.replay-single/f-17362 -rw-r--r-- 1 root root 524288 Apr 18 05:34 /mnt/lustre/d88.replay-single/f-17363 -rw-r--r-- 1 root root 524288 Apr 18 05:34 /mnt/lustre/d88.replay-single/f-17364 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0298837 s, 17.5 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0239077 s, 21.9 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0261184 s, 20.1 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0248703 s, 21.1 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0274488 s, 19.1 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0259066 s, 20.2 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0255322 s, 20.5 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0287834 s, 18.2 MB/s PASS 88 (52s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 89: no disk space leak on late ost connection ========================================================== 05:34:20 (1713432860) Waiting for orphan cleanup... osp.lustre-OST0000-osc-MDT0000.old_sync_processed osp.lustre-OST0000-osc-MDT0001.old_sync_processed osp.lustre-OST0001-osc-MDT0000.old_sync_processed osp.lustre-OST0001-osc-MDT0001.old_sync_processed wait 40 secs maximumly for oleg328-server mds-ost sync done. Waiting for MDT destroys to complete 10+0 records in 10+0 records out 10485760 bytes (10 MB) copied, 0.0698155 s, 150 MB/s Stopping /mnt/lustre-ost1 (opts:) on oleg328-server Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 05:34:26 (1713432866) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 05:34:39 (1713432879) targets are mounted 05:34:39 (1713432879) facet_failover done Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-OST0000 Starting client: oleg328-client.virtnet: -o user_xattr,flock oleg328-server@tcp:/lustre /mnt/lustre osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 68 sec Waiting for orphan cleanup... osp.lustre-OST0000-osc-MDT0000.old_sync_processed osp.lustre-OST0000-osc-MDT0001.old_sync_processed osp.lustre-OST0001-osc-MDT0000.old_sync_processed osp.lustre-OST0001-osc-MDT0001.old_sync_processed wait 40 secs maximumly for oleg328-server mds-ost sync done. Waiting for MDT destroys to complete free_before: 7646308 free_after: 7646308 PASS 89 (105s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 90: lfs find identifies the missing striped file segments ========================================================== 05:36:07 (1713432967) Create the files Fail ost2 lustre-OST0001_UUID, display the list of affected files Stopping /mnt/lustre-ost2 (opts:) on oleg328-server General Query: lfs find /mnt/lustre/d90.replay-single /mnt/lustre/d90.replay-single /mnt/lustre/d90.replay-single/f0 /mnt/lustre/d90.replay-single/f1 /mnt/lustre/d90.replay-single/all Querying files on shutdown ost2: lfs find --obd lustre-OST0001_UUID /mnt/lustre/d90.replay-single/f1 /mnt/lustre/d90.replay-single/all Check getstripe: /home/green/git/lustre-release/lustre/utils/lfs getstripe -r --obd lustre-OST0001_UUID /mnt/lustre/d90.replay-single/f1 lmm_stripe_count: 1 lmm_stripe_size: 4194304 lmm_pattern: raid0 lmm_layout_gen: 0 lmm_stripe_offset: 1 obdidx objid objid group 1 17315 0x43a3 0x2c0000401 * /mnt/lustre/d90.replay-single/all lmm_stripe_count: 2 lmm_stripe_size: 4194304 lmm_pattern: raid0 lmm_layout_gen: 0 lmm_stripe_offset: 1 obdidx objid objid group 1 17314 0x43a2 0x2c0000401 * /mnt/lustre/d90.replay-single/all /mnt/lustre/d90.replay-single/f1 Failover ost2 to oleg328-server Starting ost2: -o localrecov /dev/mapper/ost2_flakey /mnt/lustre-ost2 seq.cli-lustre-OST0001-super.width=65536 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-OST0001 PASS 90 (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 93a: replay + reconnect ============ 05:36:28 (1713432988) 1+0 records in 1+0 records out 1024 bytes (1.0 kB) copied, 0.00152072 s, 673 kB/s fail_val=40 fail_loc=0x715 Failing ost1 on oleg328-server Stopping /mnt/lustre-ost1 (opts:) on oleg328-server 05:36:30 (1713432990) shut down Failover ost1 to oleg328-server mount facets: ost1 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-OST0000 05:36:44 (1713433004) targets are mounted 05:36:44 (1713433004) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec PASS 93a (59s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 93b: replay + reconnect on mds ===== 05:37:28 (1713433048) total: 20 open/close in 0.07 seconds: 305.42 ops/second fail_val=80 fail_loc=0x715 Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 05:37:31 (1713433051) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 05:37:44 (1713433064) targets are mounted 05:37:44 (1713433064) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 93b (103s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 100a: DNE: create striped dir, drop update rep from MDT1, fail MDT1 ========================================================== 05:39:13 (1713433153) fail_loc=0x1701 Failing mds2 on oleg328-server Stopping /mnt/lustre-mds2 (opts:) on oleg328-server 05:39:14 (1713433154) shut down Failover mds2 to oleg328-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0001 05:39:27 (1713433167) targets are mounted 05:39:27 (1713433167) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec lmv_stripe_count: 2 lmv_stripe_offset: 0 lmv_hash_type: crush mdtidx FID[seq:oid:ver] 0 [0x2000347f0:0x1:0x0] 1 [0x24000abfa:0x1:0x0] total: 20 open/close in 0.05 seconds: 428.94 ops/second PASS 100a (21s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 100b: DNE: create striped dir, fail MDT0 ========================================================== 05:39:36 (1713433176) fail_loc=0x119 Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 05:39:38 (1713433178) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 05:39:51 (1713433191) targets are mounted 05:39:51 (1713433191) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec lmv_stripe_count: 2 lmv_stripe_offset: 0 lmv_hash_type: crush mdtidx FID[seq:oid:ver] 0 [0x2000347f0:0x4:0x0] 1 [0x24000abfa:0x4:0x0] total: 20 open/close in 0.05 seconds: 411.18 ops/second PASS 100b (22s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 100c: DNE: create striped dir, abort_recov_mdt mds2 ========================================================== 05:40:00 (1713433200) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2568 1285120 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2168 1285520 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18172 3588848 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1772 3605248 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19944 7194096 1% /mnt/lustre Stopping /mnt/lustre-mds2 (opts:) on oleg328-server Failover mds2 to oleg328-server Starting mds2: -o localrecov -o abort_recov_mdt /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0001 total: 20 open/close in 0.09 seconds: 229.21 ops/second Failing mds2 on oleg328-server Stopping /mnt/lustre-mds2 (opts:) on oleg328-server 05:40:21 (1713433221) shut down Failover mds2 to oleg328-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0001 05:40:35 (1713433235) targets are mounted 05:40:35 (1713433235) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec lmv_stripe_count: 2 lmv_stripe_offset: 1 lmv_hash_type: crush mdtidx FID[seq:oid:ver] 1 [0x24000b3c8:0x3:0x0] 0 [0x2000347f1:0x3:0x0] total: 20 open/close in 0.07 seconds: 267.73 ops/second PASS 100c (42s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 100d: DNE: cancel update logs upon recovery abort ========================================================== 05:40:44 (1713433244) striped dir -i0 -c2 -H all_char /mnt/lustre/d100d.replay-single total: 100 mkdir in 0.33 seconds: 304.28 ops/second lustre-MDT0001-osd [catalog]: [0x24000bb98:0x3:0x0] [index]: 00001 [logid]: [0x24000c368:0x2:0x0] lustre-MDT0000-osp-MDT0001 [catalog]: [0x200034fc1:0x3:0x0] [index]: 00001 [logid]: [0x200034fc2:0x2:0x0] [index]: 00002 [logid]: [0x200034fc2:0x3:0x0] Stopping /mnt/lustre-mds2 (opts:) on oleg328-server Failover mds2 to oleg328-server Starting mds2: -o localrecov -o abort_recovery /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0001 pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 find: '/mnt/lustre/d100d.replay-single': No such file or directory find: '/mnt/lustre/d100d.replay-single': No such file or directory PASS 100d (25s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 100e: DNE: create striped dir on MDT0 and MDT1, fail MDT0, MDT1 ========================================================== 05:41:11 (1713433271) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2988 1284700 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2548 1285140 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18172 3588848 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1772 3605248 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19944 7194096 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2988 1284700 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2548 1285140 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18172 3588848 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1772 3605248 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19944 7194096 1% /mnt/lustre lmv_stripe_count: 2 lmv_stripe_offset: 0 lmv_hash_type: crush mdtidx FID[seq:oid:ver] 0 [0x200034fc0:0x36:0x0] 1 [0x24000b3c9:0x36:0x0] Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server Failing mds2 on oleg328-server Stopping /mnt/lustre-mds2 (opts:) on oleg328-server 05:41:18 (1713433278) shut down Failover mds1 to oleg328-server mount facets: mds1 Failover mds2 to oleg328-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0001 Started lustre-MDT0000 05:41:44 (1713433304) targets are mounted 05:41:44 (1713433304) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid,mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec lmv_stripe_count: 2 lmv_stripe_offset: 0 lmv_hash_type: crush mdtidx FID[seq:oid:ver] 0 [0x200034fc0:0x36:0x0] 1 [0x24000b3c9:0x36:0x0] PASS 100e (42s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 101: Shouldn't reassign precreated objs to other files after recovery ========================================================== 05:41:55 (1713433315) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3032 1284656 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2584 1285104 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18172 3588848 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1772 3605248 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19944 7194096 1% /mnt/lustre Stopping /mnt/lustre-mds1 (opts:) on oleg328-server Failover mds1 to oleg328-server Starting mds1: -o localrecov -o abort_recovery /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 PASS 101 (37s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 102a: check resend (request lost) with multiple modify RPCs in flight ========================================================== 05:42:33 (1713433353) creating 7 files ... fail_loc=0x159 launch 7 chmod in parallel (05:42:33) ... fail_loc=0 done (05:42:49) /mnt/lustre/d102a.replay-single/file-1 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-2 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-3 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-4 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-5 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-6 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-7 has perms 0600 OK PASS 102a (18s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 102b: check resend (reply lost) with multiple modify RPCs in flight ========================================================== 05:42:52 (1713433372) creating 7 files ... fail_loc=0x15a launch 7 chmod in parallel (05:42:53) ... fail_loc=0 done (05:43:09) /mnt/lustre/d102b.replay-single/file-1 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-2 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-3 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-4 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-5 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-6 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-7 has perms 0600 OK PASS 102b (18s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 102c: check replay w/o reconstruction with multiple mod RPCs in flight ========================================================== 05:43:12 (1713433392) creating 7 files ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3196 1284492 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2656 1285032 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18172 3580644 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1772 3597100 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19944 7177744 1% /mnt/lustre fail_loc=0x15a launch 7 chmod in parallel (05:43:15) ... fail_loc=0 Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 05:43:17 (1713433397) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 05:43:32 (1713433412) targets are mounted 05:43:32 (1713433412) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec done (05:43:38) /mnt/lustre/d102c.replay-single/file-1 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-2 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-3 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-4 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-5 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-6 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-7 has perms 0600 OK PASS 102c (27s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 102d: check replay & reconstruction with multiple mod RPCs in flight ========================================================== 05:43:40 (1713433420) creating 7 files ... fail_loc=0x15a launch 7 chmod in parallel (05:43:41) ... fail_loc=0 Failing mds2 on oleg328-server Stopping /mnt/lustre-mds2 (opts:) on oleg328-server 05:43:44 (1713433424) shut down Failover mds2 to oleg328-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0001 05:43:57 (1713433437) targets are mounted 05:43:57 (1713433437) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec done (05:44:02) /mnt/lustre/d102d.replay-single/file-1 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-2 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-3 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-4 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-5 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-6 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-7 has perms 0600 OK PASS 102d (24s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 103: Check otr_next_id overflow ==== 05:44:05 (1713433445) fail_loc=0x80000162 total: 30 open/close in 0.09 seconds: 332.26 ops/second Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 05:44:07 (1713433447) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 05:44:20 (1713433460) targets are mounted 05:44:20 (1713433460) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 103 (22s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 110a: DNE: create striped dir, fail MDT1 ========================================================== 05:44:29 (1713433469) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3144 1284544 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2656 1285032 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18172 3580644 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1772 3597100 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19944 7177744 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 05:44:34 (1713433474) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 05:44:48 (1713433488) targets are mounted 05:44:48 (1713433488) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d110a.replay-single/striped_dir has type dir OK PASS 110a (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 110b: DNE: create striped dir, fail MDT1 and client ========================================================== 05:44:57 (1713433497) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3152 1284536 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2664 1285024 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18172 3580644 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1772 3605248 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19944 7185892 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 05:45:00 (1713433500) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 05:45:15 (1713433515) targets are mounted 05:45:15 (1713433515) facet_failover done pdsh@oleg328-client: oleg328-client: ssh exited with exit code 95 oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid pdsh@oleg328-client: oleg328-client: ssh exited with exit code 95 Starting client: oleg328-client.virtnet: -o user_xattr,flock oleg328-server@tcp:/lustre /mnt/lustre /mnt/lustre/d110b.replay-single/striped_dir has type dir OK PASS 110b (95s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 110c: DNE: create striped dir, fail MDT2 ========================================================== 05:46:34 (1713433594) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3156 1284532 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2664 1285024 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18172 3588848 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1772 3605248 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19944 7194096 1% /mnt/lustre Failing mds2 on oleg328-server Stopping /mnt/lustre-mds2 (opts:) on oleg328-server 05:46:37 (1713433597) shut down Failover mds2 to oleg328-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0001 05:46:51 (1713433611) targets are mounted 05:46:51 (1713433611) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d110c.replay-single/striped_dir has type dir OK PASS 110c (25s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 110d: DNE: create striped dir, fail MDT2 and client ========================================================== 05:47:01 (1713433621) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3196 1284492 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2708 1284980 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18172 3588848 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1772 3605248 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19944 7194096 1% /mnt/lustre Failing mds2 on oleg328-server Stopping /mnt/lustre-mds2 (opts:) on oleg328-server 05:47:05 (1713433625) shut down Failover mds2 to oleg328-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0001 05:47:19 (1713433639) targets are mounted 05:47:19 (1713433639) facet_failover done pdsh@oleg328-client: oleg328-client: ssh exited with exit code 95 oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid pdsh@oleg328-client: oleg328-client: ssh exited with exit code 95 Starting client: oleg328-client.virtnet: -o user_xattr,flock oleg328-server@tcp:/lustre /mnt/lustre /mnt/lustre/d110d.replay-single/striped_dir has type dir OK PASS 110d (93s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 110e: DNE: create striped dir, uncommit on MDT2, fail client/MDT1/MDT2 ========================================================== 05:48:36 (1713433716) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3144 1284544 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2656 1285032 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18172 3588848 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1772 3605248 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19944 7194096 1% /mnt/lustre df:/mnt/lustre Not a Lustre filesystem Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server Failing mds2 on oleg328-server Stopping /mnt/lustre-mds2 (opts:) on oleg328-server 05:48:43 (1713433723) shut down Failover mds1 to oleg328-server mount facets: mds1 Failover mds2 to oleg328-server mount facets: mds2 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0001 Started lustre-MDT0000 05:49:08 (1713433748) targets are mounted 05:49:08 (1713433748) facet_failover done pdsh@oleg328-client: oleg328-client: ssh exited with exit code 95 oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid,mdc.lustre-MDT0001-mdc-*.mds_server_uuid pdsh@oleg328-client: oleg328-client: ssh exited with exit code 95 Starting client: oleg328-client.virtnet: -o user_xattr,flock oleg328-server@tcp:/lustre /mnt/lustre /mnt/lustre/d110e.replay-single/striped_dir has type dir OK PASS 110e (107s) debug_raw_pointers=0 debug_raw_pointers=0 SKIP: replay-single test_110f skipping excluded test 110f debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 110g: DNE: create striped dir, uncommit on MDT1, fail client/MDT1/MDT2 ========================================================== 05:50:25 (1713433825) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3184 1284504 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2692 1284996 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18172 3588848 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1772 3605248 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19944 7194096 1% /mnt/lustre df:/mnt/lustre Not a Lustre filesystem Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server Failing mds2 on oleg328-server Stopping /mnt/lustre-mds2 (opts:) on oleg328-server 05:50:38 (1713433838) shut down Failover mds1 to oleg328-server mount facets: mds1 Failover mds2 to oleg328-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 Started lustre-MDT0001 05:50:58 (1713433858) targets are mounted 05:50:58 (1713433858) facet_failover done pdsh@oleg328-client: oleg328-client: ssh exited with exit code 95 oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid,mdc.lustre-MDT0001-mdc-*.mds_server_uuid pdsh@oleg328-client: oleg328-client: ssh exited with exit code 95 Starting client: oleg328-client.virtnet: -o user_xattr,flock oleg328-server@tcp:/lustre /mnt/lustre /mnt/lustre/d110g.replay-single/striped_dir has type dir OK PASS 110g (109s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 111a: DNE: unlink striped dir, fail MDT1 ========================================================== 05:52:16 (1713433936) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3220 1284468 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2668 1285020 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18172 3588848 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1772 3605248 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19944 7194096 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 05:52:20 (1713433940) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 05:52:35 (1713433955) targets are mounted 05:52:35 (1713433955) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/d111a.replay-single/striped_dir: No such file or directory PASS 111a (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 111b: DNE: unlink striped dir, fail MDT2 ========================================================== 05:52:44 (1713433964) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3164 1284524 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2676 1285012 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18172 3588848 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1772 3605248 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19944 7194096 1% /mnt/lustre Failing mds2 on oleg328-server Stopping /mnt/lustre-mds2 (opts:) on oleg328-server 05:52:48 (1713433968) shut down Failover mds2 to oleg328-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0001 05:53:02 (1713433982) targets are mounted 05:53:02 (1713433982) facet_failover done pdsh@oleg328-client: oleg328-client: ssh exited with exit code 95 oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid pdsh@oleg328-client: oleg328-client: ssh exited with exit code 95 Starting client: oleg328-client.virtnet: -o user_xattr,flock oleg328-server@tcp:/lustre /mnt/lustre Can't lstat /mnt/lustre/d111b.replay-single/striped_dir: No such file or directory PASS 111b (92s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 111c: DNE: unlink striped dir, uncommit on MDT1, fail client/MDT1/MDT2 ========================================================== 05:54:18 (1713434058) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3208 1284480 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2672 1285016 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18172 3588848 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1772 3605248 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19944 7194096 1% /mnt/lustre df:/mnt/lustre Not a Lustre filesystem Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server Failing mds2 on oleg328-server Stopping /mnt/lustre-mds2 (opts:) on oleg328-server 05:54:29 (1713434069) shut down Failover mds1 to oleg328-server mount facets: mds1 Failover mds2 to oleg328-server mount facets: mds2 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 Started lustre-MDT0001 05:54:50 (1713434090) targets are mounted 05:54:50 (1713434090) facet_failover done pdsh@oleg328-client: oleg328-client: ssh exited with exit code 95 oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid,mdc.lustre-MDT0001-mdc-*.mds_server_uuid pdsh@oleg328-client: oleg328-client: ssh exited with exit code 95 Starting client: oleg328-client.virtnet: -o user_xattr,flock oleg328-server@tcp:/lustre /mnt/lustre Can't lstat /mnt/lustre/d111c.replay-single/striped_dir: No such file or directory PASS 111c (108s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 111d: DNE: unlink striped dir, uncommit on MDT2, fail client/MDT1/MDT2 ========================================================== 05:56:08 (1713434168) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3236 1284452 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2672 1285016 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18172 3588848 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1772 3605248 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19944 7194096 1% /mnt/lustre df:/mnt/lustre Not a Lustre filesystem Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server Failing mds2 on oleg328-server Stopping /mnt/lustre-mds2 (opts:) on oleg328-server 05:56:15 (1713434175) shut down Failover mds1 to oleg328-server mount facets: mds1 Failover mds2 to oleg328-server mount facets: mds2 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 Started lustre-MDT0001 05:56:41 (1713434201) targets are mounted 05:56:41 (1713434201) facet_failover done pdsh@oleg328-client: oleg328-client: ssh exited with exit code 95 oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid,mdc.lustre-MDT0001-mdc-*.mds_server_uuid pdsh@oleg328-client: oleg328-client: ssh exited with exit code 95 Starting client: oleg328-client.virtnet: -o user_xattr,flock oleg328-server@tcp:/lustre /mnt/lustre Can't lstat /mnt/lustre/d111d.replay-single/striped_dir: No such file or directory PASS 111d (108s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 111e: DNE: unlink striped dir, uncommit on MDT2, fail MDT1/MDT2 ========================================================== 05:57:58 (1713434278) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3164 1284524 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2708 1284980 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18172 3588848 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1772 3605248 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19944 7194096 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3212 1284476 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2700 1284988 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18172 3588848 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1772 3605248 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19944 7194096 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server Failing mds2 on oleg328-server Stopping /mnt/lustre-mds2 (opts:) on oleg328-server 05:58:06 (1713434286) shut down Failover mds1 to oleg328-server mount facets: mds1 Failover mds2 to oleg328-server mount facets: mds2 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 Started lustre-MDT0001 05:58:32 (1713434312) targets are mounted 05:58:32 (1713434312) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid,mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/d111e.replay-single/striped_dir: No such file or directory PASS 111e (52s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 111f: DNE: unlink striped dir, uncommit on MDT1, fail MDT1/MDT2 ========================================================== 05:58:52 (1713434332) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3252 1284436 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2744 1284944 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18172 3588848 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1772 3605248 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19944 7194096 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3244 1284444 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2736 1284952 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18172 3588848 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1772 3605248 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19944 7194096 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server Failing mds2 on oleg328-server Stopping /mnt/lustre-mds2 (opts:) on oleg328-server 05:59:06 (1713434346) shut down Failover mds1 to oleg328-server mount facets: mds1 Failover mds2 to oleg328-server mount facets: mds2 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 Started lustre-MDT0001 05:59:26 (1713434366) targets are mounted 05:59:26 (1713434366) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid,mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/d111f.replay-single/striped_dir: No such file or directory PASS 111f (43s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 111g: DNE: unlink striped dir, fail MDT1/MDT2 ========================================================== 05:59:37 (1713434377) UUID Inodes IUsed IFree IUse% Mounted on lustre-MDT0000_UUID 1024000 555 1023445 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1024000 323 1023677 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 262144 557 261587 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 262144 486 261658 1% /mnt/lustre[OST:1] filesystem_summary: 524123 878 523245 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3180 1284508 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2672 1285016 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18172 3588848 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1772 3605248 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19944 7194096 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3180 1284508 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2672 1285016 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18172 3588848 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1772 3605248 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19944 7194096 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server Failing mds2 on oleg328-server Stopping /mnt/lustre-mds2 (opts:) on oleg328-server 05:59:46 (1713434386) shut down Failover mds1 to oleg328-server mount facets: mds1 Failover mds2 to oleg328-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 Started lustre-MDT0001 06:00:12 (1713434412) targets are mounted 06:00:12 (1713434412) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid,mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/d111g.replay-single/striped_dir: No such file or directory PASS 111g (53s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112a: DNE: cross MDT rename, fail MDT1 ========================================================== 06:00:32 (1713434432) SKIP: replay-single test_112a needs >= 4 MDTs SKIP 112a (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112b: DNE: cross MDT rename, fail MDT2 ========================================================== 06:00:35 (1713434435) SKIP: replay-single test_112b needs >= 4 MDTs SKIP 112b (2s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112c: DNE: cross MDT rename, fail MDT3 ========================================================== 06:00:39 (1713434439) SKIP: replay-single test_112c needs >= 4 MDTs SKIP 112c (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112d: DNE: cross MDT rename, fail MDT4 ========================================================== 06:00:42 (1713434442) SKIP: replay-single test_112d needs >= 4 MDTs SKIP 112d (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112e: DNE: cross MDT rename, fail MDT1 and MDT2 ========================================================== 06:00:45 (1713434445) SKIP: replay-single test_112e needs >= 4 MDTs SKIP 112e (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112f: DNE: cross MDT rename, fail MDT1 and MDT3 ========================================================== 06:00:48 (1713434448) SKIP: replay-single test_112f needs >= 4 MDTs SKIP 112f (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112g: DNE: cross MDT rename, fail MDT1 and MDT4 ========================================================== 06:00:52 (1713434452) SKIP: replay-single test_112g needs >= 4 MDTs SKIP 112g (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112h: DNE: cross MDT rename, fail MDT2 and MDT3 ========================================================== 06:00:55 (1713434455) SKIP: replay-single test_112h needs >= 4 MDTs SKIP 112h (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112i: DNE: cross MDT rename, fail MDT2 and MDT4 ========================================================== 06:00:59 (1713434459) SKIP: replay-single test_112i needs >= 4 MDTs SKIP 112i (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112j: DNE: cross MDT rename, fail MDT3 and MDT4 ========================================================== 06:01:02 (1713434462) SKIP: replay-single test_112j needs >= 4 MDTs SKIP 112j (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112k: DNE: cross MDT rename, fail MDT1,MDT2,MDT3 ========================================================== 06:01:06 (1713434466) SKIP: replay-single test_112k needs >= 4 MDTs SKIP 112k (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112l: DNE: cross MDT rename, fail MDT1,MDT2,MDT4 ========================================================== 06:01:09 (1713434469) SKIP: replay-single test_112l needs >= 4 MDTs SKIP 112l (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112m: DNE: cross MDT rename, fail MDT1,MDT3,MDT4 ========================================================== 06:01:12 (1713434472) SKIP: replay-single test_112m needs >= 4 MDTs SKIP 112m (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112n: DNE: cross MDT rename, fail MDT2,MDT3,MDT4 ========================================================== 06:01:15 (1713434475) SKIP: replay-single test_112n needs >= 4 MDTs SKIP 112n (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 115: failover for create/unlink striped directory ========================================================== 06:01:18 (1713434478) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3176 1284512 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2664 1285024 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18172 3588848 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1772 3605248 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19944 7194096 1% /mnt/lustre striped dir -i1 -c2 -H crush2 /mnt/lustre/d115.replay-single/test_0 striped dir -i1 -c2 -H all_char /mnt/lustre/d115.replay-single/test_1 striped dir -i1 -c2 -H fnv_1a_64 /mnt/lustre/d115.replay-single/test_2 striped dir -i1 -c2 -H fnv_1a_64 /mnt/lustre/d115.replay-single/test_3 striped dir -i1 -c2 -H all_char /mnt/lustre/d115.replay-single/test_4 Failing mds2 on oleg328-server Stopping /mnt/lustre-mds2 (opts:) on oleg328-server 06:01:22 (1713434482) shut down Failover mds2 to oleg328-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0001 06:01:36 (1713434496) targets are mounted 06:01:36 (1713434496) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3208 1284480 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2768 1284920 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18172 3588848 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1772 3605248 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19944 7194096 1% /mnt/lustre striped dir -i0 -c2 -H crush /mnt/lustre/d115.replay-single/test_0 striped dir -i0 -c2 -H crush2 /mnt/lustre/d115.replay-single/test_1 striped dir -i0 -c2 -H fnv_1a_64 /mnt/lustre/d115.replay-single/test_2 striped dir -i0 -c2 -H fnv_1a_64 /mnt/lustre/d115.replay-single/test_3 striped dir -i0 -c2 -H fnv_1a_64 /mnt/lustre/d115.replay-single/test_4 Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 06:01:47 (1713434507) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 06:02:01 (1713434521) targets are mounted 06:02:01 (1713434521) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 115 (51s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 116a: large update log master MDT recovery ========================================================== 06:02:11 (1713434531) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3268 1284420 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2796 1284892 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18172 3588848 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1772 3605248 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19944 7194096 1% /mnt/lustre fail_loc=0x80001702 Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 06:02:16 (1713434536) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 06:02:31 (1713434551) targets are mounted 06:02:31 (1713434551) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d116a.replay-single/striped_dir has type dir OK PASS 116a (27s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 116b: large update log slave MDT recovery ========================================================== 06:02:40 (1713434560) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3320 1284368 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2968 1284720 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18172 3588848 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1772 3605248 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19944 7194096 1% /mnt/lustre fail_loc=0x80001702 Failing mds2 on oleg328-server Stopping /mnt/lustre-mds2 (opts:) on oleg328-server 06:02:45 (1713434565) shut down Failover mds2 to oleg328-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0001 06:02:59 (1713434579) targets are mounted 06:02:59 (1713434579) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d116b.replay-single/striped_dir has type dir OK PASS 116b (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 117: DNE: cross MDT unlink, fail MDT1 and MDT2 ========================================================== 06:03:08 (1713434588) SKIP: replay-single test_117 needs >= 4 MDTs SKIP 117 (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 118: invalidate osp update will not cause update log corruption ========================================================== 06:03:11 (1713434591) fail_loc=0x1705 lfs setdirstripe: dirstripe error on '/mnt/lustre/d118.replay-single/striped_dir': Input/output error lfs setdirstripe: cannot create dir '/mnt/lustre/d118.replay-single/striped_dir': Input/output error lfs setdirstripe: dirstripe error on '/mnt/lustre/d118.replay-single/striped_dir1': Input/output error lfs setdirstripe: cannot create dir '/mnt/lustre/d118.replay-single/striped_dir1': Input/output error fail_loc=0x0 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3416 1284272 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2848 1284840 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18172 3588848 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1772 3605248 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19944 7194096 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 06:03:16 (1713434596) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 06:03:30 (1713434610) targets are mounted 06:03:30 (1713434610) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 118 (27s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 119: timeout of normal replay does not cause DNE replay fails ========================================================== 06:03:40 (1713434620) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3484 1284204 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2908 1284780 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18172 3588848 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1772 3605248 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19944 7194096 1% /mnt/lustre Stopping /mnt/lustre-mds1 (opts:) on oleg328-server Failover mds1 to oleg328-server fail_loc=0x80000714 fail_val=65 Starting mds1: -o localrecov -o recovery_time_hard=60 /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg328-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 68 sec mdt.lustre-MDT0000.recovery_time_hard=180 PASS 119 (83s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 120: DNE fail abort should stop both normal and DNE replay ========================================================== 06:05:05 (1713434705) Replay barrier on lustre-MDT0000 Stopping /mnt/lustre-mds1 (opts:) on oleg328-server Failover mds1 to oleg328-server Starting mds1: -o localrecov -o abort_recovery /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 find: '/mnt/lustre/d120.replay-single': No such file or directory find: '/mnt/lustre/d120.replay-single': No such file or directory PASS 120 (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 121: lock replay timed out and race ========================================================== 06:05:33 (1713434733) multiop /mnt/lustre/f121.replay-single vs_s TMPPIPE=/tmp/multiop_open_wait_pipe.7515 Stopping /mnt/lustre-mds1 (opts:) on oleg328-server Failover mds1 to oleg328-server fail_loc=0x721 fail_val=0 at_max=0 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 pdsh@oleg328-client: oleg328-client: ssh exited with exit code 5 fail_loc=0x0 at_max=600 PASS 121 (198s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 130a: DoM file create (setstripe) replay ========================================================== 06:08:53 (1713434933) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3876 1283812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 3208 1284480 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18172 3588848 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1772 3605248 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19944 7194096 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 06:08:56 (1713434936) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 06:09:12 (1713434952) targets are mounted 06:09:12 (1713434952) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 130a (27s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 130b: DoM file create (inherited) replay ========================================================== 06:09:22 (1713434962) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3888 1283800 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 3208 1284480 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18172 3588848 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1772 3605248 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19944 7194096 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 06:09:27 (1713434967) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 06:09:43 (1713434983) targets are mounted 06:09:43 (1713434983) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 130b (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 131a: DoM file write lock replay === 06:09:53 (1713434993) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3888 1283800 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 3208 1284480 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18172 3588848 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1772 3605248 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19944 7194096 1% /mnt/lustre 1+0 records in 1+0 records out 8 bytes (8 B) copied, 0.00245217 s, 3.3 kB/s Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 06:09:58 (1713434998) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 06:10:13 (1713435013) targets are mounted 06:10:13 (1713435013) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 131a (28s) debug_raw_pointers=0 debug_raw_pointers=0 SKIP: replay-single test_131b skipping excluded test 131b debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 132a: PFL new component instantiate replay ========================================================== 06:10:24 (1713435024) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3892 1283796 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 3208 1284480 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18172 3588848 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1772 3605248 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19944 7194096 1% /mnt/lustre 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0233543 s, 44.9 MB/s /mnt/lustre/f132a.replay-single lcm_layout_gen: 3 lcm_mirror_count: 1 lcm_entry_count: 2 lcme_id: 1 lcme_mirror_id: 0 lcme_flags: init lcme_extent.e_start: 0 lcme_extent.e_end: 1048576 lmm_stripe_count: 1 lmm_stripe_size: 1048576 lmm_pattern: raid0 lmm_layout_gen: 0 lmm_stripe_offset: 0 lmm_objects: - 0: { l_ost_idx: 0, l_fid: [0x280000401:0x4b0a:0x0] } lcme_id: 2 lcme_mirror_id: 0 lcme_flags: init lcme_extent.e_start: 1048576 lcme_extent.e_end: EOF lmm_stripe_count: 2 lmm_stripe_size: 4194304 lmm_pattern: raid0 lmm_layout_gen: 0 lmm_stripe_offset: 1 lmm_objects: - 0: { l_ost_idx: 1, l_fid: [0x2c0000401:0x4ac2:0x0] } - 1: { l_ost_idx: 0, l_fid: [0x280000401:0x4b0b:0x0] } Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 06:10:29 (1713435029) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 06:10:44 (1713435044) targets are mounted 06:10:44 (1713435044) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f132a.replay-single lcm_layout_gen: 1 lcm_mirror_count: 1 lcm_entry_count: 2 lcme_id: 1 lcme_mirror_id: 0 lcme_flags: init lcme_extent.e_start: 0 lcme_extent.e_end: 1048576 lmm_stripe_count: 1 lmm_stripe_size: 1048576 lmm_pattern: raid0 lmm_layout_gen: 65535 lmm_stripe_offset: 0 lmm_objects: - 0: { l_ost_idx: 0, l_fid: [0x280000401:0x4b0a:0x0] } lcme_id: 2 lcme_mirror_id: 0 lcme_flags: init lcme_extent.e_start: 1048576 lcme_extent.e_end: EOF lmm_stripe_count: 2 lmm_stripe_size: 4194304 lmm_pattern: raid0 lmm_layout_gen: 65535 lmm_stripe_offset: 1 lmm_objects: - 0: { l_ost_idx: 1, l_fid: [0x2c0000401:0x4ac2:0x0] } - 1: { l_ost_idx: 0, l_fid: [0x280000401:0x4b0b:0x0] } PASS 132a (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 133: check resend of ongoing requests for lwp during failover ========================================================== 06:10:55 (1713435055) seq.srv-lustre-MDT0001.space=clear Starting client: oleg328-client.virtnet: -o user_xattr,flock oleg328-server@tcp:/lustre /mnt/lustre fail_val=700 fail_loc=0x80000123 Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 06:11:00 (1713435060) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 06:11:14 (1713435074) targets are mounted 06:11:14 (1713435074) facet_failover done PASS 133 (25s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 134: replay creation of a file created in a pool ========================================================== 06:11:21 (1713435081) Creating new pool oleg328-server: Pool lustre.pool_134 created Adding targets to pool oleg328-server: OST lustre-OST0001_UUID added to pool lustre.pool_134 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3976 1283712 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 3252 1284436 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18172 3588848 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 2796 3604224 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 20968 7193072 1% /mnt/lustre Failing mds1 on oleg328-server Stopping /mnt/lustre-mds1 (opts:) on oleg328-server 06:11:33 (1713435093) shut down Failover mds1 to oleg328-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-MDT0000 06:11:48 (1713435108) targets are mounted 06:11:48 (1713435108) facet_failover done oleg328-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Destroy the created pools: pool_134 lustre.pool_134 oleg328-server: OST lustre-OST0001_UUID removed from pool lustre.pool_134 oleg328-server: Pool lustre.pool_134 destroyed PASS 134 (40s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 135: Server failure in lock replay phase ========================================================== 06:12:03 (1713435123) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3956 1283732 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 3252 1284436 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18172 3588848 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 2796 3604224 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 20968 7193072 1% /mnt/lustre ldlm.cancel_unused_locks_before_replay=0 Stopping /mnt/lustre-ost1 (opts:) on oleg328-server Failover ost1 to oleg328-server oleg328-server: oleg328-server.virtnet: executing load_module ../libcfs/libcfs/libcfs fail_loc=0x32d fail_val=20 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-OST0000 oleg328-client.virtnet: executing wait_import_state_mount REPLAY_LOCKS osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in REPLAY_LOCKS state after 0 sec Stopping /mnt/lustre-ost1 (opts:) on oleg328-server Failover ost1 to oleg328-server oleg328-server: oleg328-server.virtnet: executing load_module ../libcfs/libcfs/libcfs fail_loc=0 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 End of sync oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-OST0000 Stopping /mnt/lustre-ost1 (opts:-f) on oleg328-server Stopping /mnt/lustre-ost2 (opts:-f) on oleg328-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-OST0000 Starting ost2: -o localrecov /dev/mapper/ost2_flakey /mnt/lustre-ost2 seq.cli-lustre-OST0001-super.width=65536 oleg328-server: oleg328-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg328-client: oleg328-server: ssh exited with exit code 1 Started lustre-OST0001 pdsh@oleg328-client: oleg328-client: ssh exited with exit code 5 ldlm.cancel_unused_locks_before_replay=1 PASS 135 (72s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 136: MDS to disconnect all OSPs first, then cleanup ldlm ========================================================== 06:13:17 (1713435197) SKIP: replay-single test_136 needs > 2 MDTs SKIP 136 (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 200: Dropping one OBD_PING should not cause disconnect ========================================================== 06:13:20 (1713435200) SKIP: replay-single test_200 Need remote client SKIP 200 (1s) debug_raw_pointers=0 debug_raw_pointers=0 == replay-single test complete, duration 8711 sec ======== 06:13:22 (1713435202) === replay-single: start cleanup 06:13:22 (1713435202) === === replay-single: finish cleanup 06:13:26 (1713435206) ===