-----============= acceptance-small: replay-single ============----- Thu Apr 18 18:51:15 EDT 2024 excepting tests: 110f 131b 59 36 === replay-single: start setup 18:51:19 (1713480679) === oleg154-client.virtnet: executing check_config_client /mnt/lustre oleg154-client.virtnet: Checking config lustre mounted on /mnt/lustre Checking servers environments Checking clients oleg154-client.virtnet environments Using TIMEOUT=20 osc.lustre-OST0000-osc-ffff88012a457800.idle_timeout=debug osc.lustre-OST0001-osc-ffff88012a457800.idle_timeout=debug disable quota as required oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all osd-ldiskfs.track_declares_assert=1 === replay-single: finish setup 18:51:25 (1713480685) === debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 0a: empty replay =================== 18:51:26 (1713480686) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1772 1285916 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1612 1286076 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1524 3605496 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1524 3605496 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3048 7210992 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 18:51:30 (1713480690) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 18:51:43 (1713480703) targets are mounted 18:51:43 (1713480703) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 0a (24s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 0b: ensure object created after recover exists. (3284) ========================================================== 18:51:52 (1713480712) Failing ost1 on oleg154-server Stopping /mnt/lustre-ost1 (opts:) on oleg154-server 18:51:53 (1713480713) shut down Failover ost1 to oleg154-server mount facets: ost1 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-OST0000 18:52:06 (1713480726) targets are mounted 18:52:06 (1713480726) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec total: 20 open/close in 0.07 seconds: 289.50 ops/second - unlinked 0 (time 1713480729 ; total 0 ; last 0) total: 20 unlinks in 0 seconds: inf unlinks/second PASS 0b (18s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 0c: check replay-barrier =========== 18:52:12 (1713480732) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1768 1285920 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1612 1286076 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1540 3605480 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1524 3605496 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3064 7210976 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 18:52:15 (1713480735) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 18:52:28 (1713480748) targets are mounted 18:52:28 (1713480748) facet_failover done Starting client: oleg154-client.virtnet: -o user_xattr,flock oleg154-server@tcp:/lustre /mnt/lustre rm: cannot remove '/mnt/lustre/f0c.replay-single': No such file or directory PASS 0c (89s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 0d: expired recovery with no clients ========================================================== 18:53:43 (1713480823) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1768 1285920 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1612 1286076 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1540 3605480 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1524 3605496 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3064 7210976 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 18:53:47 (1713480827) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 18:54:01 (1713480841) targets are mounted 18:54:01 (1713480841) facet_failover done Starting client: oleg154-client.virtnet: -o user_xattr,flock oleg154-server@tcp:/lustre /mnt/lustre PASS 0d (91s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 1: simple create =================== 18:55:16 (1713480916) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1768 1285920 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1612 1286076 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1540 3605480 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1524 3605496 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3064 7210976 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 18:55:20 (1713480920) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 18:55:35 (1713480935) targets are mounted 18:55:35 (1713480935) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f1.replay-single has type file OK PASS 1 (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 2a: touch ========================== 18:55:45 (1713480945) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1768 1285920 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1612 1286076 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1540 3605480 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1524 3605496 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3064 7210976 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 18:55:49 (1713480949) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 18:56:04 (1713480964) targets are mounted 18:56:04 (1713480964) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f2a.replay-single has type file OK PASS 2a (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 2b: touch ========================== 18:56:13 (1713480973) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1768 1285920 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1612 1286076 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1540 3605480 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1524 3605496 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3064 7210976 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 18:56:18 (1713480978) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 18:56:33 (1713480993) targets are mounted 18:56:33 (1713480993) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f2b.replay-single has type file OK PASS 2b (27s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 2c: setstripe replay =============== 18:56:42 (1713481002) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1768 1285920 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1612 1286076 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1540 3605480 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1524 3605496 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3064 7210976 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 18:56:47 (1713481007) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 18:57:02 (1713481022) targets are mounted 18:57:02 (1713481022) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f2c.replay-single has type file OK PASS 2c (27s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 2d: setdirstripe replay ============ 18:57:10 (1713481030) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1768 1285920 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1612 1286076 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1540 3605480 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1524 3605496 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3064 7210976 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 18:57:13 (1713481033) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 18:57:28 (1713481048) targets are mounted 18:57:28 (1713481048) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d2d.replay-single has type dir OK PASS 2d (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 2e: O_CREAT|O_EXCL create replay === 18:57:38 (1713481058) fail_loc=0x8000013b UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1852 1285836 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1540 3605480 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1524 3605496 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3064 7210976 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 18:57:44 (1713481064) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 18:57:59 (1713481079) targets are mounted 18:57:59 (1713481079) facet_failover done Succeed in opening file "/mnt/lustre/f2e.replay-single"(flags=O_CREAT) oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f2e.replay-single has type file OK PASS 2e (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 3a: replay failed open(O_DIRECTORY) ========================================================== 18:58:08 (1713481088) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1852 1285836 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1540 3605480 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1524 3605496 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3064 7210976 1% /mnt/lustre Error in opening file "/mnt/lustre/f3a.replay-single"(flags=O_DIRECTORY) 20: Not a directory Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 18:58:11 (1713481091) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 18:58:25 (1713481105) targets are mounted 18:58:25 (1713481105) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f3a.replay-single has type file OK PASS 3a (24s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 3b: replay failed open -ENOMEM ===== 18:58:34 (1713481114) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1852 1285836 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1540 3605480 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1524 3605496 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3064 7210976 1% /mnt/lustre fail_loc=0x80000114 touch: cannot touch '/mnt/lustre/f3b.replay-single': Cannot allocate memory fail_loc=0 Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 18:58:39 (1713481119) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 18:58:54 (1713481134) targets are mounted 18:58:54 (1713481134) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/f3b.replay-single: No such file or directory PASS 3b (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 3c: replay failed open -ENOMEM ===== 18:59:04 (1713481144) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1852 1285836 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1540 3605480 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1524 3605496 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3064 7210976 1% /mnt/lustre fail_loc=0x80000128 touch: cannot touch '/mnt/lustre/f3c.replay-single': Cannot allocate memory fail_loc=0 Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 18:59:08 (1713481148) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 18:59:22 (1713481162) targets are mounted 18:59:22 (1713481162) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/f3c.replay-single: No such file or directory PASS 3c (25s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 4a: |x| 10 open(O_CREAT)s ========== 18:59:31 (1713481171) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1852 1285836 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1540 3605480 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1524 3605496 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3064 7210976 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 18:59:36 (1713481176) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 18:59:50 (1713481190) targets are mounted 18:59:50 (1713481190) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 4a (27s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 4b: |x| rm 10 files ================ 19:00:00 (1713481200) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1852 1285836 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1560 3605320 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1544 3605336 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3104 7210656 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:00:04 (1713481204) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:00:18 (1713481218) targets are mounted 19:00:18 (1713481218) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/f4b.replay-single-*: No such file or directory PASS 4b (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 5: |x| 220 open(O_CREAT) =========== 19:00:28 (1713481228) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1852 1285836 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1560 3605460 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1544 3605476 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3104 7210936 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:00:34 (1713481234) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:00:49 (1713481249) targets are mounted 19:00:49 (1713481249) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 5 (37s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 6a: mkdir + contained create ======= 19:01:06 (1713481266) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1884 1285804 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1996 3605024 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1544 3605476 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3540 7210500 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:01:11 (1713481271) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:01:26 (1713481286) targets are mounted 19:01:26 (1713481286) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d6a.replay-single has type dir OK /mnt/lustre/d6a.replay-single/f6a.replay-single has type file OK PASS 6a (29s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 6b: |X| rmdir ====================== 19:01:36 (1713481296) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1880 1285808 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1560 3605460 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1544 3605476 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3104 7210936 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:01:39 (1713481299) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:01:54 (1713481314) targets are mounted 19:01:54 (1713481314) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/d6b.replay-single: No such file or directory PASS 6b (25s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 7: mkdir |X| contained create ====== 19:02:03 (1713481323) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1880 1285808 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1560 3605460 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1544 3605476 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3104 7210936 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:02:07 (1713481327) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:02:22 (1713481342) targets are mounted 19:02:22 (1713481342) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d7.replay-single has type dir OK /mnt/lustre/d7.replay-single/f7.replay-single has type file OK PASS 7 (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 8: creat open |X| close ============ 19:02:31 (1713481351) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1560 3605460 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1544 3605476 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3104 7210936 1% /mnt/lustre multiop /mnt/lustre/f8.replay-single vmo_c TMPPIPE=/tmp/multiop_open_wait_pipe.7506 Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:02:35 (1713481355) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:02:50 (1713481370) targets are mounted 19:02:50 (1713481370) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f8.replay-single /mnt/lustre/f8.replay-single has type file OK PASS 8 (27s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 9: |X| create (same inum/gen) ====== 19:03:00 (1713481380) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1560 3605460 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1544 3605476 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3104 7210936 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:03:03 (1713481383) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:03:17 (1713481397) targets are mounted 19:03:17 (1713481397) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec old_inum == 144115305935798546, new_inum == 144115305935798546 old_inum and new_inum match PASS 9 (25s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 10: create |X| rename unlink ======= 19:03:26 (1713481406) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1560 3605460 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1544 3605476 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3104 7210936 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:03:30 (1713481410) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:03:45 (1713481425) targets are mounted 19:03:45 (1713481425) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/f10.replay-single: No such file or directory PASS 10 (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 11: create open write rename |X| create-old-name read ========================================================== 19:03:54 (1713481434) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1560 3605460 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1544 3605476 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3104 7210936 1% /mnt/lustre new old Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:03:58 (1713481438) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:04:13 (1713481453) targets are mounted 19:04:13 (1713481453) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec new old PASS 11 (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 12: open, unlink |X| close ========= 19:04:22 (1713481462) multiop /mnt/lustre/f12.replay-single vo_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7506 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1548 3605472 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3112 7210928 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:04:26 (1713481466) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:04:40 (1713481480) targets are mounted 19:04:40 (1713481480) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 12 (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 13: open chmod 0 |x| write close === 19:04:50 (1713481490) multiop /mnt/lustre/f13.replay-single vO_wc TMPPIPE=/tmp/multiop_open_wait_pipe.7506 /mnt/lustre/f13.replay-single has perms 00 OK UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1548 3605472 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3112 7210928 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:04:54 (1713481494) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:05:08 (1713481508) targets are mounted 19:05:08 (1713481508) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f13.replay-single has perms 00 OK /mnt/lustre/f13.replay-single has size 1 OK PASS 13 (27s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 14: open(O_CREAT), unlink |X| close ========================================================== 19:05:19 (1713481519) multiop /mnt/lustre/f14.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7506 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1548 3605472 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3112 7210928 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:05:24 (1713481524) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:05:38 (1713481538) targets are mounted 19:05:38 (1713481538) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 14 (27s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 15: open(O_CREAT), unlink |X| touch new, close ========================================================== 19:05:48 (1713481548) multiop /mnt/lustre/f15.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7506 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1548 3605472 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3112 7210928 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:05:53 (1713481553) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:06:08 (1713481568) targets are mounted 19:06:08 (1713481568) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 15 (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 16: |X| open(O_CREAT), unlink, touch new, unlink new ========================================================== 19:06:18 (1713481578) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1548 3605472 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3112 7210928 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:06:23 (1713481583) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:06:38 (1713481598) targets are mounted 19:06:38 (1713481598) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 16 (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 17: |X| open(O_CREAT), |replay| close ========================================================== 19:06:48 (1713481608) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1548 3605472 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3112 7210928 1% /mnt/lustre multiop /mnt/lustre/f17.replay-single vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.7506 Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:06:53 (1713481613) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:07:08 (1713481628) targets are mounted 19:07:08 (1713481628) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f17.replay-single has type file OK PASS 17 (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 18: open(O_CREAT), unlink, touch new, close, touch, unlink ========================================================== 19:07:18 (1713481638) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1548 3605472 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3112 7210928 1% /mnt/lustre multiop /mnt/lustre/f18.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7506 pid: 24956 will close Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:07:23 (1713481643) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:07:38 (1713481658) targets are mounted 19:07:38 (1713481658) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 18 (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 19: mcreate, open, write, rename === 19:07:48 (1713481668) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1548 3605472 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3112 7210928 1% /mnt/lustre old Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:07:52 (1713481672) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:08:06 (1713481686) targets are mounted 19:08:06 (1713481686) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec old PASS 19 (25s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 20a: |X| open(O_CREAT), unlink, replay, close (test mds_cleanup_orphans) ========================================================== 19:08:16 (1713481696) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1552 3605468 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3116 7210924 1% /mnt/lustre multiop /mnt/lustre/f20a.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7506 Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:08:20 (1713481700) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:08:35 (1713481715) targets are mounted 19:08:35 (1713481715) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 20a (27s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 20b: write, unlink, eviction, replay (test mds_cleanup_orphans) ========================================================== 19:08:44 (1713481724) /mnt/lustre/f20b.replay-single lmm_stripe_count: 1 lmm_stripe_size: 4194304 lmm_pattern: raid0 lmm_layout_gen: 0 lmm_stripe_offset: 0 obdidx objid objid group 0 1090 0x442 0x280000401 10000+0 records in 10000+0 records out 40960000 bytes (41 MB) copied, 1.39061 s, 29.5 MB/s Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:08:48 (1713481728) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:09:02 (1713481742) targets are mounted 19:09:02 (1713481742) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec affected facets: mds1 oleg154-server: oleg154-server.virtnet: executing _wait_recovery_complete *.lustre-MDT0000.recovery_status 1475 oleg154-server: *.lustre-MDT0000.recovery_status status: COMPLETE Waiting for MDT destroys to complete before 3116, after 3116 PASS 20b (27s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 20c: check that client eviction does not affect file content ========================================================== 19:09:13 (1713481753) multiop /mnt/lustre/f20c.replay-single vOw_c TMPPIPE=/tmp/multiop_open_wait_pipe.7506 -rw-r--r-- 1 root root 1 Apr 18 19:09 /mnt/lustre/f20c.replay-single PASS 20c (4s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 21: |X| open(O_CREAT), unlink touch new, replay, close (test mds_cleanup_orphans) ========================================================== 19:09:19 (1713481759) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605408 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1552 3605468 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3116 7210876 1% /mnt/lustre multiop /mnt/lustre/f21.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7506 Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:09:23 (1713481763) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:09:38 (1713481778) targets are mounted 19:09:38 (1713481778) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 21 (27s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 22: open(O_CREAT), |X| unlink, replay, close (test mds_cleanup_orphans) ========================================================== 19:09:48 (1713481788) multiop /mnt/lustre/f22.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7506 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605408 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3120 7210872 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:09:53 (1713481793) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:10:08 (1713481808) targets are mounted 19:10:08 (1713481808) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 22 (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 23: open(O_CREAT), |X| unlink touch new, replay, close (test mds_cleanup_orphans) ========================================================== 19:10:18 (1713481818) multiop /mnt/lustre/f23.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7506 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605408 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3120 7210872 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:10:23 (1713481823) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:10:38 (1713481838) targets are mounted 19:10:38 (1713481838) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 23 (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 24: open(O_CREAT), replay, unlink, close (test mds_cleanup_orphans) ========================================================== 19:10:48 (1713481848) multiop /mnt/lustre/f24.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7506 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605408 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3120 7210872 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:10:53 (1713481853) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:11:08 (1713481868) targets are mounted 19:11:08 (1713481868) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 24 (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 25: open(O_CREAT), unlink, replay, close (test mds_cleanup_orphans) ========================================================== 19:11:18 (1713481878) multiop /mnt/lustre/f25.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7506 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605408 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3120 7210872 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:11:22 (1713481882) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:11:37 (1713481897) targets are mounted 19:11:37 (1713481897) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 25 (27s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 26: |X| open(O_CREAT), unlink two, close one, replay, close one (test mds_cleanup_orphans) ========================================================== 19:11:47 (1713481907) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605408 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3120 7210872 1% /mnt/lustre multiop /mnt/lustre/f26.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7506 multiop /mnt/lustre/f26.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7506 Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:11:51 (1713481911) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:12:06 (1713481926) targets are mounted 19:12:06 (1713481926) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 26 (27s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 27: |X| open(O_CREAT), unlink two, replay, close two (test mds_cleanup_orphans) ========================================================== 19:12:16 (1713481936) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605408 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3120 7210872 1% /mnt/lustre multiop /mnt/lustre/f27.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7506 multiop /mnt/lustre/f27.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7506 Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:12:20 (1713481940) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:12:35 (1713481955) targets are mounted 19:12:35 (1713481955) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 27 (27s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 28: open(O_CREAT), |X| unlink two, close one, replay, close one (test mds_cleanup_orphans) ========================================================== 19:12:45 (1713481965) multiop /mnt/lustre/f28.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7506 multiop /mnt/lustre/f28.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7506 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605408 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3120 7210872 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:12:50 (1713481970) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:13:05 (1713481985) targets are mounted 19:13:05 (1713481985) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 28 (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 29: open(O_CREAT), |X| unlink two, replay, close two (test mds_cleanup_orphans) ========================================================== 19:13:15 (1713481995) multiop /mnt/lustre/f29.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7506 multiop /mnt/lustre/f29.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7506 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605408 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3120 7210872 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:13:19 (1713481999) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:13:35 (1713482015) targets are mounted 19:13:35 (1713482015) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 29 (27s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 30: open(O_CREAT) two, unlink two, replay, close two (test mds_cleanup_orphans) ========================================================== 19:13:44 (1713482024) multiop /mnt/lustre/f30.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7506 multiop /mnt/lustre/f30.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7506 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605408 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3120 7210872 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:13:49 (1713482029) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:14:05 (1713482045) targets are mounted 19:14:05 (1713482045) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 30 (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 31: open(O_CREAT) two, unlink one, |X| unlink one, close two (test mds_cleanup_orphans) ========================================================== 19:14:14 (1713482054) multiop /mnt/lustre/f31.replay-single-1 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7506 multiop /mnt/lustre/f31.replay-single-2 vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7506 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1876 1285812 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1688 1286000 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605408 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3120 7210872 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:14:19 (1713482059) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:14:34 (1713482074) targets are mounted 19:14:34 (1713482074) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 31 (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 32: close() notices client eviction; close() after client eviction ========================================================== 19:14:44 (1713482084) multiop /mnt/lustre/f32.replay-single vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.7506 multiop /mnt/lustre/f32.replay-single vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.7506 pdsh@oleg154-client: oleg154-client: ssh exited with exit code 5 PASS 32 (6s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 33a: fid seq shouldn't be reused after abort recovery ========================================================== 19:14:52 (1713482092) total: 10 open/close in 0.07 seconds: 137.30 ops/second Replay barrier on lustre-MDT0000 Stopping /mnt/lustre-mds1 (opts:) on oleg154-server Failover mds1 to oleg154-server Starting mds1: -o localrecov -o abort_recovery /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg154-client: error: invalid path '/mnt/lustre': Input/output error pdsh@oleg154-client: oleg154-client: ssh exited with exit code 5 first stat failed: 5 total: 10 open/close in 0.07 seconds: 133.33 ops/second PASS 33a (22s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 33b: test fid seq allocation ======= 19:15:16 (1713482116) fail_loc=0x1311 total: 10 open/close in 0.07 seconds: 151.44 ops/second Replay barrier on lustre-MDT0000 Stopping /mnt/lustre-mds1 (opts:) on oleg154-server Failover mds1 to oleg154-server Starting mds1: -o localrecov -o abort_recovery /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 total: 10 open/close in 0.08 seconds: 132.08 ops/second PASS 33b (21s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 34: abort recovery before client does replay (test mds_cleanup_orphans) ========================================================== 19:15:39 (1713482139) multiop /mnt/lustre/f34.replay-single vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.7506 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1908 1285780 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1720 1285968 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605408 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3120 7210872 1% /mnt/lustre Stopping /mnt/lustre-mds1 (opts:) on oleg154-server Failover mds1 to oleg154-server Starting mds1: -o localrecov -o abort_recovery /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 PASS 34 (21s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 35: test recovery from llog for unlink op ========================================================== 19:16:03 (1713482163) fail_loc=0x80000119 Stopping /mnt/lustre-mds1 (opts:) on oleg154-server Failover mds1 to oleg154-server Starting mds1: -o localrecov -o abort_recovery /dev/mapper/mds1_flakey /mnt/lustre-mds1 rm: cannot remove '/mnt/lustre/f35.replay-single': Input/output error oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 Can't lstat /mnt/lustre/f35.replay-single: No such file or directory PASS 35 (19s) debug_raw_pointers=0 debug_raw_pointers=0 SKIP: replay-single test_36 skipping ALWAYS excluded test 36 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 37: abort recovery before client does replay (test mds_cleanup_orphans for directories) ========================================================== 19:16:25 (1713482185) multiop /mnt/lustre/d37.replay-single/f37.replay-single vdD_c TMPPIPE=/tmp/multiop_open_wait_pipe.7506 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 1980 1285708 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1784 1285904 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605408 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3120 7210872 1% /mnt/lustre Stopping /mnt/lustre-mds1 (opts:) on oleg154-server Failover mds1 to oleg154-server Starting mds1: -o localrecov -o abort_recovery /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 PASS 37 (21s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 38: test recovery from unlink llog (test llog_gen_rec) ========================================================== 19:16:49 (1713482209) total: 800 open/close in 4.85 seconds: 164.98 ops/second - unlinked 0 (time 1713482216 ; total 0 ; last 0) total: 400 unlinks in 2 seconds: 200.000000 unlinks/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2108 1285580 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1816 1285872 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605408 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3120 7210872 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:17:02 (1713482222) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:17:18 (1713482238) targets are mounted 19:17:18 (1713482238) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec - unlinked 0 (time 1713482244 ; total 0 ; last 0) total: 400 unlinks in 3 seconds: 133.333328 unlinks/second Can't lstat /mnt/lustre/f38.replay-single-*: No such file or directory PASS 38 (42s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 39: test recovery from unlink llog (test llog_gen_rec) ========================================================== 19:17:33 (1713482253) total: 800 open/close in 4.70 seconds: 170.19 ops/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2108 1285580 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1816 1285872 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1564 3605408 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3120 7210872 1% /mnt/lustre - unlinked 0 (time 1713482263 ; total 0 ; last 0) total: 400 unlinks in 1 seconds: 400.000000 unlinks/second Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:17:46 (1713482266) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:18:01 (1713482281) targets are mounted 19:18:01 (1713482281) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec - unlinked 0 (time 1713482289 ; total 0 ; last 0) total: 400 unlinks in 2 seconds: 200.000000 unlinks/second Can't lstat /mnt/lustre/f39.replay-single-*: No such file or directory PASS 39 (42s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 41: read from a valid osc while other oscs are invalid ========================================================== 19:18:18 (1713482298) 1+0 records in 1+0 records out 4096 bytes (4.1 kB) copied, 0.00288816 s, 1.4 MB/s 1+0 records in 1+0 records out 4096 bytes (4.1 kB) copied, 0.00635194 s, 645 kB/s PASS 41 (3s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 42: recovery after ost failure ===== 19:18:23 (1713482303) total: 800 open/close in 4.87 seconds: 164.22 ops/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2136 1285552 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1816 1285872 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre - unlinked 0 (time 1713482314 ; total 0 ; last 0) total: 400 unlinks in 1 seconds: 400.000000 unlinks/second debug=-1 Failing ost1 on oleg154-server Stopping /mnt/lustre-ost1 (opts:) on oleg154-server 19:18:37 (1713482317) shut down Failover ost1 to oleg154-server mount facets: ost1 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-OST0000 19:18:53 (1713482333) targets are mounted 19:18:53 (1713482333) facet_failover done wait for MDS to timeout and recover debug=super ioctl neterror warning dlmtrace error emerg ha rpctrace vfstrace config console lfsck debug=super ioctl neterror warning dlmtrace error emerg ha rpctrace vfstrace config console lfsck - unlinked 0 (time 1713482375 ; total 0 ; last 0) total: 400 unlinks in 1 seconds: 400.000000 unlinks/second Can't lstat /mnt/lustre/f42.replay-single-*: No such file or directory PASS 42 (75s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 43: mds osc import failure during recovery; don't LBUG ========================================================== 19:19:41 (1713482381) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2192 1285496 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1816 1285872 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre fail_loc=0x80000204 Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:19:46 (1713482386) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:20:01 (1713482401) targets are mounted 19:20:01 (1713482401) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 43 (38s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 44a: race in target handle connect ========================================================== 19:20:21 (1713482421) at_max=40 1 of 10 (1713482422) service : cur 5 worst 5 (at 1713480641, 1783s ago) 4 4 4 4 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2088 1285600 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1816 1285872 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre 2 of 10 (1713482428) service : cur 5 worst 5 (at 1713480641, 1788s ago) 5 4 4 4 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2088 1285600 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1816 1285872 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre 3 of 10 (1713482434) service : cur 5 worst 5 (at 1713480641, 1794s ago) 5 4 4 4 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2088 1285600 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1816 1285872 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre 4 of 10 (1713482439) service : cur 5 worst 5 (at 1713480641, 1800s ago) 5 5 4 4 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2088 1285600 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1816 1285872 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre 5 of 10 (1713482445) service : cur 5 worst 5 (at 1713480641, 1805s ago) 5 5 4 4 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2088 1285600 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1816 1285872 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre 6 of 10 (1713482451) service : cur 5 worst 5 (at 1713480641, 1811s ago) 5 5 4 4 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2088 1285600 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1816 1285872 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre 7 of 10 (1713482457) service : cur 5 worst 5 (at 1713480641, 1817s ago) 5 5 4 4 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2088 1285600 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1816 1285872 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre 8 of 10 (1713482462) service : cur 5 worst 5 (at 1713480641, 1822s ago) 5 5 4 4 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2088 1285600 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1816 1285872 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre 9 of 10 (1713482468) service : cur 5 worst 5 (at 1713480641, 1828s ago) 5 5 4 4 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2088 1285600 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1816 1285872 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre 10 of 10 (1713482474) service : cur 5 worst 5 (at 1713480641, 1834s ago) 5 5 4 4 fail_loc=0x80000701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2088 1285600 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1816 1285872 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre fail_loc=0 at_max=600 PASS 44a (60s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 44b: race in target handle connect ========================================================== 19:21:23 (1713482483) 1 of 10 (1713482484) service : cur 5 worst 5 (at 1713480641, 1844s ago) 5 5 4 4 fail_loc=0x80000704 error: recover: Connection timed out Filesystem 1K-blocks Used Available Use% Mounted on 192.168.201.154@tcp:/lustre 7666232 3124 7210916 1% /mnt/lustre 2 of 10 (1713482525) service : cur 40 worst 40 (at 1713482525, 1s ago) 40 5 4 4 fail_loc=0x80000704 error: recover: Connection timed out Filesystem 1K-blocks Used Available Use% Mounted on 192.168.201.154@tcp:/lustre 7666232 3124 7210916 1% /mnt/lustre 3 of 10 (1713482545) service : cur 40 worst 40 (at 1713482525, 22s ago) 40 5 4 4 fail_loc=0x80000704 error: recover: Connection timed out Filesystem 1K-blocks Used Available Use% Mounted on 192.168.201.154@tcp:/lustre 7666232 3124 7210916 1% /mnt/lustre 4 of 10 (1713482566) service : cur 40 worst 40 (at 1713482525, 42s ago) 40 5 4 4 fail_loc=0x80000704 error: recover: Connection timed out Filesystem 1K-blocks Used Available Use% Mounted on 192.168.201.154@tcp:/lustre 7666232 3124 7210916 1% /mnt/lustre 5 of 10 (1713482587) service : cur 40 worst 40 (at 1713482525, 63s ago) 40 5 4 4 fail_loc=0x80000704 error: recover: Connection timed out Filesystem 1K-blocks Used Available Use% Mounted on 192.168.201.154@tcp:/lustre 7666232 3124 7210916 1% /mnt/lustre 6 of 10 (1713482607) service : cur 40 worst 40 (at 1713482525, 84s ago) 40 40 5 4 fail_loc=0x80000704 error: recover: Connection timed out Filesystem 1K-blocks Used Available Use% Mounted on 192.168.201.154@tcp:/lustre 7666232 3124 7210916 1% /mnt/lustre 7 of 10 (1713482628) service : cur 40 worst 40 (at 1713482525, 104s ago) 40 40 5 4 fail_loc=0x80000704 error: recover: Connection timed out Filesystem 1K-blocks Used Available Use% Mounted on 192.168.201.154@tcp:/lustre 7666232 3124 7210916 1% /mnt/lustre 8 of 10 (1713482649) service : cur 40 worst 40 (at 1713482525, 125s ago) 40 40 5 4 fail_loc=0x80000704 error: recover: Connection timed out Filesystem 1K-blocks Used Available Use% Mounted on 192.168.201.154@tcp:/lustre 7666232 3124 7210916 1% /mnt/lustre 9 of 10 (1713482669) service : cur 40 worst 40 (at 1713482525, 146s ago) 40 40 5 4 fail_loc=0x80000704 error: recover: Connection timed out Filesystem 1K-blocks Used Available Use% Mounted on 192.168.201.154@tcp:/lustre 7666232 3124 7210916 1% /mnt/lustre 10 of 10 (1713482690) service : cur 40 worst 40 (at 1713482525, 166s ago) 40 40 5 4 fail_loc=0x80000704 error: recover: Connection timed out Filesystem 1K-blocks Used Available Use% Mounted on 192.168.201.154@tcp:/lustre 7666232 3124 7210916 1% /mnt/lustre PASS 44b (229s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 44c: race in target handle connect ========================================================== 19:25:15 (1713482715) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2088 1285600 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1816 1285872 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre total: 100 create in 0.46 seconds: 217.19 ops/second fail_loc=0x80000712 Stopping /mnt/lustre-mds1 (opts:) on oleg154-server Failover mds1 to oleg154-server Starting mds1: -o localrecov -o abort_recovery /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 unlink(/mnt/lustre/f44c.replay-single-0) error: No such file or directory total: 0 unlinks in 0 seconds: -nan unlinks/second Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:25:36 (1713482736) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:25:50 (1713482750) targets are mounted 19:25:50 (1713482750) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec unlink(/mnt/lustre/f44c.replay-single-0) error: No such file or directory total: 0 unlinks in 0 seconds: -nan unlinks/second PASS 44c (43s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 45: Handle failed close ============ 19:25:59 (1713482759) multiop /mnt/lustre/f45.replay-single vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.7506 /mnt/lustre/f45.replay-single has type file OK PASS 45 (3s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 46: Don't leak file handle after open resend (3325) ========================================================== 19:26:04 (1713482764) fail_loc=0x122 fail_loc=0 Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:26:22 (1713482782) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:26:36 (1713482796) targets are mounted 19:26:36 (1713482796) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec lfs path2fid: cannot get fid for 'f46.replay-single': No such file or directory PASS 46 (40s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 47: MDS->OSC failure during precreate cleanup (2824) ========================================================== 19:26:46 (1713482806) total: 20 open/close in 0.15 seconds: 135.93 ops/second Failing ost1 on oleg154-server Stopping /mnt/lustre-ost1 (opts:) on oleg154-server 19:26:48 (1713482808) shut down Failover ost1 to oleg154-server mount facets: ost1 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-OST0000 19:27:02 (1713482822) targets are mounted 19:27:02 (1713482822) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec fail_loc=0x80000204 total: 20 open/close in 0.11 seconds: 188.89 ops/second - unlinked 0 (time 1713482888 ; total 0 ; last 0) total: 20 unlinks in 0 seconds: inf unlinks/second PASS 47 (84s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 48: MDS->OSC failure during precreate cleanup (2824) ========================================================== 19:28:12 (1713482892) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2120 1285568 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1848 1285840 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre total: 20 open/close in 0.12 seconds: 173.33 ops/second Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:28:16 (1713482896) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:28:31 (1713482911) targets are mounted 19:28:31 (1713482911) facet_failover done fail_loc=0x80000216 total: 20 open/close in 0.14 seconds: 138.67 ops/second - unlinked 0 (time 1713482974 ; total 0 ; last 0) total: 40 unlinks in 0 seconds: inf unlinks/second PASS 48 (84s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 50: Double OSC recovery, don't LASSERT (3812) ========================================================== 19:29:39 (1713482979) PASS 50 (8s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 52: time out lock replay (3764) ==== 19:29:49 (1713482989) multiop /mnt/lustre/f52.replay-single vs_s TMPPIPE=/tmp/multiop_open_wait_pipe.7506 fail_loc=0x80000157 Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:29:51 (1713482991) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:30:05 (1713483005) targets are mounted 19:30:05 (1713483005) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec fail_loc=0x0 fail_loc=0x0 PASS 52 (82s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53a: |X| close request while two MDC requests in flight ========================================================== 19:31:13 (1713483073) fail_loc=0x80000115 fail_loc=0 Replay barrier on lustre-MDT0000 Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:31:19 (1713483079) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:31:35 (1713483095) targets are mounted 19:31:35 (1713483095) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d53a.replay-single-1/f has type file OK /mnt/lustre/d53a.replay-single-2/f has type file OK PASS 53a (29s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53b: |X| open request while two MDC requests in flight ========================================================== 19:31:45 (1713483105) multiop /mnt/lustre/d53b.replay-single-1/f vO_c TMPPIPE=/tmp/multiop_open_wait_pipe.7506 fail_loc=0x80000107 fail_loc=0 Replay barrier on lustre-MDT0000 Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:31:51 (1713483111) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:32:06 (1713483126) targets are mounted 19:32:06 (1713483126) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d53b.replay-single-1/f has type file OK /mnt/lustre/d53b.replay-single-2/f has type file OK PASS 53b (29s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53c: |X| open request and close request while two MDC requests in flight ========================================================== 19:32:16 (1713483136) fail_loc=0x80000107 fail_loc=0x80000115 Replay barrier on lustre-MDT0000 Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:32:22 (1713483142) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:32:37 (1713483157) targets are mounted 19:32:37 (1713483157) facet_failover done fail_loc=0 /mnt/lustre/d53c.replay-single-1/f has type file OK /mnt/lustre/d53c.replay-single-2/f has type file OK PASS 53c (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53d: close reply while two MDC requests in flight ========================================================== 19:32:46 (1713483166) fail_loc=0x8000013b fail_loc=0 Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:32:50 (1713483170) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:33:03 (1713483183) targets are mounted 19:33:03 (1713483183) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d53d.replay-single-1/f has type file OK /mnt/lustre/d53d.replay-single-2/f has type file OK PASS 53d (25s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53e: |X| open reply while two MDC requests in flight ========================================================== 19:33:13 (1713483193) fail_loc=0x119 fail_loc=0 Replay barrier on lustre-MDT0000 Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:33:19 (1713483199) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:33:34 (1713483214) targets are mounted 19:33:34 (1713483214) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d53e.replay-single-1/f has type file OK /mnt/lustre/d53e.replay-single-2/f has type file OK PASS 53e (29s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53f: |X| open reply and close reply while two MDC requests in flight ========================================================== 19:33:44 (1713483224) fail_loc=0x119 fail_loc=0x8000013b Replay barrier on lustre-MDT0000 Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:33:50 (1713483230) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:34:05 (1713483245) targets are mounted 19:34:05 (1713483245) facet_failover done fail_loc=0 /mnt/lustre/d53f.replay-single-1/f has type file OK /mnt/lustre/d53f.replay-single-2/f has type file OK PASS 53f (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53g: |X| drop open reply and close request while close and open are both in flight ========================================================== 19:34:14 (1713483254) fail_loc=0x119 fail_loc=0x80000115 fail_loc=0 Replay barrier on lustre-MDT0000 Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:34:20 (1713483260) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:34:36 (1713483276) targets are mounted 19:34:36 (1713483276) facet_failover done /mnt/lustre/d53g.replay-single-1/f has type file OK /mnt/lustre/d53g.replay-single-2/f has type file OK PASS 53g (29s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 53h: open request and close reply while two MDC requests in flight ========================================================== 19:34:45 (1713483285) fail_loc=0x80000107 fail_loc=0x8000013b Replay barrier on lustre-MDT0000 Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:34:52 (1713483292) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:35:07 (1713483307) targets are mounted 19:35:07 (1713483307) facet_failover done fail_loc=0 /mnt/lustre/d53h.replay-single-1/f has type file OK /mnt/lustre/d53h.replay-single-2/f has type file OK PASS 53h (29s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 55: let MDS_CHECK_RESENT return the original return code instead of 0 ========================================================== 19:35:17 (1713483317) fail_loc=0x8000012b fail_loc=0x0 touch: rm: cannot remove '/mnt/lustre/f55.replay-single'cannot touch '/mnt/lustre/f55.replay-single': No such file or directory : No such file or directory PASS 55 (62s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 56: don't replay a symlink open request (3440) ========================================================== 19:36:21 (1713483381) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2128 1285560 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1848 1285840 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:36:26 (1713483386) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:36:42 (1713483402) targets are mounted 19:36:42 (1713483402) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 56 (38s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 57: test recovery from llog for setattr op ========================================================== 19:37:01 (1713483421) fail_loc=0x8000012c UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2128 1285560 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1848 1285840 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:37:07 (1713483427) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:37:22 (1713483442) targets are mounted 19:37:22 (1713483442) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec affected facets: mds1 oleg154-server: oleg154-server.virtnet: executing _wait_recovery_complete *.lustre-MDT0000.recovery_status 1475 oleg154-server: *.lustre-MDT0000.recovery_status status: COMPLETE Waiting for orphan cleanup... osp.lustre-OST0000-osc-MDT0000.old_sync_processed osp.lustre-OST0000-osc-MDT0001.old_sync_processed osp.lustre-OST0001-osc-MDT0000.old_sync_processed osp.lustre-OST0001-osc-MDT0001.old_sync_processed wait 40 secs maximumly for oleg154-server mds-ost sync done. /mnt/lustre/f57.replay-single has type file OK fail_loc=0x0 PASS 57 (33s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 58a: test recovery from llog for setattr op (test llog_gen_rec) ========================================================== 19:37:36 (1713483456) fail_loc=0x8000012c total: 2500 open/close in 9.47 seconds: 264.05 ops/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2424 1285264 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1848 1285840 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:37:52 (1713483472) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:38:07 (1713483487) targets are mounted 19:38:07 (1713483487) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec fail_loc=0x0 - unlinked 0 (time 1713483508 ; total 0 ; last 0) total: 2500 unlinks in 6 seconds: 416.666656 unlinks/second PASS 58a (60s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 58b: test replay of setxattr op ==== 19:38:39 (1713483519) Starting client: oleg154-client.virtnet: -o user_xattr,flock oleg154-server@tcp:/lustre /mnt/lustre2 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2408 1285280 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1848 1285840 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:38:44 (1713483524) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:38:59 (1713483539) targets are mounted 19:38:59 (1713483539) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Stopping client oleg154-client.virtnet /mnt/lustre2 (opts:) oleg154-client.virtnet: executing wait_import_state_mount FULL mgc.*.mgs_server_uuid mgc.*.mgs_server_uuid in FULL state after 0 sec PASS 58b (29s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 58c: resend/reconstruct setxattr op ========================================================== 19:39:11 (1713483551) Starting client: oleg154-client.virtnet: -o user_xattr,flock oleg154-server@tcp:/lustre /mnt/lustre2 fail_val=0 fail_loc=0x123 fail_loc=0 fail_loc=0x119 fail_loc=0 Stopping client oleg154-client.virtnet /mnt/lustre2 (opts:) PASS 58c (129s) debug_raw_pointers=0 debug_raw_pointers=0 SKIP: replay-single test_59 skipping ALWAYS excluded test 59 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 60: test llog post recovery init vs llog unlink ========================================================== 19:41:23 (1713483683) total: 200 open/close in 0.81 seconds: 247.66 ops/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2264 1285424 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 1848 1285840 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre - unlinked 0 (time 1713483688 ; total 0 ; last 0) total: 100 unlinks in 1 seconds: 100.000000 unlinks/second Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:41:30 (1713483690) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:41:45 (1713483705) targets are mounted 19:41:46 (1713483706) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec - unlinked 0 (time 1713483712 ; total 0 ; last 0) total: 100 unlinks in 0 seconds: inf unlinks/second PASS 60 (31s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 61a: test race llog recovery vs llog cleanup ========================================================== 19:41:56 (1713483716) total: 800 open/close in 3.19 seconds: 250.63 ops/second UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2348 1285340 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2020 1285668 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1556 3605464 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3124 7210916 1% /mnt/lustre - unlinked 0 (time 1713483725 ; total 0 ; last 0) total: 800 unlinks in 2 seconds: 400.000000 unlinks/second fail_val=0 fail_loc=0x80000221 Failing ost1 on oleg154-server Stopping /mnt/lustre-ost1 (opts:) on oleg154-server 19:42:09 (1713483729) shut down Failover ost1 to oleg154-server mount facets: ost1 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-OST0000 19:42:25 (1713483745) targets are mounted 19:42:25 (1713483745) facet_failover done Failing ost1 on oleg154-server Stopping /mnt/lustre-ost1 (opts:) on oleg154-server 19:42:36 (1713483756) shut down Failover ost1 to oleg154-server mount facets: ost1 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-OST0000 19:42:50 (1713483770) targets are mounted 19:42:50 (1713483770) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec fail_val=0 fail_loc=0x0 Can't lstat /mnt/lustre/d61a.replay-single/f61a.replay-single-*: No such file or directory PASS 61a (89s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 61b: test race mds llog sync vs llog cleanup ========================================================== 19:43:28 (1713483808) fail_loc=0x8000013a Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:43:30 (1713483810) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:43:44 (1713483824) targets are mounted 19:43:44 (1713483824) facet_failover done Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:43:56 (1713483836) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:44:10 (1713483850) targets are mounted 19:44:10 (1713483850) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec 1+0 records in 1+0 records out 4096 bytes (4.1 kB) copied, 0.00467535 s, 876 kB/s PASS 61b (49s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 61c: test race mds llog sync vs llog cleanup ========================================================== 19:44:19 (1713483859) fail_val=0 fail_loc=0x80000222 Failing ost1 on oleg154-server Stopping /mnt/lustre-ost1 (opts:) on oleg154-server 19:44:32 (1713483872) shut down Failover ost1 to oleg154-server mount facets: ost1 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-OST0000 19:44:47 (1713483887) targets are mounted 19:44:47 (1713483887) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec fail_val=0 fail_loc=0x0 PASS 61c (34s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 61d: error in llog_setup should cleanup the llog context correctly ========================================================== 19:44:55 (1713483895) Stopping /mnt/lustre-mds1 (opts:) on oleg154-server fail_loc=0x80000605 Starting mgs: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: mount.lustre: mount /dev/mapper/mds1_flakey at /mnt/lustre-mds1 failed: Operation not supported pdsh@oleg154-client: oleg154-server: ssh exited with exit code 95 Start of /dev/mapper/mds1_flakey on mgs failed 95 fail_loc=0 Starting mgs: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 PASS 61d (11s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 62: don't mis-drop resent replay === 19:45:08 (1713483908) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2340 1285348 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2004 1285684 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1568 3605452 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1560 3605460 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3128 7210912 1% /mnt/lustre total: 25 open/close in 0.13 seconds: 197.53 ops/second fail_loc=0x80000707 Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:45:13 (1713483913) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:45:28 (1713483928) targets are mounted 19:45:28 (1713483928) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec fail_loc=0 - unlinked 0 (time 1713483992 ; total 0 ; last 0) total: 25 unlinks in 0 seconds: inf unlinks/second PASS 62 (85s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 65a: AT: verify early replies ====== 19:46:35 (1713483995) at_history=8 at_history=8 debug=other fail_val=11000 fail_loc=0x8000050a 00000100:00001000:1.0:1713484026.050412:0:14690:0:(client.c:537:ptlrpc_at_recv_early_reply()) @@@ Early reply #1, new deadline in 35s (25s) req@ffff8800a79a9880 x1796714641742144/t0(0) o101->lustre-MDT0000-mdc-ffff8800b5819000@192.168.201.154@tcp:12/10 lens 664/66320 e 1 to 0 dl 1713484061 ref 2 fl Rpc:PQr/200/ffffffff rc 0/-1 job:'createmany.0' uid:0 gid:0 portal 12 : cur 36 worst 40 (at 1713482525, 1506s ago) 40 40 40 40 portal 29 : cur 5 worst 5 (at 1713480911, 3120s ago) 5 0 0 0 portal 23 : cur 5 worst 5 (at 1713480911, 3120s ago) 5 5 5 5 portal 30 : cur 5 worst 5 (at 1713480916, 3115s ago) 5 0 0 0 portal 17 : cur 5 worst 5 (at 1713481006, 3025s ago) 5 5 5 5 portal 13 : cur 5 worst 5 (at 1713481409, 2622s ago) 5 5 0 0 portal 12 : cur 5 worst 40 (at 1713482525, 1515s ago) 40 40 40 40 portal 29 : cur 5 worst 5 (at 1713480911, 3129s ago) 5 0 0 0 portal 23 : cur 5 worst 5 (at 1713480911, 3129s ago) 5 5 5 5 portal 30 : cur 5 worst 5 (at 1713480916, 3124s ago) 5 0 0 0 portal 17 : cur 5 worst 5 (at 1713481006, 3034s ago) 5 5 5 5 portal 13 : cur 5 worst 5 (at 1713481409, 2631s ago) 5 5 0 0 PASS 65a (47s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 65b: AT: verify early replies on packed reply / bulk ========================================================== 19:47:24 (1713484044) at_history=8 at_history=8 debug=other trace fail_val=11 fail_loc=0x224 fail_loc=0 00000100:00001000:0.0:1713484074.514317:0:1956:0:(client.c:537:ptlrpc_at_recv_early_reply()) @@@ Early reply #1, new deadline in 35s (25s) req@ffff88012bfd5180 x1796714641752192/t0(0) o4->lustre-OST0000-osc-ffff8800b5819000@192.168.201.154@tcp:6/4 lens 4584/448 e 1 to 0 dl 1713484109 ref 2 fl Rpc:Qr/200/ffffffff rc 0/-1 job:'multiop.0' uid:0 gid:0 debug=super ioctl neterror warning dlmtrace error emerg ha rpctrace vfstrace config console lfsck debug=super ioctl neterror warning dlmtrace error emerg ha rpctrace vfstrace config console lfsck portal 28 : cur 5 worst 5 (at 1713480911, 3169s ago) 5 5 5 5 portal 7 : cur 5 worst 5 (at 1713480912, 3168s ago) 5 5 5 5 portal 17 : cur 5 worst 5 (at 1713480974, 3106s ago) 5 5 0 0 portal 6 : cur 36 worst 36 (at 1713484079, 1s ago) 36 0 0 0 PASS 65b (38s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 66a: AT: verify MDT service time adjusts with no early replies ========================================================== 19:48:04 (1713484084) at_history=8 at_history=8 portal 12 : cur 5 worst 40 (at 1713482525, 1582s ago) 5 40 40 40 fail_val=5000 fail_loc=0x8000050a portal 12 : cur 5 worst 40 (at 1713482525, 1588s ago) 5 40 40 40 fail_val=10000 fail_loc=0x8000050a portal 12 : cur 36 worst 40 (at 1713482525, 1599s ago) 36 40 40 40 fail_loc=0 portal 12 : cur 5 worst 40 (at 1713482525, 1608s ago) 36 40 40 40 Current MDT timeout 5, worst 40 PASS 66a (51s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 66b: AT: verify net latency adjusts ========================================================== 19:48:57 (1713484137) at_history=8 at_history=8 fail_val=10 fail_loc=0x50c fail_loc=0 network timeout orig 5, cur 10, worst 10 PASS 66b (85s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 67a: AT: verify slow request processing doesn't induce reconnects ========================================================== 19:50:24 (1713484224) at_history=8 at_history=8 fail_val=400 fail_loc=0x50a fail_loc=0 0 osc reconnect attempts on gradual slow PASS 67a (74s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 67b: AT: verify instant slowdown doesn't induce reconnects ========================================================== 19:51:41 (1713484301) at_history=8 at_history=8 Creating to objid 5057 on ost lustre-OST0000... fail_val=20000 fail_loc=0x80000223 total: 18 open/close in 0.10 seconds: 187.62 ops/second Connected clients: oleg154-client.virtnet oleg154-client.virtnet service : cur 5 worst 5 (at 1713480652, 3674s ago) 1 1 1 1 phase 2 0 osc reconnect attempts on instant slow fail_loc=0x80000223 fail_loc=0 Connected clients: oleg154-client.virtnet oleg154-client.virtnet service : cur 5 worst 5 (at 1713480652, 3676s ago) 1 1 1 1 0 osc reconnect attempts on 2nd slow PASS 67b (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 68: AT: verify slowing locks ======= 19:52:11 (1713484331) at_history=8 at_history=8 /home/green/git/lustre-release/lustre/tests/replay-single.sh: line 1968: $ldlm_enqueue_min: ambiguous redirect fail_val=19 fail_loc=0x80000312 fail_val=25 fail_loc=0x80000312 fail_loc=0 /home/green/git/lustre-release/lustre/tests/replay-single.sh: line 1983: $ldlm_enqueue_min: ambiguous redirect PASS 68 (70s) debug_raw_pointers=0 debug_raw_pointers=0 Cleaning up AT ... debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 70a: check multi client t-f ======== 19:53:23 (1713484403) SKIP: replay-single test_70a Need two or more clients, have 1 SKIP 70a (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 70b: dbench 2mdts recovery; 1 clients ========================================================== 19:53:27 (1713484407) Starting client oleg154-client.virtnet: -o user_xattr,flock oleg154-server@tcp:/lustre /mnt/lustre Started clients oleg154-client.virtnet: 192.168.201.154@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) striped dir -i0 -c2 -H all_char /mnt/lustre/d70b.replay-single + MISSING_DBENCH_OK= + PATH=/opt/iozone/bin:/opt/iozone/bin:/home/green/git/lustre-release/lustre/tests/mpi:/home/green/git/lustre-release/lustre/tests/racer:/home/green/git/lustre-release/lustre/../lustre-iokit/sgpdd-survey:/home/green/git/lustre-release/lustre/tests:/home/green/git/lustre-release/lustre/utils/gss:/home/green/git/lustre-release/lustre/utils:/opt/iozone/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin::/home/green/git/lustre-release/lustre/utils:/home/green/git/lustre-release/lustre/tests:/sbin:/usr/sbin:/home/green/git/lustre-release/lustre/utils:/home/green/git/lustre-release/lustre/tests/: + DBENCH_LIB= + TESTSUITE=replay-single + TESTNAME=test_70b + MOUNT=/mnt/lustre ++ hostname + DIR=/mnt/lustre/d70b.replay-single/oleg154-client.virtnet + LCTL=/home/green/git/lustre-release/lustre/utils/lctl + rundbench 1 -t 300 dbench: no process found Started rundbench load pid=21814 ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2464 1285224 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2160 1285528 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 26148 3556152 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3592876 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 27712 7149028 1% /mnt/lustre test_70b fail mds1 1 times Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:53:36 (1713484416) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:53:51 (1713484431) targets are mounted 19:53:51 (1713484431) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2424 1285264 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2128 1285560 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 27440 3568832 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1724 3593016 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 29164 7161848 1% /mnt/lustre oleg154-client.virtnet: looking for dbench program oleg154-client.virtnet: /usr/bin/dbench oleg154-client.virtnet: creating output directory /mnt/lustre/d70b.replay-single/oleg154-client.virtnet oleg154-client.virtnet: mkdir: created directory '/mnt/lustre/d70b.replay-single/oleg154-client.virtnet' oleg154-client.virtnet: found dbench client file /usr/share/dbench/client.txt oleg154-client.virtnet: '/usr/share/dbench/client.txt' -> 'client.txt' oleg154-client.virtnet: running 'dbench 1 -t 300' on /mnt/lustre/d70b.replay-single/oleg154-client.virtnet at Thu Apr 18 19:53:28 EDT 2024 oleg154-client.virtnet: waiting for dbench pid 21838 oleg154-client.virtnet: dbench version 4.00 - Copyright Andrew Tridgell 1999-2004 oleg154-client.virtnet: oleg154-client.virtnet: Running for 300 seconds with load 'client.txt' and minimum warmup 60 secs oleg154-client.virtnet: failed to create barrier semaphore oleg154-client.virtnet: 0 of 1 processes prepared for launch 0 sec oleg154-client.virtnet: 1 of 1 processes prepared for launch 0 sec oleg154-client.virtnet: releasing clients oleg154-client.virtnet: 1 305 11.22 MB/sec warmup 1 sec latency 23.103 ms oleg154-client.virtnet: 1 664 10.58 MB/sec warmup 2 sec latency 18.738 ms oleg154-client.virtnet: 1 965 7.27 MB/sec warmup 3 sec latency 16.207 ms oleg154-client.virtnet: 1 1405 6.50 MB/sec warmup 4 sec latency 15.864 ms oleg154-client.virtnet: 1 1762 5.26 MB/sec warmup 5 sec latency 17.708 ms oleg154-client.virtnet: 1 2160 4.45 MB/sec warmup 6 sec latency 11.537 ms oleg154-client.virtnet: 1 2566 4.13 MB/sec warmup 7 sec latency 16.099 ms oleg154-client.virtnet: 1 3273 4.42 MB/sec warmup 8 sec latency 15.888 ms oleg154-client.virtnet: 1 3768 4.42 MB/sec warmup 9 sec latency 23.669 ms oleg154-client.virtnet: 1 4002 4.01 MB/sec warmup 10 sec latency 15.563 ms oleg154-client.virtnet: 1 4083 3.64 MB/sec warmup 11 sec latency 681.612 ms oleg154-client.virtnet: 1 4083 3.34 MB/sec warmup 12 sec latency 1681.802 ms oleg154-client.virtnet: 1 4083 3.08 MB/sec warmup 13 sec latency 2681.995 ms oleg154-client.virtnet: 1 4083 2.86 MB/sec warmup 14 sec latency 3682.217 ms oleg154-client.virtnet: 1 4083 2.67 MB/sec warmup 15 sec latency 4682.394 ms oleg154-client.virtnet: 1 4083 2.51 MB/sec warmup 16 sec latency 5682.608 ms oleg154-client.virtnet: 1 4083 2.36 MB/sec warmup 17 sec latency 6682.823 ms oleg154-client.virtnet: 1 4083 2.23 MB/sec warmup 18 sec latency 7683.034 ms oleg154-client.virtnet: 1 4083 2.11 MB/sec warmup 19 sec latency 8683.213 ms oleg154-client.virtnet: 1 4083 2.00 MB/sec warmup 20 sec latency 9683.373 ms oleg154-client.virtnet: 1 4083 1.91 MB/sec warmup 21 sec latency 10683.549 ms oleg154-client.virtnet: 1 4083 1.82 MB/sec warmup 22 sec latency 11683.728 ms oleg154-client.virtnet: 1 4083 1.74 MB/sec warmup 23 sec latency 12683.993 ms oleg154-client.virtnet: 1 4083 1.67 MB/sec warmup 24 sec latency 13684.255 ms oleg154-client.virtnet: 1 4083 1.60 MB/sec warmup 25 sec latency 14684.468 ms oleg154-client.virtnet: 1 4083 1.54 MB/sec warmup 26 sec latency 15684.676 ms oleg154-client.virtnet: 1 4360 1.50 MB/sec warmup 27 sec latency 15943.569 ms oleg154-client.virtnet: 1 4699 1.50 MB/sec warmup 28 sec latency 15.833 ms oleg154-client.virtnet: 1 5075 1.56 MB/sec warmup 29 sec latency 16.179 ms oleg154-client.virtnet: 1 5375 1.51 MB/sec warmup 30 sec latency 17.674 ms oleg154-client.virtnet: 1 5761 1.48 MB/sec warmup 31 sec latency 11.924 ms oleg154-client.virtnet: 1 6146 1.50 MB/sec warmup 32 sec latency 16.188 ms oleg154-client.virtnet: 1 6726 1.61 MB/sec warmup 33 sec latency 222.246 mstest_70b fail mds2 2 times Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 19:54:04 (1713484444) shut down Failover mds2 to oleg154-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0001 19:54:19 (1713484459) targets are mounted 19:54:19 (1713484459) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2384 1285304 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2052 1285636 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 36392 3570112 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 13236 3592872 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 49628 7162984 1% /mnt/lustre test_70b fail mds1 3 times Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:54:32 (1713484472) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-client.virtnet: 1 7230 1.70 MB/sec warmup 34 sec latency 15.649 ms oleg154-client.virtnet: 1 7495 1.69 MB/sec warmup 35 sec latency 16.469 ms oleg154-client.virtnet: 1 7514 1.64 MB/sec warmup 36 sec latency 907.508 ms oleg154-client.virtnet: 1 7514 1.60 MB/sec warmup 37 sec latency 1907.694 ms oleg154-client.virtnet: 1 7514 1.56 MB/sec warmup 38 sec latency 2907.894 ms oleg154-client.virtnet: 1 7514 1.52 MB/sec warmup 39 sec latency 3908.110 ms oleg154-client.virtnet: 1 7514 1.48 MB/sec warmup 40 sec latency 4908.356 ms oleg154-client.virtnet: 1 7514 1.44 MB/sec warmup 41 sec latency 5908.593 ms oleg154-client.virtnet: 1 7514 1.41 MB/sec warmup 42 sec latency 6908.763 ms oleg154-client.virtnet: 1 7514 1.38 MB/sec warmup 43 sec latency 7908.964 ms oleg154-client.virtnet: 1 7514 1.34 MB/sec warmup 44 sec latency 8909.165 ms oleg154-client.virtnet: 1 7514 1.31 MB/sec warmup 45 sec latency 9909.360 ms oleg154-client.virtnet: 1 7514 1.29 MB/sec warmup 46 sec latency 10909.462 ms oleg154-client.virtnet: 1 7514 1.26 MB/sec warmup 47 sec latency 11909.617 ms oleg154-client.virtnet: 1 7514 1.23 MB/sec warmup 48 sec latency 12909.791 ms oleg154-client.virtnet: 1 7514 1.21 MB/sec warmup 49 sec latency 13909.958 ms oleg154-client.virtnet: 1 7514 1.18 MB/sec warmup 50 sec latency 14910.140 ms oleg154-client.virtnet: 1 7514 1.16 MB/sec warmup 51 sec latency 15910.313 ms oleg154-client.virtnet: 1 7514 1.14 MB/sec warmup 52 sec latency 16910.511 ms oleg154-client.virtnet: 1 7514 1.12 MB/sec warmup 53 sec latency 17910.693 ms oleg154-client.virtnet: 1 7514 1.10 MB/sec warmup 54 sec latency 18910.906 ms oleg154-client.virtnet: 1 7519 1.08 MB/sec warmup 55 sec latency 19882.829 ms oleg154-client.virtnet: 1 7790 1.06 MB/sec warmup 56 sec latency 16.884 ms oleg154-client.virtnet: 1 8075 1.05 MB/sec warmup 57 sec latency 17.326 ms oleg154-client.virtnet: 1 8487 1.11 MB/sec warmup 58 sec latency 16.582 ms oleg154-client.virtnet: 1 8761 1.09 MB/sec warmup 59 sec latency 16.793 ms oleg154-client.virtnet: 1 9499 1.47 MB/sec execute 1 sec latency 16.696 ms oleg154-client.virtnet: 1 10016 3.48 MB/sec execute 2 sec latency 16.060 ms oleg154-client.virtnet: 1 10639 3.96 MB/sec execute 3 sec latency 15.752 ms oleg154-client.virtnet: 1 10938 3.32 MB/sec execute 4 sec latency 16.364 ms oleg154-client.virtnet: 1 11173 2.72 MB/sec execute 5 sec latency 18.259 ms oleg154-client.virtnet: 1 11523 2.37 MB/sec execute 6 sec latency 16.472 ms oleg154-client.virtnet: 1 11833 2.23 MB/sec execute 7 sec latency 16.264 ms oleg154-client.virtnet: 1 12225 2.34 MB/sec execute 8 sec latency 15.995 ms oleg154-client.virtnet: 1 12520 2.11 MB/sec execute 9 sec latency 17.874 ms oleg154-client.virtnet: 1 12923 1.96 MB/sec execute 10 sec latency 11.122 ms oleg154-client.virtnet: 1 13260 1.97 MB/sec execute 11 sec latency 16.322 ms oleg154-client.virtnet: 1 13988 2.33 MB/sec execute 12 sec latency 16.367 ms oleg154-client.virtnet: 1 14390 2.47 MB/sec execute 13 sec latency 13.592 ms oleg154-client.virtnet: 1 14622 2.33 MB/sec execute 14 sec latency 16.479 ms oleg154-client.virtnet: 1 14935 2.20 MB/sec execute 15 sec latency 20.125 ms oleg154-client.virtnet: 1 15234 2.16 MB/sec execute 16 sec latency 16.214 ms oleg154-client.virtnet: 1 15653 2.22 MB/sec execute 17 sec latency 16.885 ms oleg154-client.virtnet: 1 15963 2.11 MB/sec execute 18 sec latency 15.968 ms oleg154-client.virtnet: 1 16350 2.02 MB/sec execute 19 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:54:48 (1713484488) targets are mounted 19:54:48 (1713484488) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2384 1285304 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2076 1285612 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 36400 3570136 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 13088 3592932 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 49488 7163068 1% /mnt/lustre test_70b fail mds2 4 times Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 19:55:07 (1713484507) shut down Failover mds2 to oleg154-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0001 19:55:22 (1713484522) targets are mounted 19:55:22 (1713484522) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2384 1285304 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2052 1285636 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 36712 3569844 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 12924 3593408 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 49636 7163252 1% /mnt/lustre sec latency 10.837 ms oleg154-client.virtnet: 1 16698 2.02 MB/sec execute 20 sec latency 15.557 ms oleg154-client.virtnet: 1 17368 2.18 MB/sec execute 21 sec latency 16.325 ms oleg154-client.virtnet: 1 17879 2.29 MB/sec execute 22 sec latency 15.736 ms oleg154-client.virtnet: 1 18145 2.25 MB/sec execute 23 sec latency 16.132 ms oleg154-client.virtnet: 1 18463 2.17 MB/sec execute 24 sec latency 21.887 ms oleg154-client.virtnet: 1 18743 2.11 MB/sec execute 25 sec latency 16.091 ms oleg154-client.virtnet: 1 19170 2.19 MB/sec execute 26 sec latency 16.206 ms oleg154-client.virtnet: 1 19467 2.11 MB/sec execute 27 sec latency 17.846 ms oleg154-client.virtnet: 1 19837 2.05 MB/sec execute 28 sec latency 11.313 ms oleg154-client.virtnet: 1 20161 2.03 MB/sec execute 29 sec latency 198.553 ms oleg154-client.virtnet: 1 20349 1.99 MB/sec execute 30 sec latency 651.809 ms oleg154-client.virtnet: 1 20988 2.10 MB/sec execute 31 sec latency 15.977 ms oleg154-client.virtnet: 1 21467 2.18 MB/sec execute 32 sec latency 15.692 ms oleg154-client.virtnet: 1 21494 2.11 MB/sec execute 33 sec latency 913.204 ms oleg154-client.virtnet: 1 21494 2.05 MB/sec execute 34 sec latency 1913.394 ms oleg154-client.virtnet: 1 21494 1.99 MB/sec execute 35 sec latency 2913.569 ms oleg154-client.virtnet: 1 21494 1.94 MB/sec execute 36 sec latency 3913.746 ms oleg154-client.virtnet: 1 21494 1.88 MB/sec execute 37 sec latency 4913.941 ms oleg154-client.virtnet: 1 21494 1.83 MB/sec execute 38 sec latency 5914.135 ms oleg154-client.virtnet: 1 21494 1.79 MB/sec execute 39 sec latency 6914.320 ms oleg154-client.virtnet: 1 21494 1.74 MB/sec execute 40 sec latency 7914.551 ms oleg154-client.virtnet: 1 21494 1.70 MB/sec execute 41 sec latency 8914.780 ms oleg154-client.virtnet: 1 21494 1.66 MB/sec execute 42 sec latency 9914.981 ms oleg154-client.virtnet: 1 21494 1.62 MB/sec execute 43 sec latency 10915.177 ms oleg154-client.virtnet: 1 21494 1.58 MB/sec execute 44 sec latency 11915.396 ms oleg154-client.virtnet: 1 21494 1.55 MB/sec execute 45 sec latency 12915.574 ms oleg154-client.virtnet: 1 21494 1.52 MB/sec execute 46 sec latency 13915.765 ms oleg154-client.virtnet: 1 21494 1.48 MB/sec execute 47 sec latency 14915.956 ms oleg154-client.virtnet: 1 21494 1.45 MB/sec execute 48 sec latency 15916.152 ms oleg154-client.virtnet: 1 21494 1.42 MB/sec execute 49 sec latency 16916.349 ms oleg154-client.virtnet: 1 21494 1.39 MB/sec execute 50 sec latency 17916.513 ms oleg154-client.virtnet: 1 21494 1.37 MB/sec execute 51 sec latency 18916.710 ms oleg154-client.virtnet: 1 21494 1.34 MB/sec execute 52 sec latency 19916.876 ms oleg154-client.virtnet: 1 21494 1.31 MB/sec execute 53 sec latency 20917.063 ms oleg154-client.virtnet: 1 21494 1.29 MB/sec execute 54 sec latency 21917.254 ms oleg154-client.virtnet: 1 21494 1.27 MB/sec execute 55 sec latency 22917.422 ms oleg154-client.virtnet: 1 21494 1.24 MB/sec execute 56 sec latency 23917.588 ms oleg154-client.virtnet: 1 21494 1.22 MB/sec execute 57 sec latency 24917.797 ms oleg154-client.virtnet: 1 21536 1.22 MB/sec execute 58 sec latency 25823.099 ms oleg154-client.virtnet: 1 21772 1.20 MB/sec execute 59 sec latency 16.096 ms oleg154-client.virtnet: 1 22041 1.19 MB/sec execute 60 sec latency 22.028 ms oleg154-client.virtnet: 1 22315 1.18 MB/sec execute 61 sec latency 16.324 ms oleg154-client.virtnet: 1 22734 1.23 MB/sec execute 62 sec latency 16.318 ms oleg154-client.virtnet: 1 23026 1.21 MB/sec execute 63 sec latency 16.621 ms test_70b fail mds1 5 times Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:55:36 (1713484536) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:55:51 (1713484551) targets are mounted 19:55:51 (1713484551) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2384 1285304 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2076 1285612 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 36628 3569880 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 12924 3593104 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 49552 7162984 1% /mnt/lustre test_70b fail mds2 6 times Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 19:56:04 (1713484564) shut down Failover mds2 to oleg154-server mount facets: mds2 oleg154-client.virtnet: 1 23381 1.20 MB/sec execute 64 sec latency 11.068 ms oleg154-client.virtnet: 1 23755 1.20 MB/sec execute 65 sec latency 16.250 ms oleg154-client.virtnet: 1 24312 1.27 MB/sec execute 66 sec latency 16.385 ms oleg154-client.virtnet: 1 24944 1.32 MB/sec execute 67 sec latency 15.684 ms oleg154-client.virtnet: 1 25225 1.32 MB/sec execute 68 sec latency 15.952 ms oleg154-client.virtnet: 1 25536 1.31 MB/sec execute 69 sec latency 15.444 ms oleg154-client.virtnet: 1 25808 1.30 MB/sec execute 70 sec latency 21.503 ms oleg154-client.virtnet: 1 26155 1.30 MB/sec execute 71 sec latency 16.883 ms oleg154-client.virtnet: 1 26549 1.32 MB/sec execute 72 sec latency 16.208 ms oleg154-client.virtnet: 1 26889 1.31 MB/sec execute 73 sec latency 17.360 ms oleg154-client.virtnet: 1 27283 1.31 MB/sec execute 74 sec latency 16.330 ms oleg154-client.virtnet: 1 27846 1.37 MB/sec execute 75 sec latency 16.080 ms oleg154-client.virtnet: 1 28478 1.42 MB/sec execute 76 sec latency 16.209 ms oleg154-client.virtnet: 1 28764 1.42 MB/sec execute 77 sec latency 15.833 ms oleg154-client.virtnet: 1 29071 1.40 MB/sec execute 78 sec latency 15.156 ms oleg154-client.virtnet: 1 29362 1.39 MB/sec execute 79 sec latency 18.389 ms oleg154-client.virtnet: 1 29804 1.43 MB/sec execute 80 sec latency 15.395 ms oleg154-client.virtnet: 1 30114 1.41 MB/sec execute 81 sec latency 16.633 ms oleg154-client.virtnet: 1 30529 1.40 MB/sec execute 82 sec latency 10.162 ms oleg154-client.virtnet: 1 30896 1.41 MB/sec execute 83 sec latency 15.598 ms oleg154-client.virtnet: 1 31435 1.45 MB/sec execute 84 sec latency 15.980 ms oleg154-client.virtnet: 1 32057 1.49 MB/sec execute 85 sec latency 16.041 ms oleg154-client.virtnet: 1 32329 1.49 MB/sec execute 86 sec latency 16.228 ms oleg154-client.virtnet: 1 32643 1.48 MB/sec execute 87 sec latency 21.730 ms oleg154-client.virtnet: 1 32943 1.47 MB/sec execute 88 sec latency 16.378 ms oleg154-client.virtnet: 1 33362 1.50 MB/sec execute 89 sec latency 16.745 ms oleg154-client.virtnet: 1 33661 1.48 MB/sec execute 90 sec latency 18.202 ms oleg154-client.virtnet: 1 34009 1.47 MB/sec execute 91 sec latency 12.080 ms oleg154-client.virtnet: 1 34382 1.47 MB/sec execute 92 sec latency 81.698 ms oleg154-client.virtnet: 1 34558 1.47 MB/sec execute 93 sec latency 604.250 ms oleg154-client.virtnet: 1 35268 1.52 MB/sec execute 94 sec latency 15.939 ms oleg154-client.virtnet: 1 35724 1.55 MB/sec execute 95 sec latency 15.520 ms oleg154-client.virtnet: 1 35778 1.53 MB/sec execute 96 sec latency 760.673 ms oleg154-client.virtnet: 1 35778 1.52 MB/sec execute 97 sec latency 1760.841 ms oleg154-client.virtnet: 1 35778 1.50 MB/sec execute 98 sec latency 2761.032 ms oleg154-client.virtnet: 1 35778 1.49 MB/sec execute 99 sec latency 3761.274 ms oleg154-client.virtnet: 1 35778 1.47 MB/sec execute 100 sec latency 4761.552 ms oleg154-client.virtnet: 1 35778 1.46 MB/sec execute 101 sec latency 5761.760 ms oleg154-client.virtnet: 1 35778 1.44 MB/sec execute 102 sec latency 6761.905 ms oleg154-client.virtnet: 1 35778 1.43 MB/sec execute 103 sec latency 7762.058 ms oleg154-client.virtnet: 1 35778 1.41 MB/sec execute 104 sec latency 8762.302 ms oleg154-client.virtnet: 1 35778 1.40 MB/sec execute 105 sec latency 9762.520 ms oleg154-client.virtnet: 1 35778 1.39 MB/sec execute 106 sec latency 10762.700 ms oleg154-client.virtnet: 1 35778 1.38 MB/sec execute 107 sec latency 11762.860 ms oleg154-client.virtnet: 1 35778 1.36 MB/sec execute 10Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0001 19:56:19 (1713484579) targets are mounted 19:56:19 (1713484579) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2384 1285304 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2052 1285636 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 36396 3570316 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 13084 3593192 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 49480 7163508 1% /mnt/lustre test_70b fail mds1 7 times Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:56:32 (1713484592) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:56:47 (1713484607) targets are mounted 19:56:47 (1713484607) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2384 1285304 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2076 1285612 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 36392 3570040 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 13160 3592972 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 49552 7163012 1% /mnt/lustre test_70b fail mds2 8 times Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 19:57:00 (1713484620) shut down 8 sec latency 12763.061 ms oleg154-client.virtnet: 1 35778 1.35 MB/sec execute 109 sec latency 13763.266 ms oleg154-client.virtnet: 1 35778 1.34 MB/sec execute 110 sec latency 14763.477 ms oleg154-client.virtnet: 1 35778 1.33 MB/sec execute 111 sec latency 15763.667 ms oleg154-client.virtnet: 1 35778 1.31 MB/sec execute 112 sec latency 16763.872 ms oleg154-client.virtnet: 1 35778 1.30 MB/sec execute 113 sec latency 17764.108 ms oleg154-client.virtnet: 1 35778 1.29 MB/sec execute 114 sec latency 18764.322 ms oleg154-client.virtnet: 1 35783 1.28 MB/sec execute 115 sec latency 19739.843 ms oleg154-client.virtnet: 1 36030 1.27 MB/sec execute 116 sec latency 15.188 ms oleg154-client.virtnet: 1 36322 1.26 MB/sec execute 117 sec latency 15.558 ms oleg154-client.virtnet: 1 36658 1.27 MB/sec execute 118 sec latency 15.355 ms oleg154-client.virtnet: 1 37031 1.28 MB/sec execute 119 sec latency 17.507 ms oleg154-client.virtnet: 1 37348 1.27 MB/sec execute 120 sec latency 18.331 ms oleg154-client.virtnet: 1 37810 1.27 MB/sec execute 121 sec latency 11.468 ms oleg154-client.virtnet: 1 38275 1.30 MB/sec execute 122 sec latency 16.641 ms oleg154-client.virtnet: 1 39031 1.34 MB/sec execute 123 sec latency 21.431 ms oleg154-client.virtnet: 1 39339 1.34 MB/sec execute 124 sec latency 15.712 ms oleg154-client.virtnet: 1 39565 1.33 MB/sec execute 125 sec latency 19.756 ms oleg154-client.virtnet: 1 39917 1.33 MB/sec execute 126 sec latency 20.179 ms oleg154-client.virtnet: 1 40235 1.33 MB/sec execute 127 sec latency 18.824 ms oleg154-client.virtnet: 1 40620 1.34 MB/sec execute 128 sec latency 24.935 ms oleg154-client.virtnet: 1 40900 1.33 MB/sec execute 129 sec latency 17.646 ms oleg154-client.virtnet: 1 41270 1.33 MB/sec execute 130 sec latency 11.245 ms oleg154-client.virtnet: 1 41648 1.33 MB/sec execute 131 sec latency 16.484 ms oleg154-client.virtnet: 1 42328 1.37 MB/sec execute 132 sec latency 16.363 ms oleg154-client.virtnet: 1 42764 1.39 MB/sec execute 133 sec latency 13.033 ms oleg154-client.virtnet: 1 43022 1.39 MB/sec execute 134 sec latency 16.445 ms oleg154-client.virtnet: 1 43397 1.38 MB/sec execute 135 sec latency 16.376 ms oleg154-client.virtnet: 1 43750 1.38 MB/sec execute 136 sec latency 16.105 ms oleg154-client.virtnet: 1 44142 1.39 MB/sec execute 137 sec latency 15.380 ms oleg154-client.virtnet: 1 44444 1.38 MB/sec execute 138 sec latency 17.574 ms oleg154-client.virtnet: 1 44855 1.38 MB/sec execute 139 sec latency 11.145 ms oleg154-client.virtnet: 1 45375 1.40 MB/sec execute 140 sec latency 12.920 ms oleg154-client.virtnet: 1 46089 1.44 MB/sec execute 141 sec latency 16.855 ms oleg154-client.virtnet: 1 46416 1.44 MB/sec execute 142 sec latency 13.603 ms oleg154-client.virtnet: 1 46651 1.43 MB/sec execute 143 sec latency 16.228 ms oleg154-client.virtnet: 1 47022 1.43 MB/sec execute 144 sec latency 21.721 ms oleg154-client.virtnet: 1 47346 1.43 MB/sec execute 145 sec latency 15.803 ms oleg154-client.virtnet: 1 47733 1.44 MB/sec execute 146 sec latency 15.916 ms oleg154-client.virtnet: 1 48029 1.43 MB/sec execute 147 sec latency 18.039 ms oleg154-client.virtnet: 1 48486 1.43 MB/sec execute 148 sec latency 11.256 ms oleg154-client.virtnet: 1 48625 1.43 MB/sec execute 149 sec latency 490.924 ms oleg154-client.virtnet: 1 49158 1.45 MB/sec execute 150 sec latency 554.436 ms oleg154-client.virtnet: 1 49799 1.47 MB/sec execute 151 sec latency 15.712 ms oleg154-client.virtnet: 1 49999 1.47 MB/sec execute 152 sec latency 274.344 ms oleg154-client.virtnet: 1 49999 Failover mds2 to oleg154-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0001 19:57:16 (1713484636) targets are mounted 19:57:16 (1713484636) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2384 1285304 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2052 1285636 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 36392 3570200 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 13244 3592904 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 49636 7163104 1% /mnt/lustre test_70b fail mds1 9 times Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:57:29 (1713484649) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:57:44 (1713484664) targets are mounted 19:57:45 (1713484665) facet_failover done 1.46 MB/sec execute 153 sec latency 1274.597 ms oleg154-client.virtnet: 1 49999 1.45 MB/sec execute 154 sec latency 2274.776 ms oleg154-client.virtnet: 1 49999 1.44 MB/sec execute 155 sec latency 3275.020 ms oleg154-client.virtnet: 1 49999 1.43 MB/sec execute 156 sec latency 4275.240 ms oleg154-client.virtnet: 1 49999 1.42 MB/sec execute 157 sec latency 5275.474 ms oleg154-client.virtnet: 1 49999 1.42 MB/sec execute 158 sec latency 6275.695 ms oleg154-client.virtnet: 1 49999 1.41 MB/sec execute 159 sec latency 7275.993 ms oleg154-client.virtnet: 1 49999 1.40 MB/sec execute 160 sec latency 8276.181 ms oleg154-client.virtnet: 1 49999 1.39 MB/sec execute 161 sec latency 9276.375 ms oleg154-client.virtnet: 1 49999 1.38 MB/sec execute 162 sec latency 10276.578 ms oleg154-client.virtnet: 1 49999 1.37 MB/sec execute 163 sec latency 11276.770 ms oleg154-client.virtnet: 1 49999 1.36 MB/sec execute 164 sec latency 12276.995 ms oleg154-client.virtnet: 1 49999 1.36 MB/sec execute 165 sec latency 13277.172 ms oleg154-client.virtnet: 1 49999 1.35 MB/sec execute 166 sec latency 14277.360 ms oleg154-client.virtnet: 1 49999 1.34 MB/sec execute 167 sec latency 15277.570 ms oleg154-client.virtnet: 1 49999 1.33 MB/sec execute 168 sec latency 16277.838 ms oleg154-client.virtnet: 1 49999 1.32 MB/sec execute 169 sec latency 17278.014 ms oleg154-client.virtnet: 1 49999 1.32 MB/sec execute 170 sec latency 18278.198 ms oleg154-client.virtnet: 1 49999 1.31 MB/sec execute 171 sec latency 19278.385 ms oleg154-client.virtnet: 1 50109 1.30 MB/sec execute 172 sec latency 19835.999 ms oleg154-client.virtnet: 1 50379 1.30 MB/sec execute 173 sec latency 17.896 ms oleg154-client.virtnet: 1 50667 1.29 MB/sec execute 174 sec latency 16.602 ms oleg154-client.virtnet: 1 51086 1.31 MB/sec execute 175 sec latency 16.119 ms oleg154-client.virtnet: 1 51382 1.30 MB/sec execute 176 sec latency 16.065 ms oleg154-client.virtnet: 1 51717 1.30 MB/sec execute 177 sec latency 18.568 ms oleg154-client.virtnet: 1 52121 1.30 MB/sec execute 178 sec latency 15.983 ms oleg154-client.virtnet: 1 52665 1.32 MB/sec execute 179 sec latency 15.572 ms oleg154-client.virtnet: 1 53267 1.34 MB/sec execute 180 sec latency 16.197 ms oleg154-client.virtnet: 1 53671 1.34 MB/sec execute 181 sec latency 12.949 ms oleg154-client.virtnet: 1 53995 1.34 MB/sec execute 182 sec latency 20.975 ms oleg154-client.virtnet: 1 54305 1.34 MB/sec execute 183 sec latency 16.423 ms oleg154-client.virtnet: 1 54706 1.35 MB/sec execute 184 sec latency 15.769 ms oleg154-client.virtnet: 1 55013 1.34 MB/sec execute 185 sec latency 17.117 ms oleg154-client.virtnet: 1 55384 1.34 MB/sec execute 186 sec latency 10.989 ms oleg154-client.virtnet: 1 55727 1.34 MB/sec execute 187 sec latency 15.638 ms oleg154-client.virtnet: 1 56406 1.36 MB/sec execute 188 sec latency 16.210 ms oleg154-client.virtnet: 1 56903 1.38 MB/sec execute 189 sec latency 15.864 ms oleg154-client.virtnet: 1 57168 1.38 MB/sec execute 190 sec latency 16.066 ms oleg154-client.virtnet: 1 57466 1.37 MB/sec execute 191 sec latency 19.655 ms oleg154-client.virtnet: 1 57758 1.37 MB/sec execute 192 sec latency 15.973 ms oleg154-client.virtnet: 1 58188 1.38 MB/sec execute 193 sec latency 16.042 ms oleg154-client.virtnet: 1 58497 1.38 MB/sec execute 194 sec latency 16.763 ms oleg154-client.virtnet: 1 58872 1.37 MB/sec execute 195 sec latency 10.835 ms oleg154-client.virtnet: 1 59249 1.38 MB/sec execute 196 sec latency 15.405 ms oleg154-client.virtnet: 1 59802 1.39 MB/sec execute 197 sec latency oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2384 1285304 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2076 1285612 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 36392 3570088 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 13160 3592832 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 49552 7162920 1% /mnt/lustre test_70b fail mds2 10 times Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 19:58:03 (1713484683) shut down Failover mds2 to oleg154-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0001 19:58:19 (1713484699) targets are mounted 19:58:19 (1713484699) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2384 1285304 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2052 1285636 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 36392 3570508 2% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 13244 3593592 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 49636 7164100 1% /mnt/lustre test_70b fail mds1 11 times 15.629 ms oleg154-client.virtnet: 1 60445 1.41 MB/sec execute 198 sec latency 15.360 ms oleg154-client.virtnet: 1 60719 1.41 MB/sec execute 199 sec latency 15.801 ms oleg154-client.virtnet: 1 61035 1.41 MB/sec execute 200 sec latency 18.255 ms oleg154-client.virtnet: 1 61321 1.40 MB/sec execute 201 sec latency 16.433 ms oleg154-client.virtnet: 1 61782 1.42 MB/sec execute 202 sec latency 15.566 ms oleg154-client.virtnet: 1 62112 1.41 MB/sec execute 203 sec latency 18.130 ms oleg154-client.virtnet: 1 62489 1.41 MB/sec execute 204 sec latency 12.868 ms oleg154-client.virtnet: 1 62883 1.41 MB/sec execute 205 sec latency 15.777 ms oleg154-client.virtnet: 1 63248 1.43 MB/sec execute 206 sec latency 434.310 ms oleg154-client.virtnet: 1 63855 1.44 MB/sec execute 207 sec latency 521.733 ms oleg154-client.virtnet: 1 64185 1.44 MB/sec execute 208 sec latency 17.843 ms oleg154-client.virtnet: 1 64358 1.44 MB/sec execute 209 sec latency 348.817 ms oleg154-client.virtnet: 1 64358 1.43 MB/sec execute 210 sec latency 1349.062 ms oleg154-client.virtnet: 1 64358 1.42 MB/sec execute 211 sec latency 2349.327 ms oleg154-client.virtnet: 1 64358 1.42 MB/sec execute 212 sec latency 3349.548 ms oleg154-client.virtnet: 1 64358 1.41 MB/sec execute 213 sec latency 4349.753 ms oleg154-client.virtnet: 1 64358 1.40 MB/sec execute 214 sec latency 5349.977 ms oleg154-client.virtnet: 1 64358 1.40 MB/sec execute 215 sec latency 6350.164 ms oleg154-client.virtnet: 1 64358 1.39 MB/sec execute 216 sec latency 7350.338 ms oleg154-client.virtnet: 1 64358 1.38 MB/sec execute 217 sec latency 8350.492 ms oleg154-client.virtnet: 1 64358 1.38 MB/sec execute 218 sec latency 9350.652 ms oleg154-client.virtnet: 1 64358 1.37 MB/sec execute 219 sec latency 10350.814 ms oleg154-client.virtnet: 1 64358 1.37 MB/sec execute 220 sec latency 11351.021 ms oleg154-client.virtnet: 1 64358 1.36 MB/sec execute 221 sec latency 12351.275 ms oleg154-client.virtnet: 1 64358 1.35 MB/sec execute 222 sec latency 13351.501 ms oleg154-client.virtnet: 1 64358 1.35 MB/sec execute 223 sec latency 14351.710 ms oleg154-client.virtnet: 1 64358 1.34 MB/sec execute 224 sec latency 15351.908 ms oleg154-client.virtnet: 1 64358 1.33 MB/sec execute 225 sec latency 16352.121 ms oleg154-client.virtnet: 1 64358 1.33 MB/sec execute 226 sec latency 17352.289 ms oleg154-client.virtnet: 1 64358 1.32 MB/sec execute 227 sec latency 18352.476 ms oleg154-client.virtnet: 1 64358 1.32 MB/sec execute 228 sec latency 19352.653 ms oleg154-client.virtnet: 1 64358 1.31 MB/sec execute 229 sec latency 20352.846 ms oleg154-client.virtnet: 1 64358 1.31 MB/sec execute 230 sec latency 21353.120 ms oleg154-client.virtnet: 1 64358 1.30 MB/sec execute 231 sec latency 22353.297 ms oleg154-client.virtnet: 1 64358 1.29 MB/sec execute 232 sec latency 23353.460 ms oleg154-client.virtnet: 1 64358 1.29 MB/sec execute 233 sec latency 24353.629 ms oleg154-client.virtnet: 1 64358 1.28 MB/sec execute 234 sec latency 25353.787 ms oleg154-client.virtnet: 1 64546 1.28 MB/sec execute 235 sec latency 25774.691 ms oleg154-client.virtnet: 1 64838 1.28 MB/sec execute 236 sec latency 17.514 ms oleg154-client.virtnet: 1 65254 1.29 MB/sec execute 237 sec latency 15.906 ms oleg154-client.virtnet: 1 65531 1.28 MB/sec execute 238 sec latency 16.084 ms oleg154-client.virtnet: 1 65878 1.28 MB/sec execute 239 sec latency 17.021 ms oleg154-client.virtnet: 1 66278 1.28 MB/sec execute 240 sec latency 14.360 ms oleg154-client.virtnet: 1 66836 1.30 MB/sec execute 241 sec latency 15.354 ms oleg154-clFailing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 19:58:32 (1713484712) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 19:58:47 (1713484727) targets are mounted 19:58:47 (1713484727) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec ient.virtnet: 1 67446 1.31 MB/sec execute 242 sec latency 15.529 ms oleg154-client.virtnet: 1 67758 1.31 MB/sec execute 243 sec latency 16.402 ms oleg154-client.virtnet: 1 67944 1.31 MB/sec execute 244 sec latency 285.677 ms oleg154-client.virtnet: 1 67944 1.30 MB/sec execute 245 sec latency 1285.860 ms oleg154-client.virtnet: 1 67944 1.30 MB/sec execute 246 sec latency 2286.009 ms oleg154-client.virtnet: 1 67944 1.29 MB/sec execute 247 sec latency 3286.204 ms oleg154-client.virtnet: 1 67944 1.29 MB/sec execute 248 sec latency 4286.400 ms oleg154-client.virtnet: 1 67944 1.28 MB/sec execute 249 sec latency 5286.601 ms oleg154-client.virtnet: 1 67944 1.28 MB/sec execute 250 sec latency 6286.787 ms oleg154-client.virtnet: 1 67944 1.27 MB/sec execute 251 sec latency 7286.938 ms oleg154-client.virtnet: 1 67944 1.27 MB/sec execute 252 sec latency 8287.103 ms oleg154-client.virtnet: 1 67944 1.26 MB/sec execute 253 sec latency 9287.294 ms oleg154-client.virtnet: 1 67944 1.26 MB/sec execute 254 sec latency 10287.414 ms oleg154-client.virtnet: 1 67944 1.25 MB/sec execute 255 sec latency 11287.614 ms oleg154-client.virtnet: 1 67944 1.25 MB/sec execute 256 sec latency 12287.818 ms oleg154-client.virtnet: 1 67944 1.24 MB/sec execute 257 sec latency 13287.996 ms oleg154-client.virtnet: 1 67944 1.24 MB/sec execute 258 sec latency 14288.167 ms oleg154-client.virtnet: 1 67944 1.23 MB/sec execute 259 sec latency 15288.328 ms oleg154-client.virtnet: 1 67944 1.23 MB/sec execute 260 sec latency 16288.562 ms oleg154-client.virtnet: 1 67944 1.22 MB/sec execute 261 sec latency 17288.778 ms oleg154-client.virtnet: 1 67944 1.22 MB/sec execute 262 sec latency 18289.001 ms oleg154-client.virtnet: 1 68211 1.22 MB/sec execute 263 sec latency 18600.914 ms oleg154-client.virtnet: 1 68547 1.22 MB/sec execute 264 sec latency 16.074 ms oleg154-client.virtnet: 1 68931 1.22 MB/sec execute 265 sec latency 15.918 ms oleg154-client.virtnet: 1 69233 1.22 MB/sec execute 266 sec latency 17.596 ms oleg154-client.virtnet: 1 69609 1.22 MB/sec execute 267 sec latency 10.995 ms oleg154-client.virtnet: 1 70002 1.22 MB/sec execute 268 sec latency 16.282 ms oleg154-client.virtnet: 1 70689 1.24 MB/sec execute 269 sec latency 17.131 ms oleg154-client.virtnet: 1 71127 1.25 MB/sec execute 270 sec latency 13.179 ms oleg154-client.virtnet: 1 71388 1.25 MB/sec execute 271 sec latency 16.054 ms oleg154-client.virtnet: 1 71712 1.25 MB/sec execute 272 sec latency 19.688 ms oleg154-client.virtnet: 1 72030 1.25 MB/sec execute 273 sec latency 15.915 ms oleg154-client.virtnet: 1 72425 1.25 MB/sec execute 274 sec latency 15.776 ms oleg154-client.virtnet: 1 72724 1.25 MB/sec execute 275 sec latency 17.721 ms oleg154-client.virtnet: 1 73070 1.25 MB/sec execute 276 sec latency 11.230 ms oleg154-client.virtnet: 1 73439 1.25 MB/sec execute 277 sec latency 15.990 ms oleg154-client.virtnet: 1 73993 1.26 MB/sec execute 278 sec latency 16.604 ms oleg154-client.virtnet: 1 74619 1.28 MB/sec execute 279 sec latency 16.067 ms oleg154-client.virtnet: 1 74904 1.28 MB/sec execute 280 sec latency 15.055 ms oleg154-client.virtnet: 1 75227 1.27 MB/sec execute 281 sec latency 19.227 ms oleg154-client.virtnet: 1 75522 1.27 MB/sec execute 282 sec latency 15.657 ms oleg154-client.virtnet: 1 75941 1.28 MB/sec execute 283 sec latency 16.501 ms oleg154-client.virtnet: 1 76238 1.28 MB/sec execute 284 sec latency 17.949 ms oleg154-client.virtnet: 1 76575 1.28 MB/sec execute 285 sec latency 15.374 ms oleg154-client.virtnet: 1 76973 1.28 MB/sec execute 286 sec latency 15.358 ms oleg154-client.virtnet: 1 77524 1.29 MB/sec execute 287 sec latency 16.230 ms oleg154-client.virtnet: 1 78150 1.30 MB/sec execute 288 sec latency 15.589 ms oleg154-client.virtnet: 1 78430 1.30 MB/sec execute 289 sec latency 16.336 ms oleg154-client.virtnet: 1 78748 1.30 MB/sec execute 290 sec latency 16.593 ms oleg154-client.virtnet: 1 79023 1.30 MB/sec execute 291 sec latency 21.822 ms oleg154-client.virtnet: 1 79361 1.30 MB/sec execute 292 sec latency 16.452 ms oleg154-client.virtnet: 1 79752 1.30 MB/sec execute 293 sec latency 15.822 ms oleg154-client.virtnet: 1 80105 1.30 MB/sec execute 294 sec latency 18.204 ms oleg154-client.virtnet: 1 80512 1.30 MB/sec execute 295 sec latency 12.821 ms oleg154-client.virtnet: 1 81107 1.32 MB/sec execute 296 sec latency 15.448 ms oleg154-client.virtnet: 1 81740 1.33 MB/sec execute 297 sec latency 15.656 ms oleg154-client.virtnet: 1 82009 1.33 MB/sec execute 298 sec latency 15.999 ms oleg154-client.virtnet: 1 82333 1.33 MB/sec execute 299 sec latency 19.531 ms oleg154-client.virtnet: 1 cleanup 300 sec oleg154-client.virtnet: 0 cleanup 300 sec oleg154-client.virtnet: oleg154-client.virtnet: Operation Count AvgLat MaxLat oleg154-client.virtnet: ---------------------------------------- oleg154-client.virtnet: NTCreateX 12704 13.747 25823.084 oleg154-client.virtnet: Close 9312 1.832 10.913 oleg154-client.virtnet: Rename 539 11.843 21.412 oleg154-client.virtnet: Unlink 2586 10.594 19739.816 oleg154-client.virtnet: Qpathinfo 11484 3.854 18600.893 oleg154-client.virtnet: Qfileinfo 2002 0.204 2.541 oleg154-client.virtnet: Qfsinfo 2126 0.355 18.244 oleg154-client.virtnet: Sfileinfo 1032 6.318 14.241 oleg154-client.virtnet: Find 4461 0.626 12.193 oleg154-client.virtnet: WriteX 6277 1.369 7.586 oleg154-client.virtnet: ReadX 20005 0.024 1.129 oleg154-client.virtnet: LockX 42 1.895 2.229 oleg154-client.virtnet: UnlockX 42 1.993 2.337 oleg154-client.virtnet: Flush 895 11.047 651.793 oleg154-client.virtnet: oleg154-client.virtnet: Throughput 1.3253 MB/sec 1 clients 1 procs max_latency=25823.099 ms oleg154-client.virtnet: stopping dbench on /mnt/lustre/d70b.replay-single/oleg154-client.virtnet at Thu Apr 18 19:59:28 EDT 2024 with return code 0 oleg154-client.virtnet: clean dbench files on /mnt/lustre/d70b.replay-single/oleg154-client.virtnet oleg154-client.virtnet: /mnt/lustre/d70b.replay-single/oleg154-client.virtnet /mnt/lustre/d70b.replay-single/oleg154-client.virtnet oleg154-client.virtnet: removed directory: 'clients/client0/~dmtmp/PWRPNT' oleg154-client.virtnet: removed directory: 'clients/client0/~dmtmp/ACCESS' oleg154-client.virtnet: removed directory: 'clients/client0/~dmtmp/COREL' oleg154-client.virtnet: removed directory: 'clients/client0/~dmtmp/SEED' oleg154-client.virtnet: removed directory: 'clients/client0/~dmtmp/WORDPRO' oleg154-client.virtnet: removed directory: 'clients/client0/~dmtmp/PM' oleg154-client.virtnet: removed directory: 'clients/client0/~dmtmp/WORD' oleg154-client.virtnet: removed directory: 'clients/client0/~dmtmp/EXCEL' oleg154-client.virtnet: removed directory: 'clients/client0/~dmtmp/PARADOX' oleg154-client.virtnet: removed directory: 'clients/client0/~dmtmp' oleg154-client.virtnet: removed directory: 'clients/client0' oleg154-client.virtnet: removed directory: 'clients' oleg154-client.virtnet: removed 'client.txt' oleg154-client.virtnet: /mnt/lustre/d70b.replay-single/oleg154-client.virtnet oleg154-client.virtnet: dbench successfully finished PASS 70b (363s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 70c: tar 2mdts recovery ============ 19:59:32 (1713484772) Starting client oleg154-client.virtnet: -o user_xattr,flock oleg154-server@tcp:/lustre /mnt/lustre Started clients oleg154-client.virtnet: 192.168.201.154@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Started tar 2431 striped dir -i0 -c2 -H crush /mnt/lustre/d70c.replay-single tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets tar: Removing leading `/' from member names UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 6552 1281136 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 4368 1283320 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 6052 3589144 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 17560 3574480 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 23612 7163624 1% /mnt/lustre tar: Removing leading `/' from hard link targets test_70c fail mds2 1 times Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 20:01:53 (1713484913) shut down Failover mds2 to oleg154-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0001 20:02:09 (1713484929) targets are mounted 20:02:09 (1713484929) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 6840 1280848 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 4320 1283368 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 20932 3572000 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 6004 3589220 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 26936 7161220 1% /mnt/lustre test_70c fail mds2 2 times Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 20:04:31 (1713485071) shut down Failover mds2 to oleg154-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0001 20:04:46 (1713485086) targets are mounted 20:04:46 (1713485086) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 70c (362s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 70d: mkdir/rmdir striped dir 2mdts recovery ========================================================== 20:05:37 (1713485137) Starting client oleg154-client.virtnet: -o user_xattr,flock oleg154-server@tcp:/lustre /mnt/lustre Started clients oleg154-client.virtnet: 192.168.201.154@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Started 5779 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 5364 1282324 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 4304 1283384 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3603780 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3604120 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7207900 1% /mnt/lustre test_70d fail mds2 1 times Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 20:07:52 (1713485272) shut down Failover mds2 to oleg154-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0001 20:08:07 (1713485287) targets are mounted 20:08:07 (1713485287) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 4424 1283264 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 3656 1284032 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3603780 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3604120 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7207900 1% /mnt/lustre test_70d fail mds2 2 times Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 20:10:28 (1713485428) shut down Failover mds2 to oleg154-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0001 20:10:42 (1713485442) targets are mounted 20:10:42 (1713485442) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec /home/green/git/lustre-release/lustre/tests/test-framework.sh: line 4697: 5779 Killed ( while true; do $LFS mkdir -i0 -c2 $DIR/$tdir/test || { echo "mkdir fails"; break; }; $LFS mkdir -i1 -c2 $DIR/$tdir/test1 || { echo "mkdir fails"; break; }; touch $DIR/$tdir/test/a || { echo "touch fails"; break; }; mkdir $DIR/$tdir/test/b || { echo "mkdir fails"; break; }; rm -rf $DIR/$tdir/test || { echo "rmdir fails"; ls -lR $DIR/$tdir; break; }; touch $DIR/$tdir/test1/a || { echo "touch fails"; break; }; mkdir $DIR/$tdir/test1/b || { echo "mkdir fails"; break; }; rm -rf $DIR/$tdir/test1 || { echo "rmdir fails"; ls -lR $DIR/$tdir/test1; break; }; done ) (wd: ~) PASS 70d (313s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 70e: rename cross-MDT with random fails ========================================================== 20:10:51 (1713485451) debug=+ha Starting client oleg154-client.virtnet: -o user_xattr,flock oleg154-server@tcp:/lustre /mnt/lustre Started clients oleg154-client.virtnet: 192.168.201.154@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Started PID=26581 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 4556 1283132 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 3768 1283920 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3603780 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3604120 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7207900 1% /mnt/lustre test_70e fail mds2 1 times Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 20:13:05 (1713485585) shut down Failover mds2 to oleg154-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0001 20:13:19 (1713485599) targets are mounted 20:13:19 (1713485599) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 4744 1282944 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 3452 1284236 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3603780 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3604120 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7207900 1% /mnt/lustre test_70e fail mds2 2 times Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 20:15:39 (1713485739) shut down Failover mds2 to oleg154-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0001 20:15:53 (1713485753) targets are mounted 20:15:53 (1713485753) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 70e (310s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 70f: OSS O_DIRECT recovery with 1 clients ========================================================== 20:16:02 (1713485762) mount clients oleg154-client.virtnet ... Starting client oleg154-client.virtnet: -o user_xattr,flock oleg154-server@tcp:/lustre /mnt/lustre Started clients oleg154-client.virtnet: 192.168.201.154@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 4812 1282876 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 3452 1284236 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3586776 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3593024 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7179800 1% /mnt/lustre ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... test_70f failing OST 1 times Failing ost1 on oleg154-server Stopping /mnt/lustre-ost1 (opts:) on oleg154-server 20:16:11 (1713485771) shut down Failover ost1 to oleg154-server mount facets: ost1 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... Started lustre-OST0000 20:16:26 (1713485786) targets are mounted 20:16:26 (1713485786) facet_failover done ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 4812 1282876 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 3452 1284236 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3571152 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3584712 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7155864 1% /mnt/lustre ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... test_70f failing OST 2 times Failing ost1 on oleg154-server Stopping /mnt/lustre-ost1 (opts:) on oleg154-server 20:16:40 (1713485800) shut down Failover ost1 to oleg154-server mount facets: ost1 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... Started lustre-OST0000 20:16:55 (1713485815) targets are mounted 20:16:55 (1713485815) facet_failover done ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 4812 1282876 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 3452 1284236 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3599256 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3601336 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7200592 1% /mnt/lustre ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... test_70f failing OST 3 times Failing ost1 on oleg154-server Stopping /mnt/lustre-ost1 (opts:) on oleg154-server 20:17:09 (1713485829) shut down Failover ost1 to oleg154-server mount facets: ost1 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Started lustre-OST0000 20:17:24 (1713485844) targets are mounted 20:17:24 (1713485844) facet_failover done Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 4812 1282876 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 3452 1284236 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3561816 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3576400 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7138216 1% /mnt/lustre Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... test_70f failing OST 4 times Failing ost1 on oleg154-server Stopping /mnt/lustre-ost1 (opts:) on oleg154-server 20:17:38 (1713485858) shut down Failover ost1 to oleg154-server mount facets: ost1 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... Started lustre-OST0000 20:17:54 (1713485874) targets are mounted 20:17:54 (1713485874) facet_failover done ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 4812 1282876 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 3452 1284236 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3568008 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3588832 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7156840 1% /mnt/lustre ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... test_70f failing OST 5 times Failing ost1 on oleg154-server Stopping /mnt/lustre-ost1 (opts:) on oleg154-server 20:18:07 (1713485887) shut down Failover ost1 to oleg154-server mount facets: ost1 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-OST0000 20:18:22 (1713485902) targets are mounted 20:18:22 (1713485902) facet_failover done ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... ldlm.namespaces.MGC192.168.201.154@tcp.lru_size=clear ldlm.namespaces.lustre-MDT0000-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-MDT0001-mdc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0000-osc-ffff8800b5819000.lru_size=clear ldlm.namespaces.lustre-OST0001-osc-ffff8800b5819000.lru_size=clear Write/read files in: '/mnt/lustre/d70f.replay-single', clients: 'oleg154-client.virtnet' ... PASS 70f (147s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 71a: mkdir/rmdir striped dir with 2 mdts recovery ========================================================== 20:18:31 (1713485911) Starting client oleg154-client.virtnet: -o user_xattr,flock oleg154-server@tcp:/lustre /mnt/lustre Started clients oleg154-client.virtnet: 192.168.201.154@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Started 21852 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 5576 1282112 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 3272 1284416 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3605448 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 5776 1281912 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 3488 1284200 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3605448 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre fail mds1 mds2 1 times Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 20:20:49 (1713486049) shut down Failover mds1 to oleg154-server mount facets: mds1 Failover mds2 to oleg154-server mount facets: mds2 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0001 Started lustre-MDT0000 20:21:15 (1713486075) targets are mounted 20:21:15 (1713486075) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid,mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3544 1284144 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 4132 1283556 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3605448 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3724 1283964 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2276 1285412 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3605448 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre fail mds1 mds2 2 times Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 20:23:44 (1713486224) shut down Failover mds1 to oleg154-server mount facets: mds1 Failover mds2 to oleg154-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0001 Started lustre-MDT0000 20:24:10 (1713486250) targets are mounted 20:24:10 (1713486250) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid,mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec /home/green/git/lustre-release/lustre/tests/test-framework.sh: line 4697: 21852 Killed ( while true; do $LFS mkdir -i0 -c2 $DIR/$tdir/test; rmdir $DIR/$tdir/test; done ) (wd: ~) PASS 71a (352s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 73a: open(O_CREAT), unlink, replay, reconnect before open replay, close ========================================================== 20:24:25 (1713486265) multiop /mnt/lustre/f73a.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7506 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3912 1283776 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 3288 1284400 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3605448 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre fail_loc=0x80000302 Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 20:24:29 (1713486269) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 20:24:43 (1713486283) targets are mounted 20:24:44 (1713486284) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 73a (39s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 73b: open(O_CREAT), unlink, replay, reconnect at open_replay reply, close ========================================================== 20:25:05 (1713486305) multiop /mnt/lustre/f73b.replay-single vO_tSc TMPPIPE=/tmp/multiop_open_wait_pipe.7506 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2444 1285244 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2012 1285676 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3605448 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre fail_loc=0x80000157 Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 20:25:09 (1713486309) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 20:25:24 (1713486324) targets are mounted 20:25:24 (1713486324) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 73b (39s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 74: Ensure applications don't fail waiting for OST recovery ========================================================== 20:25:45 (1713486345) Stopping clients: oleg154-client.virtnet /mnt/lustre (opts:) Stopping client oleg154-client.virtnet /mnt/lustre opts: Stopping /mnt/lustre-ost1 (opts:) on oleg154-server Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 20:25:49 (1713486349) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 20:26:02 (1713486362) targets are mounted 20:26:02 (1713486362) facet_failover done Starting client oleg154-client.virtnet: -o user_xattr,flock oleg154-server@tcp:/lustre /mnt/lustre Started clients oleg154-client.virtnet: 192.168.201.154@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-OST0000 PASS 74 (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80a: DNE: create remote dir, drop update rep from MDT0, fail MDT0 ========================================================== 20:26:15 (1713486375) fail_loc=0x1701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2448 1285240 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2052 1285636 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3605448 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 20:26:19 (1713486379) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 20:26:33 (1713486393) targets are mounted 20:26:33 (1713486393) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec total: 20 open/close in 0.11 seconds: 189.86 ops/second PASS 80a (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80b: DNE: create remote dir, drop update rep from MDT0, fail MDT1 ========================================================== 20:26:43 (1713486403) fail_loc=0x1701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2488 1285200 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2052 1285636 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3605448 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2488 1285200 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2052 1285636 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3605448 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 20:26:54 (1713486414) shut down Failover mds2 to oleg154-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0001 20:27:09 (1713486429) targets are mounted 20:27:09 (1713486429) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec total: 20 open/close in 0.09 seconds: 231.47 ops/second PASS 80b (33s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80c: DNE: create remote dir, drop update rep from MDT1, fail MDT[0,1] ========================================================== 20:27:18 (1713486438) fail_loc=0x1701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2524 1285164 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2088 1285600 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3605448 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2524 1285164 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2088 1285600 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3605448 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 20:27:25 (1713486445) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 20:27:39 (1713486459) targets are mounted 20:27:39 (1713486459) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 20:27:46 (1713486466) shut down Failover mds2 to oleg154-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0001 20:28:00 (1713486480) targets are mounted 20:28:00 (1713486480) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec total: 20 open/close in 0.05 seconds: 363.69 ops/second PASS 80c (50s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80d: DNE: create remote dir, drop update rep from MDT1, fail 2 MDTs ========================================================== 20:28:09 (1713486489) fail_loc=0x1701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2560 1285128 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2052 1285636 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3605448 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2560 1285128 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2052 1285636 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3605448 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 20:28:24 (1713486504) shut down Failover mds1 to oleg154-server mount facets: mds1 Failover mds2 to oleg154-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0001 Started lustre-MDT0000 20:28:44 (1713486524) targets are mounted 20:28:44 (1713486524) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid,mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec total: 20 open/close in 0.07 seconds: 279.17 ops/second PASS 80d (43s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80e: DNE: create remote dir, drop MDT1 rep, fail MDT0 ========================================================== 20:28:54 (1713486534) fail_loc=0x119 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2488 1285200 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2052 1285636 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3605448 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 20:29:00 (1713486540) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 20:29:14 (1713486554) targets are mounted 20:29:14 (1713486554) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec total: 20 open/close in 0.05 seconds: 407.61 ops/second PASS 80e (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80f: DNE: create remote dir, drop MDT1 rep, fail MDT1 ========================================================== 20:29:23 (1713486563) fail_loc=0x119 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2488 1285200 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2052 1285636 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3605448 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 20:29:27 (1713486567) shut down Failover mds2 to oleg154-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0001 20:29:41 (1713486581) targets are mounted 20:29:41 (1713486581) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec total: 20 open/close in 0.05 seconds: 393.66 ops/second PASS 80f (25s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80g: DNE: create remote dir, drop MDT1 rep, fail MDT0, then MDT1 ========================================================== 20:29:49 (1713486589) fail_loc=0x119 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2488 1285200 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2052 1285636 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3605448 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2488 1285200 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2052 1285636 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3605448 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 20:29:58 (1713486598) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 20:30:12 (1713486612) targets are mounted 20:30:12 (1713486612) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 20:30:19 (1713486619) shut down Failover mds2 to oleg154-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0001 20:30:33 (1713486633) targets are mounted 20:30:33 (1713486633) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec total: 20 open/close in 0.05 seconds: 366.89 ops/second PASS 80g (51s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 80h: DNE: create remote dir, drop MDT1 rep, fail 2 MDTs ========================================================== 20:30:42 (1713486642) fail_loc=0x119 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2488 1285200 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2052 1285636 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3605448 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2488 1285200 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2052 1285636 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3605448 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 20:30:52 (1713486652) shut down Failover mds1 to oleg154-server mount facets: mds1 Failover mds2 to oleg154-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 Started lustre-MDT0001 20:31:18 (1713486678) targets are mounted 20:31:18 (1713486678) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid,mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec total: 20 open/close in 0.06 seconds: 347.73 ops/second PASS 80h (46s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81a: DNE: unlink remote dir, drop MDT0 update rep, fail MDT1 ========================================================== 20:31:29 (1713486689) fail_loc=0x1701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2484 1285204 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2052 1285636 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3605448 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 20:31:39 (1713486699) shut down Failover mds2 to oleg154-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0001 20:31:53 (1713486713) targets are mounted 20:31:53 (1713486713) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 81a (31s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81b: DNE: unlink remote dir, drop MDT0 update reply, fail MDT0 ========================================================== 20:32:02 (1713486722) fail_loc=0x1701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2520 1285168 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2056 1285632 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3605448 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 20:32:06 (1713486726) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 20:32:20 (1713486740) targets are mounted 20:32:20 (1713486740) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 81b (25s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81c: DNE: unlink remote dir, drop MDT0 update reply, fail MDT0,MDT1 ========================================================== 20:32:29 (1713486749) fail_loc=0x1701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2484 1285204 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2056 1285632 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3605448 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2484 1285204 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2056 1285632 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3605448 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 20:32:35 (1713486755) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 20:32:49 (1713486769) targets are mounted 20:32:49 (1713486769) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 20:32:56 (1713486776) shut down Failover mds2 to oleg154-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0001 20:33:10 (1713486790) targets are mounted 20:33:10 (1713486790) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 81c (48s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81d: DNE: unlink remote dir, drop MDT0 update reply, fail 2 MDTs ========================================================== 20:33:19 (1713486799) fail_loc=0x1701 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2520 1285168 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2060 1285628 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3605448 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2520 1285168 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2060 1285628 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3605448 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 20:33:31 (1713486811) shut down Failover mds1 to oleg154-server mount facets: mds1 Failover mds2 to oleg154-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0001 Started lustre-MDT0000 20:33:51 (1713486831) targets are mounted 20:33:51 (1713486831) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid,mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 81d (41s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81e: DNE: unlink remote dir, drop MDT1 req reply, fail MDT0 ========================================================== 20:34:01 (1713486841) fail_loc=0x119 fail_loc=0 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2484 1285204 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2060 1285628 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3605448 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 20:34:06 (1713486846) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 20:34:20 (1713486860) targets are mounted 20:34:20 (1713486860) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 81e (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81f: DNE: unlink remote dir, drop MDT1 req reply, fail MDT1 ========================================================== 20:34:28 (1713486868) fail_loc=0x119 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2484 1285204 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2060 1285628 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3605448 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 20:34:32 (1713486872) shut down Failover mds2 to oleg154-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0001 20:34:46 (1713486886) targets are mounted 20:34:46 (1713486886) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 81f (25s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81g: DNE: unlink remote dir, drop req reply, fail M0, then M1 ========================================================== 20:34:55 (1713486895) fail_loc=0x119 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2484 1285204 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2060 1285628 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3605448 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2484 1285204 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2060 1285628 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3605448 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 20:35:00 (1713486900) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 20:35:14 (1713486914) targets are mounted 20:35:14 (1713486914) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 20:35:21 (1713486921) shut down Failover mds2 to oleg154-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0001 20:35:36 (1713486936) targets are mounted 20:35:36 (1713486936) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 81g (48s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 81h: DNE: unlink remote dir, drop request reply, fail 2 MDTs ========================================================== 20:35:44 (1713486944) fail_loc=0x119 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2484 1285204 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2060 1285628 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3605448 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2484 1285204 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2060 1285628 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1572 3605448 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1564 3605456 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3136 7210904 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 20:35:51 (1713486951) shut down Failover mds1 to oleg154-server mount facets: mds1 Failover mds2 to oleg154-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 Started lustre-MDT0001 20:36:16 (1713486976) targets are mounted 20:36:16 (1713486976) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid,mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 81h (46s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 84a: stale open during export disconnect ========================================================== 20:36:31 (1713486991) fail_loc=0x80000144 total: 1 open/close in 0.01 seconds: 163.22 ops/second oleg154-client: error: invalid path '/mnt/lustre': Input/output error pdsh@oleg154-client: oleg154-client: ssh exited with exit code 5 PASS 84a (5s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 85a: check the cancellation of unused locks during recovery(IBITS) ========================================================== 20:36:38 (1713486998) before recovery: unused locks count = 201 Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 20:36:40 (1713487000) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 20:36:53 (1713487013) targets are mounted 20:36:53 (1713487013) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec after recovery: unused locks count = 101 PASS 85a (23s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 85b: check the cancellation of unused locks during recovery(EXTENT) ========================================================== 20:37:01 (1713487021) before recovery: unused locks count = 100 Failing ost1 on oleg154-server Stopping /mnt/lustre-ost1 (opts:) on oleg154-server 20:37:05 (1713487025) shut down Failover ost1 to oleg154-server mount facets: ost1 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-OST0000 20:37:18 (1713487038) targets are mounted 20:37:18 (1713487038) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec after recovery: unused locks count = 0 PASS 85b (22s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 86: umount server after clear nid_stats should not hit LBUG ========================================================== 20:37:25 (1713487045) Stopping clients: oleg154-client.virtnet /mnt/lustre (opts:) Stopping client oleg154-client.virtnet /mnt/lustre opts: mdt.lustre-MDT0000.exports.clear=0 mdt.lustre-MDT0001.exports.clear=0 Stopping /mnt/lustre-mds1 (opts:) on oleg154-server Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 Starting client oleg154-client.virtnet: -o user_xattr,flock oleg154-server@tcp:/lustre /mnt/lustre Started clients oleg154-client.virtnet: 192.168.201.154@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) PASS 86 (11s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 87a: write replay ================== 20:37:37 (1713487057) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2480 1285208 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2064 1285624 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 1772 3605248 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1764 3605256 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 3536 7210504 1% /mnt/lustre 8+0 records in 8+0 records out 8388608 bytes (8.4 MB) copied, 0.137306 s, 61.1 MB/s Failing ost1 on oleg154-server Stopping /mnt/lustre-ost1 (opts:) on oleg154-server 20:37:41 (1713487061) shut down Failover ost1 to oleg154-server mount facets: ost1 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-OST0000 20:37:55 (1713487075) targets are mounted 20:37:55 (1713487075) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec 8+0 records in 8+0 records out 8388608 bytes (8.4 MB) copied, 0.0419104 s, 200 MB/s PASS 87a (23s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 87b: write replay with changed data (checksum resend) ========================================================== 20:38:02 (1713487082) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2480 1285208 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2064 1285624 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 9964 3597056 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1764 3605256 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 11728 7202312 1% /mnt/lustre 8+0 records in 8+0 records out 8388608 bytes (8.4 MB) copied, 0.127814 s, 65.6 MB/s 8+0 records in 8+0 records out 8 bytes (8 B) copied, 0.00157162 s, 5.1 kB/s Failing ost1 on oleg154-server Stopping /mnt/lustre-ost1 (opts:) on oleg154-server 20:38:06 (1713487086) shut down Failover ost1 to oleg154-server mount facets: ost1 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-OST0000 20:38:21 (1713487101) targets are mounted 20:38:21 (1713487101) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec 0+1 records in 0+1 records out 72 bytes (72 B) copied, 0.0025337 s, 28.4 kB/s PASS 87b (23s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 88: MDS should not assign same objid to different files ========================================================== 20:38:27 (1713487107) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2484 1285204 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2064 1285624 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 9968 3597052 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1764 3605256 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 11732 7202308 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2484 1285204 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2064 1285624 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 9968 3597052 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1764 3605256 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 11732 7202308 1% /mnt/lustre before test: last_id = 14369, next_id = 14340 Creating to objid 14369 on ost lustre-OST0000... total: 31 open/close in 0.07 seconds: 422.58 ops/second total: 8 open/close in 0.02 seconds: 420.41 ops/second before recovery: last_id = 14401, next_id = 14378 Stopping /mnt/lustre-mds1 (opts:) on oleg154-server Stopping /mnt/lustre-ost1 (opts:) on oleg154-server Failover mds1 to oleg154-server Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 Failover ost1 to oleg154-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-OST0000 after recovery: last_id = 14409, next_id = 14378 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0248959 s, 21.1 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0236589 s, 22.2 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0265071 s, 19.8 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0222995 s, 23.5 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0286119 s, 18.3 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.023984 s, 21.9 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.026489 s, 19.8 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0321838 s, 16.3 MB/s -rw-r--r-- 1 root root 0 Apr 18 20:38 /mnt/lustre/d88.replay-single/f-14340 -rw-r--r-- 1 root root 0 Apr 18 20:38 /mnt/lustre/d88.replay-single/f-14341 -rw-r--r-- 1 root root 0 Apr 18 20:38 /mnt/lustre/d88.replay-single/f-14342 -rw-r--r-- 1 root root 0 Apr 18 20:38 /mnt/lustre/d88.replay-single/f-14343 -rw-r--r-- 1 root root 0 Apr 18 20:38 /mnt/lustre/d88.replay-single/f-14344 -rw-r--r-- 1 root root 0 Apr 18 20:38 /mnt/lustre/d88.replay-single/f-14345 -rw-r--r-- 1 root root 0 Apr 18 20:38 /mnt/lustre/d88.replay-single/f-14346 -rw-r--r-- 1 root root 0 Apr 18 20:38 /mnt/lustre/d88.replay-single/f-14347 -rw-r--r-- 1 root root 0 Apr 18 20:38 /mnt/lustre/d88.replay-single/f-14348 -rw-r--r-- 1 root root 0 Apr 18 20:38 /mnt/lustre/d88.replay-single/f-14349 -rw-r--r-- 1 root root 0 Apr 18 20:38 /mnt/lustre/d88.replay-single/f-14350 -rw-r--r-- 1 root root 0 Apr 18 20:38 /mnt/lustre/d88.replay-single/f-14351 -rw-r--r-- 1 root root 0 Apr 18 20:38 /mnt/lustre/d88.replay-single/f-14352 -rw-r--r-- 1 root root 0 Apr 18 20:38 /mnt/lustre/d88.replay-single/f-14353 -rw-r--r-- 1 root root 0 Apr 18 20:38 /mnt/lustre/d88.replay-single/f-14354 -rw-r--r-- 1 root root 0 Apr 18 20:38 /mnt/lustre/d88.replay-single/f-14355 -rw-r--r-- 1 root root 0 Apr 18 20:38 /mnt/lustre/d88.replay-single/f-14356 -rw-r--r-- 1 root root 0 Apr 18 20:38 /mnt/lustre/d88.replay-single/f-14357 -rw-r--r-- 1 root root 0 Apr 18 20:38 /mnt/lustre/d88.replay-single/f-14358 -rw-r--r-- 1 root root 0 Apr 18 20:38 /mnt/lustre/d88.replay-single/f-14359 -rw-r--r-- 1 root root 0 Apr 18 20:38 /mnt/lustre/d88.replay-single/f-14360 -rw-r--r-- 1 root root 0 Apr 18 20:38 /mnt/lustre/d88.replay-single/f-14361 -rw-r--r-- 1 root root 0 Apr 18 20:38 /mnt/lustre/d88.replay-single/f-14362 -rw-r--r-- 1 root root 0 Apr 18 20:38 /mnt/lustre/d88.replay-single/f-14363 -rw-r--r-- 1 root root 0 Apr 18 20:38 /mnt/lustre/d88.replay-single/f-14364 -rw-r--r-- 1 root root 0 Apr 18 20:38 /mnt/lustre/d88.replay-single/f-14365 -rw-r--r-- 1 root root 0 Apr 18 20:38 /mnt/lustre/d88.replay-single/f-14366 -rw-r--r-- 1 root root 0 Apr 18 20:38 /mnt/lustre/d88.replay-single/f-14367 -rw-r--r-- 1 root root 0 Apr 18 20:38 /mnt/lustre/d88.replay-single/f-14368 -rw-r--r-- 1 root root 0 Apr 18 20:38 /mnt/lustre/d88.replay-single/f-14369 -rw-r--r-- 1 root root 0 Apr 18 20:38 /mnt/lustre/d88.replay-single/f-14370 -rw-r--r-- 1 root root 0 Apr 18 20:38 /mnt/lustre/d88.replay-single/f-14371 -rw-r--r-- 1 root root 0 Apr 18 20:38 /mnt/lustre/d88.replay-single/f-14372 -rw-r--r-- 1 root root 0 Apr 18 20:38 /mnt/lustre/d88.replay-single/f-14373 -rw-r--r-- 1 root root 0 Apr 18 20:38 /mnt/lustre/d88.replay-single/f-14374 -rw-r--r-- 1 root root 0 Apr 18 20:38 /mnt/lustre/d88.replay-single/f-14375 -rw-r--r-- 1 root root 0 Apr 18 20:38 /mnt/lustre/d88.replay-single/f-14376 -rw-r--r-- 1 root root 0 Apr 18 20:38 /mnt/lustre/d88.replay-single/f-14377 -rw-r--r-- 1 root root 524288 Apr 18 20:39 /mnt/lustre/d88.replay-single/f-14381 -rw-r--r-- 1 root root 524288 Apr 18 20:39 /mnt/lustre/d88.replay-single/f-14382 -rw-r--r-- 1 root root 524288 Apr 18 20:39 /mnt/lustre/d88.replay-single/f-14383 -rw-r--r-- 1 root root 524288 Apr 18 20:39 /mnt/lustre/d88.replay-single/f-14384 -rw-r--r-- 1 root root 524288 Apr 18 20:39 /mnt/lustre/d88.replay-single/f-14385 -rw-r--r-- 1 root root 524288 Apr 18 20:39 /mnt/lustre/d88.replay-single/f-14386 -rw-r--r-- 1 root root 524288 Apr 18 20:39 /mnt/lustre/d88.replay-single/f-14387 -rw-r--r-- 1 root root 524288 Apr 18 20:39 /mnt/lustre/d88.replay-single/f-14388 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0256393 s, 20.4 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0244547 s, 21.4 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.029163 s, 18.0 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0241809 s, 21.7 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0277621 s, 18.9 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0236003 s, 22.2 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0251699 s, 20.8 MB/s 128+0 records in 128+0 records out 524288 bytes (524 kB) copied, 0.0251573 s, 20.8 MB/s PASS 88 (51s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 89: no disk space leak on late ost connection ========================================================== 20:39:20 (1713487160) Waiting for orphan cleanup... osp.lustre-OST0000-osc-MDT0000.old_sync_processed osp.lustre-OST0000-osc-MDT0001.old_sync_processed osp.lustre-OST0001-osc-MDT0000.old_sync_processed osp.lustre-OST0001-osc-MDT0001.old_sync_processed wait 40 secs maximumly for oleg154-server mds-ost sync done. Waiting for MDT destroys to complete 10+0 records in 10+0 records out 10485760 bytes (10 MB) copied, 0.0726639 s, 144 MB/s Stopping /mnt/lustre-ost1 (opts:) on oleg154-server Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 20:39:26 (1713487166) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 20:39:38 (1713487178) targets are mounted 20:39:38 (1713487178) facet_failover done Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-OST0000 Starting client: oleg154-client.virtnet: -o user_xattr,flock oleg154-server@tcp:/lustre /mnt/lustre osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 68 sec Waiting for orphan cleanup... osp.lustre-OST0000-osc-MDT0000.old_sync_processed osp.lustre-OST0000-osc-MDT0001.old_sync_processed osp.lustre-OST0001-osc-MDT0000.old_sync_processed osp.lustre-OST0001-osc-MDT0001.old_sync_processed wait 40 secs maximumly for oleg154-server mds-ost sync done. Waiting for MDT destroys to complete free_before: 7646308 free_after: 7646308 PASS 89 (103s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 90: lfs find identifies the missing striped file segments ========================================================== 20:41:05 (1713487265) Create the files Fail ost1 lustre-OST0000_UUID, display the list of affected files Stopping /mnt/lustre-ost1 (opts:) on oleg154-server General Query: lfs find /mnt/lustre/d90.replay-single /mnt/lustre/d90.replay-single /mnt/lustre/d90.replay-single/f0 /mnt/lustre/d90.replay-single/all /mnt/lustre/d90.replay-single/f1 Querying files on shutdown ost1: lfs find --obd lustre-OST0000_UUID /mnt/lustre/d90.replay-single/f0 /mnt/lustre/d90.replay-single/all Check getstripe: /home/green/git/lustre-release/lustre/utils/lfs getstripe -r --obd lustre-OST0000_UUID /mnt/lustre/d90.replay-single/f0 lmm_stripe_count: 1 lmm_stripe_size: 4194304 lmm_pattern: raid0 lmm_layout_gen: 0 lmm_stripe_offset: 0 obdidx objid objid group 0 14411 0x384b 0x280000401 * /mnt/lustre/d90.replay-single/all lmm_stripe_count: 2 lmm_stripe_size: 4194304 lmm_pattern: raid0 lmm_layout_gen: 0 lmm_stripe_offset: 0 obdidx objid objid group 0 14410 0x384a 0x280000401 * /mnt/lustre/d90.replay-single/all /mnt/lustre/d90.replay-single/f0 Failover ost1 to oleg154-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-OST0000 PASS 90 (18s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 93a: replay + reconnect ============ 20:41:25 (1713487285) 1+0 records in 1+0 records out 1024 bytes (1.0 kB) copied, 0.00156414 s, 655 kB/s fail_val=40 fail_loc=0x715 Failing ost1 on oleg154-server Stopping /mnt/lustre-ost1 (opts:) on oleg154-server 20:41:27 (1713487287) shut down Failover ost1 to oleg154-server mount facets: ost1 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-OST0000 20:41:40 (1713487300) targets are mounted 20:41:40 (1713487300) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec PASS 93a (59s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 93b: replay + reconnect on mds ===== 20:42:26 (1713487346) total: 20 open/close in 0.07 seconds: 298.31 ops/second fail_val=80 fail_loc=0x715 Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 20:42:27 (1713487347) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 20:42:41 (1713487361) targets are mounted 20:42:41 (1713487361) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 93b (102s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 100a: DNE: create striped dir, drop update rep from MDT1, fail MDT1 ========================================================== 20:44:09 (1713487449) fail_loc=0x1701 Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 20:44:11 (1713487451) shut down Failover mds2 to oleg154-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0001 20:44:24 (1713487464) targets are mounted 20:44:24 (1713487464) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec lmv_stripe_count: 2 lmv_stripe_offset: 0 lmv_hash_type: crush mdtidx FID[seq:oid:ver] 0 [0x2000320e0:0x1:0x0] 1 [0x24000cb3a:0x1:0x0] total: 20 open/close in 0.05 seconds: 394.41 ops/second PASS 100a (22s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 100b: DNE: create striped dir, fail MDT0 ========================================================== 20:44:33 (1713487473) fail_loc=0x119 Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 20:44:34 (1713487474) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 20:44:47 (1713487487) targets are mounted 20:44:47 (1713487487) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec lmv_stripe_count: 2 lmv_stripe_offset: 0 lmv_hash_type: crush mdtidx FID[seq:oid:ver] 0 [0x2000320e0:0x4:0x0] 1 [0x24000cb3a:0x4:0x0] total: 20 open/close in 0.05 seconds: 422.00 ops/second PASS 100b (21s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 100c: DNE: create striped dir, abort_recov_mdt mds2 ========================================================== 20:44:56 (1713487496) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2568 1285120 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2136 1285552 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18164 3588856 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1764 3605256 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19928 7194112 1% /mnt/lustre Stopping /mnt/lustre-mds2 (opts:) on oleg154-server Failover mds2 to oleg154-server Starting mds2: -o localrecov -o abort_recov_mdt /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0001 total: 20 open/close in 0.05 seconds: 392.63 ops/second Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 20:45:15 (1713487515) shut down Failover mds2 to oleg154-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0001 20:45:28 (1713487528) targets are mounted 20:45:28 (1713487528) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec lmv_stripe_count: 2 lmv_stripe_offset: 1 lmv_hash_type: crush mdtidx FID[seq:oid:ver] 1 [0x24000d308:0x3:0x0] 0 [0x2000320e1:0x3:0x0] total: 20 open/close in 0.05 seconds: 433.17 ops/second PASS 100c (39s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 100d: DNE: cancel update logs upon recovery abort ========================================================== 20:45:36 (1713487536) striped dir -i0 -c2 -H fnv_1a_64 /mnt/lustre/d100d.replay-single total: 100 mkdir in 0.27 seconds: 372.04 ops/second lustre-MDT0000-osd [catalog]: [0x20001a210:0x1:0x0] [index]: 00019 [logid]: [0x2000328b0:0x1:0x0] lustre-MDT0001-osp-MDT0000 [catalog]: [0x2400007ec:0x1:0x0] [index]: 00026 [logid]: [0x24000d309:0x1:0x0] [index]: 00027 [logid]: [0x24000d309:0x2:0x0] Stopping /mnt/lustre-mds1 (opts:) on oleg154-server Failover mds1 to oleg154-server Starting mds1: -o localrecov -o abort_recovery /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 find: '/mnt/lustre/d100d.replay-single': No such file or directory find: '/mnt/lustre/d100d.replay-single': No such file or directory PASS 100d (18s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 100e: DNE: create striped dir on MDT0 and MDT1, fail MDT0, MDT1 ========================================================== 20:45:56 (1713487556) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2992 1284696 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2552 1285136 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18164 3588856 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1764 3605256 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19928 7194112 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2992 1284696 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2552 1285136 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18164 3588856 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1764 3605256 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19928 7194112 1% /mnt/lustre lmv_stripe_count: 2 lmv_stripe_offset: 0 lmv_hash_type: crush mdtidx FID[seq:oid:ver] 0 [0x200033080:0x2:0x0] 1 [0x24000e2a9:0x2:0x0] Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 20:46:03 (1713487563) shut down Failover mds1 to oleg154-server mount facets: mds1 Failover mds2 to oleg154-server mount facets: mds2 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 Started lustre-MDT0001 20:46:29 (1713487589) targets are mounted 20:46:29 (1713487589) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid,mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec lmv_stripe_count: 2 lmv_stripe_offset: 0 lmv_hash_type: crush mdtidx FID[seq:oid:ver] 0 [0x200033080:0x2:0x0] 1 [0x24000e2a9:0x2:0x0] PASS 100e (38s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 101: Shouldn't reassign precreated objs to other files after recovery ========================================================== 20:46:36 (1713487596) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3072 1284616 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2624 1285064 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18164 3588856 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1764 3605256 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19928 7194112 1% /mnt/lustre Stopping /mnt/lustre-mds1 (opts:) on oleg154-server Failover mds1 to oleg154-server Starting mds1: -o localrecov -o abort_recovery /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 pdsh@oleg154-client: oleg154-client: ssh exited with exit code 5 first stat failed: 5 PASS 101 (34s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 102a: check resend (request lost) with multiple modify RPCs in flight ========================================================== 20:47:12 (1713487632) creating 7 files ... fail_loc=0x159 launch 7 chmod in parallel (20:47:13) ... fail_loc=0 done (20:47:29) /mnt/lustre/d102a.replay-single/file-1 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-2 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-3 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-4 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-5 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-6 has perms 0600 OK /mnt/lustre/d102a.replay-single/file-7 has perms 0600 OK PASS 102a (18s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 102b: check resend (reply lost) with multiple modify RPCs in flight ========================================================== 20:47:32 (1713487652) creating 7 files ... fail_loc=0x15a launch 7 chmod in parallel (20:47:33) ... fail_loc=0 done (20:47:49) /mnt/lustre/d102b.replay-single/file-1 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-2 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-3 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-4 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-5 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-6 has perms 0600 OK /mnt/lustre/d102b.replay-single/file-7 has perms 0600 OK PASS 102b (18s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 102c: check replay w/o reconstruction with multiple mod RPCs in flight ========================================================== 20:47:52 (1713487672) creating 7 files ... UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2984 1284704 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2468 1285220 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18164 3580652 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1764 3597052 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19928 7177704 1% /mnt/lustre fail_loc=0x15a launch 7 chmod in parallel (20:47:54) ... fail_loc=0 Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 20:47:57 (1713487677) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 20:48:11 (1713487691) targets are mounted 20:48:11 (1713487691) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec done (20:48:17) /mnt/lustre/d102c.replay-single/file-1 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-2 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-3 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-4 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-5 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-6 has perms 0600 OK /mnt/lustre/d102c.replay-single/file-7 has perms 0600 OK PASS 102c (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 102d: check replay & reconstruction with multiple mod RPCs in flight ========================================================== 20:48:20 (1713487700) creating 7 files ... fail_loc=0x15a launch 7 chmod in parallel (20:48:20) ... fail_loc=0 Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 20:48:23 (1713487703) shut down Failover mds2 to oleg154-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0001 20:48:36 (1713487716) targets are mounted 20:48:36 (1713487716) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec done (20:48:42) /mnt/lustre/d102d.replay-single/file-1 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-2 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-3 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-4 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-5 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-6 has perms 0600 OK /mnt/lustre/d102d.replay-single/file-7 has perms 0600 OK PASS 102d (23s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 103: Check otr_next_id overflow ==== 20:48:45 (1713487725) fail_loc=0x80000162 total: 30 open/close in 0.08 seconds: 385.68 ops/second Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 20:48:47 (1713487727) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 20:49:00 (1713487740) targets are mounted 20:49:00 (1713487740) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 103 (22s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 110a: DNE: create striped dir, fail MDT1 ========================================================== 20:49:09 (1713487749) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2936 1284752 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2472 1285216 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18164 3580652 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1764 3597052 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19928 7177704 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 20:49:13 (1713487753) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 20:49:27 (1713487767) targets are mounted 20:49:27 (1713487767) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d110a.replay-single/striped_dir has type dir OK PASS 110a (25s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 110b: DNE: create striped dir, fail MDT1 and client ========================================================== 20:49:36 (1713487776) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2936 1284752 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2472 1285216 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18164 3580652 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1764 3597052 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19928 7177704 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 20:49:40 (1713487780) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 20:49:55 (1713487795) targets are mounted 20:49:55 (1713487795) facet_failover done pdsh@oleg154-client: oleg154-client: ssh exited with exit code 95 oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid pdsh@oleg154-client: oleg154-client: ssh exited with exit code 95 Starting client: oleg154-client.virtnet: -o user_xattr,flock oleg154-server@tcp:/lustre /mnt/lustre /mnt/lustre/d110b.replay-single/striped_dir has type dir OK PASS 110b (95s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 110c: DNE: create striped dir, fail MDT2 ========================================================== 20:51:12 (1713487872) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2940 1284748 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2476 1285212 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18164 3588856 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1764 3605256 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19928 7194112 1% /mnt/lustre Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 20:51:16 (1713487876) shut down Failover mds2 to oleg154-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0001 20:51:31 (1713487891) targets are mounted 20:51:31 (1713487891) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d110c.replay-single/striped_dir has type dir OK PASS 110c (25s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 110d: DNE: create striped dir, fail MDT2 and client ========================================================== 20:51:39 (1713487899) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2984 1284704 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2520 1285168 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18164 3588856 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1764 3605256 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19928 7194112 1% /mnt/lustre Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 20:51:43 (1713487903) shut down Failover mds2 to oleg154-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0001 20:51:57 (1713487917) targets are mounted 20:51:57 (1713487917) facet_failover done pdsh@oleg154-client: oleg154-client: ssh exited with exit code 95 oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid pdsh@oleg154-client: oleg154-client: ssh exited with exit code 95 Starting client: oleg154-client.virtnet: -o user_xattr,flock oleg154-server@tcp:/lustre /mnt/lustre /mnt/lustre/d110d.replay-single/striped_dir has type dir OK PASS 110d (92s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 110e: DNE: create striped dir, uncommit on MDT2, fail client/MDT1/MDT2 ========================================================== 20:53:13 (1713487993) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3020 1284668 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2552 1285136 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18164 3588856 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1764 3605256 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19928 7194112 1% /mnt/lustre df:/mnt/lustre Not a Lustre filesystem Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 20:53:19 (1713487999) shut down Failover mds1 to oleg154-server mount facets: mds1 Failover mds2 to oleg154-server mount facets: mds2 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0001 Started lustre-MDT0000 20:53:45 (1713488025) targets are mounted 20:53:45 (1713488025) facet_failover done pdsh@oleg154-client: oleg154-client: ssh exited with exit code 95 oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid,mdc.lustre-MDT0001-mdc-*.mds_server_uuid pdsh@oleg154-client: oleg154-client: ssh exited with exit code 95 Starting client: oleg154-client.virtnet: -o user_xattr,flock oleg154-server@tcp:/lustre /mnt/lustre /mnt/lustre/d110e.replay-single/striped_dir has type dir OK PASS 110e (106s) debug_raw_pointers=0 debug_raw_pointers=0 SKIP: replay-single test_110f skipping excluded test 110f debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 110g: DNE: create striped dir, uncommit on MDT1, fail client/MDT1/MDT2 ========================================================== 20:55:01 (1713488101) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3060 1284628 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2588 1285100 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18164 3588856 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1764 3605256 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19928 7194112 1% /mnt/lustre df:/mnt/lustre Not a Lustre filesystem Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 20:55:13 (1713488113) shut down Failover mds1 to oleg154-server mount facets: mds1 Failover mds2 to oleg154-server mount facets: mds2 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0001 Started lustre-MDT0000 20:55:33 (1713488133) targets are mounted 20:55:33 (1713488133) facet_failover done pdsh@oleg154-client: oleg154-client: ssh exited with exit code 95 oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid,mdc.lustre-MDT0001-mdc-*.mds_server_uuid pdsh@oleg154-client: oleg154-client: ssh exited with exit code 95 Starting client: oleg154-client.virtnet: -o user_xattr,flock oleg154-server@tcp:/lustre /mnt/lustre /mnt/lustre/d110g.replay-single/striped_dir has type dir OK PASS 110g (108s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 111a: DNE: unlink striped dir, fail MDT1 ========================================================== 20:56:51 (1713488211) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3096 1284592 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2480 1285208 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18164 3588856 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1764 3605256 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19928 7194112 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 20:56:54 (1713488214) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 20:57:08 (1713488228) targets are mounted 20:57:08 (1713488228) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/d111a.replay-single/striped_dir: No such file or directory PASS 111a (24s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 111b: DNE: unlink striped dir, fail MDT2 ========================================================== 20:57:17 (1713488237) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2952 1284736 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2488 1285200 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18164 3588856 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1764 3605256 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19928 7194112 1% /mnt/lustre Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 20:57:21 (1713488241) shut down Failover mds2 to oleg154-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0001 20:57:34 (1713488254) targets are mounted 20:57:34 (1713488254) facet_failover done pdsh@oleg154-client: oleg154-client: ssh exited with exit code 95 oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid pdsh@oleg154-client: oleg154-client: ssh exited with exit code 95 Starting client: oleg154-client.virtnet: -o user_xattr,flock oleg154-server@tcp:/lustre /mnt/lustre Can't lstat /mnt/lustre/d111b.replay-single/striped_dir: No such file or directory PASS 111b (92s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 111c: DNE: unlink striped dir, uncommit on MDT1, fail client/MDT1/MDT2 ========================================================== 20:58:50 (1713488330) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2996 1284692 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2484 1285204 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18164 3588856 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1764 3605256 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19928 7194112 1% /mnt/lustre df:/mnt/lustre Not a Lustre filesystem Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 20:59:02 (1713488342) shut down Failover mds1 to oleg154-server mount facets: mds1 Failover mds2 to oleg154-server mount facets: mds2 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0001 Started lustre-MDT0000 20:59:23 (1713488363) targets are mounted 20:59:23 (1713488363) facet_failover done pdsh@oleg154-client: oleg154-client: ssh exited with exit code 95 oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid,mdc.lustre-MDT0001-mdc-*.mds_server_uuid pdsh@oleg154-client: oleg154-client: ssh exited with exit code 95 Starting client: oleg154-client.virtnet: -o user_xattr,flock oleg154-server@tcp:/lustre /mnt/lustre Can't lstat /mnt/lustre/d111c.replay-single/striped_dir: No such file or directory PASS 111c (109s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 111d: DNE: unlink striped dir, uncommit on MDT2, fail client/MDT1/MDT2 ========================================================== 21:00:42 (1713488442) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2956 1284732 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2484 1285204 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18164 3588856 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1764 3605256 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19928 7194112 1% /mnt/lustre df:/mnt/lustre Not a Lustre filesystem Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 21:00:50 (1713488450) shut down Failover mds1 to oleg154-server mount facets: mds1 Failover mds2 to oleg154-server mount facets: mds2 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 Started lustre-MDT0001 21:01:17 (1713488477) targets are mounted 21:01:17 (1713488477) facet_failover done pdsh@oleg154-client: oleg154-client: ssh exited with exit code 95 oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid,mdc.lustre-MDT0001-mdc-*.mds_server_uuid pdsh@oleg154-client: oleg154-client: ssh exited with exit code 95 Starting client: oleg154-client.virtnet: -o user_xattr,flock oleg154-server@tcp:/lustre /mnt/lustre Can't lstat /mnt/lustre/d111d.replay-single/striped_dir: No such file or directory PASS 111d (110s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 111e: DNE: unlink striped dir, uncommit on MDT2, fail MDT1/MDT2 ========================================================== 21:02:33 (1713488553) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3000 1284688 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2520 1285168 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18164 3588856 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1764 3605256 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19928 7194112 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2992 1284696 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2512 1285176 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18164 3588856 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1764 3605256 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19928 7194112 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 21:02:40 (1713488560) shut down Failover mds1 to oleg154-server mount facets: mds1 Failover mds2 to oleg154-server mount facets: mds2 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 Started lustre-MDT0001 21:03:05 (1713488585) targets are mounted 21:03:05 (1713488585) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid,mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/d111e.replay-single/striped_dir: No such file or directory PASS 111e (37s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 111f: DNE: unlink striped dir, uncommit on MDT1, fail MDT1/MDT2 ========================================================== 21:03:12 (1713488592) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2960 1284728 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2492 1285196 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18164 3588856 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1764 3605256 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19928 7194112 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2960 1284728 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2480 1285208 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18164 3588856 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1764 3605256 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19928 7194112 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 21:03:18 (1713488598) shut down Failover mds1 to oleg154-server mount facets: mds1 Failover mds2 to oleg154-server mount facets: mds2 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0001 Started lustre-MDT0000 21:03:44 (1713488624) targets are mounted 21:03:44 (1713488624) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid,mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/d111f.replay-single/striped_dir: No such file or directory PASS 111f (40s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 111g: DNE: unlink striped dir, fail MDT1/MDT2 ========================================================== 21:03:54 (1713488634) UUID Inodes IUsed IFree IUse% Mounted on lustre-MDT0000_UUID 1024000 555 1023445 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1024000 318 1023682 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 262144 552 261592 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 262144 492 261652 1% /mnt/lustre[OST:1] filesystem_summary: 524117 873 523244 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2968 1284720 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2488 1285200 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18164 3588856 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1764 3605256 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19928 7194112 1% /mnt/lustre UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2968 1284720 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2488 1285200 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18164 3588856 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1764 3605256 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19928 7194112 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 21:04:00 (1713488640) shut down Failover mds1 to oleg154-server mount facets: mds1 Failover mds2 to oleg154-server mount facets: mds2 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0001 Started lustre-MDT0000 21:04:26 (1713488666) targets are mounted 21:04:26 (1713488666) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid,mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec Can't lstat /mnt/lustre/d111g.replay-single/striped_dir: No such file or directory PASS 111g (40s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112a: DNE: cross MDT rename, fail MDT1 ========================================================== 21:04:36 (1713488676) SKIP: replay-single test_112a needs >= 4 MDTs SKIP 112a (0s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112b: DNE: cross MDT rename, fail MDT2 ========================================================== 21:04:38 (1713488678) SKIP: replay-single test_112b needs >= 4 MDTs SKIP 112b (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112c: DNE: cross MDT rename, fail MDT3 ========================================================== 21:04:40 (1713488680) SKIP: replay-single test_112c needs >= 4 MDTs SKIP 112c (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112d: DNE: cross MDT rename, fail MDT4 ========================================================== 21:04:43 (1713488683) SKIP: replay-single test_112d needs >= 4 MDTs SKIP 112d (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112e: DNE: cross MDT rename, fail MDT1 and MDT2 ========================================================== 21:04:45 (1713488685) SKIP: replay-single test_112e needs >= 4 MDTs SKIP 112e (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112f: DNE: cross MDT rename, fail MDT1 and MDT3 ========================================================== 21:04:47 (1713488687) SKIP: replay-single test_112f needs >= 4 MDTs SKIP 112f (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112g: DNE: cross MDT rename, fail MDT1 and MDT4 ========================================================== 21:04:49 (1713488689) SKIP: replay-single test_112g needs >= 4 MDTs SKIP 112g (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112h: DNE: cross MDT rename, fail MDT2 and MDT3 ========================================================== 21:04:52 (1713488692) SKIP: replay-single test_112h needs >= 4 MDTs SKIP 112h (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112i: DNE: cross MDT rename, fail MDT2 and MDT4 ========================================================== 21:04:54 (1713488694) SKIP: replay-single test_112i needs >= 4 MDTs SKIP 112i (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112j: DNE: cross MDT rename, fail MDT3 and MDT4 ========================================================== 21:04:56 (1713488696) SKIP: replay-single test_112j needs >= 4 MDTs SKIP 112j (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112k: DNE: cross MDT rename, fail MDT1,MDT2,MDT3 ========================================================== 21:04:59 (1713488699) SKIP: replay-single test_112k needs >= 4 MDTs SKIP 112k (0s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112l: DNE: cross MDT rename, fail MDT1,MDT2,MDT4 ========================================================== 21:05:01 (1713488701) SKIP: replay-single test_112l needs >= 4 MDTs SKIP 112l (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112m: DNE: cross MDT rename, fail MDT1,MDT3,MDT4 ========================================================== 21:05:03 (1713488703) SKIP: replay-single test_112m needs >= 4 MDTs SKIP 112m (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 112n: DNE: cross MDT rename, fail MDT2,MDT3,MDT4 ========================================================== 21:05:05 (1713488705) SKIP: replay-single test_112n needs >= 4 MDTs SKIP 112n (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 115: failover for create/unlink striped directory ========================================================== 21:05:08 (1713488708) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 2964 1284724 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2476 1285212 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18164 3588856 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1764 3605256 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19928 7194112 1% /mnt/lustre striped dir -i1 -c2 -H all_char /mnt/lustre/d115.replay-single/test_0 striped dir -i1 -c2 -H crush2 /mnt/lustre/d115.replay-single/test_1 striped dir -i1 -c2 -H all_char /mnt/lustre/d115.replay-single/test_2 striped dir -i1 -c2 -H all_char /mnt/lustre/d115.replay-single/test_3 striped dir -i1 -c2 -H fnv_1a_64 /mnt/lustre/d115.replay-single/test_4 Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 21:05:11 (1713488711) shut down Failover mds2 to oleg154-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0001 21:05:24 (1713488724) targets are mounted 21:05:24 (1713488724) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3016 1284672 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2544 1285144 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18164 3588856 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1764 3605256 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19928 7194112 1% /mnt/lustre striped dir -i0 -c2 -H crush /mnt/lustre/d115.replay-single/test_0 striped dir -i0 -c2 -H all_char /mnt/lustre/d115.replay-single/test_1 striped dir -i0 -c2 -H fnv_1a_64 /mnt/lustre/d115.replay-single/test_2 striped dir -i0 -c2 -H crush2 /mnt/lustre/d115.replay-single/test_3 striped dir -i0 -c2 -H fnv_1a_64 /mnt/lustre/d115.replay-single/test_4 Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 21:05:34 (1713488734) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 21:05:47 (1713488747) targets are mounted 21:05:47 (1713488747) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 115 (46s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 116a: large update log master MDT recovery ========================================================== 21:05:56 (1713488756) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3056 1284632 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2608 1285080 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18164 3588856 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1764 3605256 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19928 7194112 1% /mnt/lustre fail_loc=0x80001702 Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 21:06:00 (1713488760) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 21:06:14 (1713488774) targets are mounted 21:06:14 (1713488774) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d116a.replay-single/striped_dir has type dir OK PASS 116a (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 116b: large update log slave MDT recovery ========================================================== 21:06:24 (1713488784) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3108 1284580 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2780 1284908 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18164 3588856 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1764 3605256 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19928 7194112 1% /mnt/lustre fail_loc=0x80001702 Failing mds2 on oleg154-server Stopping /mnt/lustre-mds2 (opts:) on oleg154-server 21:06:28 (1713488788) shut down Failover mds2 to oleg154-server mount facets: mds2 Starting mds2: -o localrecov /dev/mapper/mds2_flakey /mnt/lustre-mds2 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0001 21:06:42 (1713488802) targets are mounted 21:06:42 (1713488802) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0001-mdc-*.mds_server_uuid mdc.lustre-MDT0001-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/d116b.replay-single/striped_dir has type dir OK PASS 116b (26s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 117: DNE: cross MDT unlink, fail MDT1 and MDT2 ========================================================== 21:06:52 (1713488812) SKIP: replay-single test_117 needs >= 4 MDTs SKIP 117 (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 118: invalidate osp update will not cause update log corruption ========================================================== 21:06:56 (1713488816) fail_loc=0x1705 lfs setdirstripe: dirstripe error on '/mnt/lustre/d118.replay-single/striped_dir': Input/output error lfs setdirstripe: cannot create dir '/mnt/lustre/d118.replay-single/striped_dir': Input/output error lfs setdirstripe: dirstripe error on '/mnt/lustre/d118.replay-single/striped_dir1': Input/output error lfs setdirstripe: cannot create dir '/mnt/lustre/d118.replay-single/striped_dir1': Input/output error fail_loc=0x0 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3204 1284484 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2660 1285028 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18164 3588856 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1764 3605256 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19928 7194112 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 21:07:02 (1713488822) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 21:07:17 (1713488837) targets are mounted 21:07:17 (1713488837) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 118 (29s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 119: timeout of normal replay does not cause DNE replay fails ========================================================== 21:07:27 (1713488847) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3272 1284416 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 2720 1284968 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18164 3588856 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1764 3605256 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19928 7194112 1% /mnt/lustre Stopping /mnt/lustre-mds1 (opts:) on oleg154-server Failover mds1 to oleg154-server fail_loc=0x80000714 fail_val=65 Starting mds1: -o localrecov -o recovery_time_hard=60 /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 oleg154-client.virtnet: executing wait_import_state_mount FULL mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 67 sec mdt.lustre-MDT0000.recovery_time_hard=180 PASS 119 (86s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 120: DNE fail abort should stop both normal and DNE replay ========================================================== 21:08:56 (1713488936) Replay barrier on lustre-MDT0000 Stopping /mnt/lustre-mds1 (opts:) on oleg154-server Failover mds1 to oleg154-server Starting mds1: -o localrecov -o abort_recovery /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 find: '/mnt/lustre/d120.replay-single': No such file or directory PASS 120 (27s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 121: lock replay timed out and race ========================================================== 21:09:25 (1713488965) multiop /mnt/lustre/f121.replay-single vs_s TMPPIPE=/tmp/multiop_open_wait_pipe.7506 Stopping /mnt/lustre-mds1 (opts:) on oleg154-server Failover mds1 to oleg154-server fail_loc=0x721 fail_val=0 at_max=0 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 pdsh@oleg154-client: oleg154-client: ssh exited with exit code 5 fail_loc=0x0 at_max=600 PASS 121 (200s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 130a: DoM file create (setstripe) replay ========================================================== 21:12:47 (1713489167) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3664 1284024 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 3024 1284664 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18164 3588856 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1764 3605256 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19928 7194112 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 21:12:52 (1713489172) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 21:13:08 (1713489188) targets are mounted 21:13:08 (1713489188) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 130a (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 130b: DoM file create (inherited) replay ========================================================== 21:13:18 (1713489198) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3668 1284020 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 3024 1284664 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18164 3588856 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1764 3605256 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19928 7194112 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 21:13:23 (1713489203) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 21:13:39 (1713489219) targets are mounted 21:13:39 (1713489219) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 130b (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 131a: DoM file write lock replay === 21:13:49 (1713489229) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3668 1284020 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 3024 1284664 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18164 3588856 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1764 3605256 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19928 7194112 1% /mnt/lustre 1+0 records in 1+0 records out 8 bytes (8 B) copied, 0.00247049 s, 3.2 kB/s Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 21:13:54 (1713489234) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 21:14:10 (1713489250) targets are mounted 21:14:10 (1713489250) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec PASS 131a (28s) debug_raw_pointers=0 debug_raw_pointers=0 SKIP: replay-single test_131b skipping excluded test 131b debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 132a: PFL new component instantiate replay ========================================================== 21:14:20 (1713489260) UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3672 1284016 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 3024 1284664 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18164 3588856 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 1764 3605256 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 19928 7194112 1% /mnt/lustre 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0364409 s, 28.8 MB/s /mnt/lustre/f132a.replay-single lcm_layout_gen: 3 lcm_mirror_count: 1 lcm_entry_count: 2 lcme_id: 1 lcme_mirror_id: 0 lcme_flags: init lcme_extent.e_start: 0 lcme_extent.e_end: 1048576 lmm_stripe_count: 1 lmm_stripe_size: 1048576 lmm_pattern: raid0 lmm_layout_gen: 0 lmm_stripe_offset: 0 lmm_objects: - 0: { l_ost_idx: 0, l_fid: [0x280000401:0x3f8a:0x0] } lcme_id: 2 lcme_mirror_id: 0 lcme_flags: init lcme_extent.e_start: 1048576 lcme_extent.e_end: EOF lmm_stripe_count: 2 lmm_stripe_size: 4194304 lmm_pattern: raid0 lmm_layout_gen: 0 lmm_stripe_offset: 1 lmm_objects: - 0: { l_ost_idx: 1, l_fid: [0x2c0000401:0x3f62:0x0] } - 1: { l_ost_idx: 0, l_fid: [0x280000401:0x3f8b:0x0] } Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 21:14:25 (1713489265) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 21:14:40 (1713489280) targets are mounted 21:14:40 (1713489280) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec /mnt/lustre/f132a.replay-single lcm_layout_gen: 1 lcm_mirror_count: 1 lcm_entry_count: 2 lcme_id: 1 lcme_mirror_id: 0 lcme_flags: init lcme_extent.e_start: 0 lcme_extent.e_end: 1048576 lmm_stripe_count: 1 lmm_stripe_size: 1048576 lmm_pattern: raid0 lmm_layout_gen: 65535 lmm_stripe_offset: 0 lmm_objects: - 0: { l_ost_idx: 0, l_fid: [0x280000401:0x3f8a:0x0] } lcme_id: 2 lcme_mirror_id: 0 lcme_flags: init lcme_extent.e_start: 1048576 lcme_extent.e_end: EOF lmm_stripe_count: 2 lmm_stripe_size: 4194304 lmm_pattern: raid0 lmm_layout_gen: 65535 lmm_stripe_offset: 1 lmm_objects: - 0: { l_ost_idx: 1, l_fid: [0x2c0000401:0x3f62:0x0] } - 1: { l_ost_idx: 0, l_fid: [0x280000401:0x3f8b:0x0] } PASS 132a (28s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 133: check resend of ongoing requests for lwp during failover ========================================================== 21:14:50 (1713489290) seq.srv-lustre-MDT0001.space=clear Starting client: oleg154-client.virtnet: -o user_xattr,flock oleg154-server@tcp:/lustre /mnt/lustre fail_val=700 fail_loc=0x80000123 Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 21:14:55 (1713489295) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 21:15:09 (1713489309) targets are mounted 21:15:09 (1713489309) facet_failover done PASS 133 (24s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 134: replay creation of a file created in a pool ========================================================== 21:15:17 (1713489317) Creating new pool oleg154-server: Pool lustre.pool_134 created Adding targets to pool oleg154-server: OST lustre-OST0001_UUID added to pool lustre.pool_134 UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3764 1283924 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 3068 1284620 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18164 3588856 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 2788 3604232 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 20952 7193088 1% /mnt/lustre Failing mds1 on oleg154-server Stopping /mnt/lustre-mds1 (opts:) on oleg154-server 21:15:27 (1713489327) shut down Failover mds1 to oleg154-server mount facets: mds1 Starting mds1: -o localrecov /dev/mapper/mds1_flakey /mnt/lustre-mds1 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-MDT0000 21:15:43 (1713489343) targets are mounted 21:15:43 (1713489343) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) mdc.lustre-MDT0000-mdc-*.mds_server_uuid mdc.lustre-MDT0000-mdc-*.mds_server_uuid in FULL state after 0 sec Destroy the created pools: pool_134 lustre.pool_134 oleg154-server: OST lustre-OST0001_UUID removed from pool lustre.pool_134 oleg154-server: Pool lustre.pool_134 destroyed Waiting 90s for 'foo' PASS 134 (40s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 135: Server failure in lock replay phase ========================================================== 21:16:00 (1713489360) Failing ost1 on oleg154-server Stopping /mnt/lustre-ost1 (opts:) on oleg154-server 21:16:02 (1713489362) shut down Failover ost1 to oleg154-server mount facets: ost1 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-OST0000 21:16:17 (1713489377) targets are mounted 21:16:17 (1713489377) facet_failover done oleg154-client.virtnet: executing wait_import_state_mount (FULL|IDLE) osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in FULL state after 0 sec UUID 1K-blocks Used Available Use% Mounted on lustre-MDT0000_UUID 1414116 3744 1283944 1% /mnt/lustre[MDT:0] lustre-MDT0001_UUID 1414116 3068 1284620 1% /mnt/lustre[MDT:1] lustre-OST0000_UUID 3833116 18164 3588856 1% /mnt/lustre[OST:0] lustre-OST0001_UUID 3833116 2788 3604232 1% /mnt/lustre[OST:1] filesystem_summary: 7666232 20952 7193088 1% /mnt/lustre ldlm.cancel_unused_locks_before_replay=0 Stopping /mnt/lustre-ost1 (opts:) on oleg154-server Failover ost1 to oleg154-server oleg154-server: oleg154-server.virtnet: executing load_module ../libcfs/libcfs/libcfs fail_loc=0x32d fail_val=20 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-OST0000 oleg154-client.virtnet: executing wait_import_state_mount REPLAY_LOCKS osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid osc.lustre-OST0000-osc-[-0-9a-f]*.ost_server_uuid in REPLAY_LOCKS state after 0 sec Stopping /mnt/lustre-ost1 (opts:) on oleg154-server Failover ost1 to oleg154-server oleg154-server: oleg154-server.virtnet: executing load_module ../libcfs/libcfs/libcfs fail_loc=0 Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 End of sync oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-OST0000 Stopping /mnt/lustre-ost1 (opts:-f) on oleg154-server Stopping /mnt/lustre-ost2 (opts:-f) on oleg154-server Starting ost1: -o localrecov /dev/mapper/ost1_flakey /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-OST0000 Starting ost2: -o localrecov /dev/mapper/ost2_flakey /mnt/lustre-ost2 seq.cli-lustre-OST0001-super.width=65536 oleg154-server: oleg154-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg154-client: oleg154-server: ssh exited with exit code 1 Started lustre-OST0001 ldlm.cancel_unused_locks_before_replay=1 PASS 135 (76s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 136: MDS to disconnect all OSPs first, then cleanup ldlm ========================================================== 21:17:18 (1713489438) SKIP: replay-single test_136 needs > 2 MDTs SKIP 136 (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == replay-single test 200: Dropping one OBD_PING should not cause disconnect ========================================================== 21:17:22 (1713489442) SKIP: replay-single test_200 Need remote client SKIP 200 (1s) debug_raw_pointers=0 debug_raw_pointers=0 == replay-single test complete, duration 8769 sec ======== 21:17:24 (1713489444) === replay-single: start cleanup 21:17:24 (1713489444) === === replay-single: finish cleanup 21:17:27 (1713489447) ===