-----============= acceptance-small: sanity-scrub ============----- Tue Apr 16 15:40:29 EDT 2024 excepting tests: Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions loading modules on: 'oleg109-server' oleg109-server: oleg109-server.virtnet: executing load_modules_local oleg109-server: Loading modules from /home/green/git/lustre-release/lustre oleg109-server: detected 4 online CPUs by sysfs oleg109-server: Force libcfs to create 2 CPU partitions client=34553367 MDS=34553367 OSS=34553367 Stopping clients: oleg109-client.virtnet /mnt/lustre (opts:) Stopping client oleg109-client.virtnet /mnt/lustre opts: Stopping clients: oleg109-client.virtnet /mnt/lustre2 (opts:) Stopping /mnt/lustre-mds1 (opts:-f) on oleg109-server Stopping /mnt/lustre-ost1 (opts:-f) on oleg109-server Stopping /mnt/lustre-ost2 (opts:-f) on oleg109-server unloading modules on: 'oleg109-server' oleg109-server: oleg109-server.virtnet: executing unload_modules_local modules unloaded. === sanity-scrub: start setup 15:40:43 (1713296443) === Stopping clients: oleg109-client.virtnet /mnt/lustre (opts:-f) Stopping clients: oleg109-client.virtnet /mnt/lustre2 (opts:-f) pdsh@oleg109-client: oleg109-server: ssh exited with exit code 2 oleg109-server: oleg109-server.virtnet: executing set_hostid Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions ../libcfs/libcfs/libcfs options: 'cpu_npartitions=2' ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' quota/lquota options: 'hash_lqs_cur_bits=3' loading modules on: 'oleg109-server' oleg109-server: oleg109-server.virtnet: executing load_modules_local oleg109-server: Loading modules from /home/green/git/lustre-release/lustre oleg109-server: detected 4 online CPUs by sysfs oleg109-server: Force libcfs to create 2 CPU partitions oleg109-server: ptlrpc/ptlrpc options: 'lbug_on_grant_miscount=1' oleg109-server: quota/lquota options: 'hash_lqs_cur_bits=3' Formatting mgs, mds, osts Format mds1: lustre-mdt1/mdt1 Format ost1: lustre-ost1/ost1 Format ost2: lustre-ost2/ost2 Checking servers environments Checking clients oleg109-client.virtnet environments Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions loading modules on: 'oleg109-server' oleg109-server: oleg109-server.virtnet: executing load_modules_local oleg109-server: Loading modules from /home/green/git/lustre-release/lustre oleg109-server: detected 4 online CPUs by sysfs oleg109-server: Force libcfs to create 2 CPU partitions Setup mgs, mdt, osts Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg109-server: oleg109-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg109-client: oleg109-server: ssh exited with exit code 1 Commit the device label on lustre-mdt1/mdt1 Started lustre-MDT0000 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg109-server: oleg109-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg109-client: oleg109-server: ssh exited with exit code 1 Commit the device label on lustre-ost1/ost1 Started lustre-OST0000 Starting ost2: -o localrecov lustre-ost2/ost2 /mnt/lustre-ost2 seq.cli-lustre-OST0001-super.width=65536 oleg109-server: oleg109-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg109-client: oleg109-server: ssh exited with exit code 1 Commit the device label on lustre-ost2/ost2 Started lustre-OST0001 Starting client: oleg109-client.virtnet: -o user_xattr,flock oleg109-server@tcp:/lustre /mnt/lustre Starting client oleg109-client.virtnet: -o user_xattr,flock oleg109-server@tcp:/lustre /mnt/lustre Started clients oleg109-client.virtnet: 192.168.201.109@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Using TIMEOUT=20 osc.lustre-OST0000-osc-ffff8800b64a9800.idle_timeout=debug osc.lustre-OST0001-osc-ffff8800b64a9800.idle_timeout=debug setting jobstats to procname_uid Setting lustre.sys.jobid_var from disable to procname_uid Waiting 90s for 'procname_uid' Updated after 3s: want 'procname_uid' got 'procname_uid' disable quota as required === sanity-scrub: finish setup 15:41:29 (1713296489) === debug_raw_pointers=Y debug_raw_pointers=Y == sanity-scrub test 0: Do not auto trigger OI scrub for non-backup/restore case ========================================================== 15:41:30 (1713296490) preparing... Tue Apr 16 15:41:31 EDT 2024 creating 0 files on mds1 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.00866735 s, 121 MB/s prepared Tue Apr 16 15:41:31 EDT 2024. stop mds1 starting MDTs without disabling OI scrub oleg109-server: oleg109-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg109-client: oleg109-server: ssh exited with exit code 1 Starting client: oleg109-client.virtnet: -o user_xattr,flock oleg109-server@tcp:/lustre /mnt/lustre PASS 0 (11s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == sanity-scrub test 1a: Auto trigger initial OI scrub when server mounts ========================================================== 15:41:42 (1713296502) preparing... Tue Apr 16 15:41:43 EDT 2024 creating 0 files on mds1 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.00878607 s, 119 MB/s prepared Tue Apr 16 15:41:43 EDT 2024. stop mds1 start mds1 without disabling OI scrub: -o localrecov -o user_xattr oleg109-server: oleg109-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg109-client: oleg109-server: ssh exited with exit code 1 Starting client: oleg109-client.virtnet: -o user_xattr,flock oleg109-server@tcp:/lustre /mnt/lustre fail_loc=0x193 fail_loc=0 192.168.201.109@tcp:/lustre /mnt/lustre lustre rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project 0 0 Stopping client oleg109-client.virtnet /mnt/lustre (opts:) stop mds1 start mds1 with disabling OI scrub: -o localrecov -o user_xattr,noscrub oleg109-server: oleg109-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg109-client: oleg109-server: ssh exited with exit code 1 PASS 1a (15s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == sanity-scrub test 1b: Trigger OI scrub when MDT mounts for OI files remove/recreate case ========================================================== 15:41:58 (1713296518) Checking servers environments Checking clients oleg109-client.virtnet environments Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions loading modules on: 'oleg109-server' oleg109-server: oleg109-server.virtnet: executing load_modules_local oleg109-server: Loading modules from /home/green/git/lustre-release/lustre oleg109-server: detected 4 online CPUs by sysfs oleg109-server: Force libcfs to create 2 CPU partitions oleg109-server: libkmod: kmod_module_get_holders: could not open '/sys/module/acpi_cpufreq/holders': No such file or directory Setup mgs, mdt, osts Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg109-server: mount.lustre: according to /etc/mtab lustre-mdt1/mdt1 is already mounted on /mnt/lustre-mds1 pdsh@oleg109-client: oleg109-server: ssh exited with exit code 17 Start of lustre-mdt1/mdt1 on mds1 failed 17 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg109-server: mount.lustre: according to /etc/mtab lustre-ost1/ost1 is already mounted on /mnt/lustre-ost1 pdsh@oleg109-client: oleg109-server: ssh exited with exit code 17 seq.cli-lustre-OST0000-super.width=65536 Start of lustre-ost1/ost1 on ost1 failed 17 Starting ost2: -o localrecov lustre-ost2/ost2 /mnt/lustre-ost2 oleg109-server: mount.lustre: according to /etc/mtab lustre-ost2/ost2 is already mounted on /mnt/lustre-ost2 pdsh@oleg109-client: oleg109-server: ssh exited with exit code 17 seq.cli-lustre-OST0001-super.width=65536 Start of lustre-ost2/ost2 on ost2 failed 17 Starting client: oleg109-client.virtnet: -o user_xattr,flock oleg109-server@tcp:/lustre /mnt/lustre Starting client oleg109-client.virtnet: -o user_xattr,flock oleg109-server@tcp:/lustre /mnt/lustre Started clients oleg109-client.virtnet: 192.168.201.109@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Using TIMEOUT=20 osc.lustre-OST0000-osc-ffff8800b649e000.idle_timeout=debug osc.lustre-OST0001-osc-ffff8800b649e000.idle_timeout=debug disable quota as required preparing... Tue Apr 16 15:42:14 EDT 2024 creating 0 files on mds1 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.011525 s, 91.0 MB/s prepared Tue Apr 16 15:42:15 EDT 2024. fail_loc=0x198 fail_loc=0 stop mds1 start MDTs without disabling OI scrub oleg109-server: oleg109-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg109-client: oleg109-server: ssh exited with exit code 1 Starting client: oleg109-client.virtnet: -o user_xattr,flock oleg109-server@tcp:/lustre /mnt/lustre PASS 1b (27s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == sanity-scrub test 1c: Auto detect kinds of OI file(s) removed/recreated cases ========================================================== 15:42:26 (1713296546) SKIP: sanity-scrub test_1c ldiskfs special test SKIP 1c (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == sanity-scrub test 2: Trigger OI scrub when MDT mounts for backup/restore case ========================================================== 15:42:29 (1713296549) SKIP: sanity-scrub test_2 ldiskfs special test SKIP 2 (0s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == sanity-scrub test 4a: Auto trigger OI scrub if bad OI mapping was found (1) ========================================================== 15:42:31 (1713296551) preparing... Tue Apr 16 15:42:32 EDT 2024 creating 0 files on mds1 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.00957948 s, 109 MB/s prepared Tue Apr 16 15:42:32 EDT 2024. fail_loc=0x193 fail_loc=0 stop mds1 starting MDTs with OI scrub disabled oleg109-server: oleg109-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg109-client: oleg109-server: ssh exited with exit code 1 Starting client: oleg109-client.virtnet: -o user_xattr,flock oleg109-server@tcp:/lustre /mnt/lustre PASS 4a (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == sanity-scrub test 4b: Auto trigger OI scrub if bad OI mapping was found (2) ========================================================== 15:42:51 (1713296571) SKIP: sanity-scrub test_4b ldiskfs special test SKIP 4b (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == sanity-scrub test 4c: Auto trigger OI scrub if bad OI mapping was found (3) ========================================================== 15:42:54 (1713296574) SKIP: sanity-scrub test_4c ldiskfs special test SKIP 4c (0s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == sanity-scrub test 4d: FID in LMA mismatch with object FID won't block create ========================================================== 15:42:56 (1713296576) SKIP: sanity-scrub test_4d ldiskfs only test SKIP 4d (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == sanity-scrub test 4e: FID reuse can be fixed ========== 15:42:59 (1713296579) SKIP: sanity-scrub test_4e ldiskfs only test SKIP 4e (0s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == sanity-scrub test 5: OI scrub state machine =========== 15:43:01 (1713296581) oleg109-server: oleg109-server.virtnet: executing set_hostid oleg109-server: oleg109-server.virtnet: executing load_modules_local oleg109-server: oleg109-server.virtnet: executing load_modules_local oleg109-server: oleg109-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg109-client: oleg109-server: ssh exited with exit code 1 oleg109-server: oleg109-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg109-client: oleg109-server: ssh exited with exit code 1 oleg109-server: oleg109-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg109-client: oleg109-server: ssh exited with exit code 1 Using TIMEOUT=20 preparing... Tue Apr 16 15:44:05 EDT 2024 creating 100 files on mds1 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.00893896 s, 117 MB/s prepared Tue Apr 16 15:44:06 EDT 2024. fail_loc=0x193 fail_loc=0 stop mds1 starting MDTs with OI scrub disabled (1) oleg109-server: oleg109-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg109-client: oleg109-server: ssh exited with exit code 1 Starting client: oleg109-client.virtnet: -o user_xattr,flock oleg109-server@tcp:/lustre /mnt/lustre fail_val=3 fail_loc=0x190 192.168.201.109@tcp:/lustre /mnt/lustre lustre rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project 0 0 Stopping client oleg109-client.virtnet /mnt/lustre (opts:) fail_loc=0x191 stopping mds1 fail_loc=0 fail_val=0 starting MDTs with OI scrub disabled (2) oleg109-server: oleg109-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg109-client: oleg109-server: ssh exited with exit code 1 stopping mds1 fail_val=3 fail_loc=0x190 starting MDTs without disabling OI scrub oleg109-server: oleg109-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg109-client: oleg109-server: ssh exited with exit code 1 fail_loc=0x192 Waiting 6s for 'failed' Starting client: oleg109-client.virtnet: -o user_xattr,flock oleg109-server@tcp:/lustre /mnt/lustre fail_loc=0 fail_val=0 File: '/mnt/lustre/d5.sanity-scrub/mds1/sanity-scrub.sh' Size: 40777 Blocks: 131 IO Block: 4194304 regular file Device: 2c54f966h/743766374d Inode: 144115205272502341 Links: 1 Access: (0755/-rwxr-xr-x) Uid: ( 0/ root) Gid: ( 0/ root) Access: 2024-04-16 15:44:06.000000000 -0400 Modify: 2024-04-16 15:44:06.000000000 -0400 Change: 2024-04-16 15:44:07.000000000 -0400 Birth: - PASS 5 (92s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == sanity-scrub test 6: OI scrub resumes from last checkpoint ========================================================== 15:44:35 (1713296675) preparing... Tue Apr 16 15:44:36 EDT 2024 creating 100 files on mds1 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.00876964 s, 120 MB/s prepared Tue Apr 16 15:44:37 EDT 2024. fail_loc=0x193 fail_loc=0 stop mds1 starting MDTs with OI scrub disabled oleg109-server: oleg109-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg109-client: oleg109-server: ssh exited with exit code 1 Starting client: oleg109-client.virtnet: -o user_xattr,flock oleg109-server@tcp:/lustre /mnt/lustre fail_val=2 fail_loc=0x190 fail_loc=0x192 Waiting 6s for 'failed' Updated after 2s: want 'failed' got 'failed' fail_val=3 fail_loc=0x190 File: '/mnt/lustre/d6.sanity-scrub/mds1/sanity-scrub.sh' Size: 40777 Blocks: 131 IO Block: 4194304 regular file Device: 2c54f966h/743766374d Inode: 144115305935798341 Links: 1 Access: (0755/-rwxr-xr-x) Uid: ( 0/ root) Gid: ( 0/ root) Access: 2024-04-16 15:44:37.000000000 -0400 Modify: 2024-04-16 15:44:37.000000000 -0400 Change: 2024-04-16 15:44:37.000000000 -0400 Birth: - 192.168.201.109@tcp:/lustre /mnt/lustre lustre rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project 0 0 Stopping client oleg109-client.virtnet /mnt/lustre (opts:) fail_loc=0x191 stopping mds1 fail_val=3 fail_loc=0x190 starting MDTs without disabling OI scrub oleg109-server: oleg109-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg109-client: oleg109-server: ssh exited with exit code 1 fail_loc=0 fail_val=0 Waiting 6s for 'completed' PASS 6 (31s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == sanity-scrub test 7: System is available during OI scrub scanning ========================================================== 15:45:08 (1713296708) Checking servers environments Checking clients oleg109-client.virtnet environments Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions loading modules on: 'oleg109-server' oleg109-server: oleg109-server.virtnet: executing load_modules_local oleg109-server: Loading modules from /home/green/git/lustre-release/lustre oleg109-server: detected 4 online CPUs by sysfs oleg109-server: Force libcfs to create 2 CPU partitions Setup mgs, mdt, osts Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg109-server: mount.lustre: according to /etc/mtab lustre-mdt1/mdt1 is already mounted on /mnt/lustre-mds1 pdsh@oleg109-client: oleg109-server: ssh exited with exit code 17 Start of lustre-mdt1/mdt1 on mds1 failed 17 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg109-server: mount.lustre: according to /etc/mtab lustre-ost1/ost1 is already mounted on /mnt/lustre-ost1 pdsh@oleg109-client: oleg109-server: ssh exited with exit code 17 seq.cli-lustre-OST0000-super.width=65536 Start of lustre-ost1/ost1 on ost1 failed 17 Starting ost2: -o localrecov lustre-ost2/ost2 /mnt/lustre-ost2 oleg109-server: mount.lustre: according to /etc/mtab lustre-ost2/ost2 is already mounted on /mnt/lustre-ost2 pdsh@oleg109-client: oleg109-server: ssh exited with exit code 17 seq.cli-lustre-OST0001-super.width=65536 Start of lustre-ost2/ost2 on ost2 failed 17 Starting client: oleg109-client.virtnet: -o user_xattr,flock oleg109-server@tcp:/lustre /mnt/lustre Starting client oleg109-client.virtnet: -o user_xattr,flock oleg109-server@tcp:/lustre /mnt/lustre Started clients oleg109-client.virtnet: 192.168.201.109@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Using TIMEOUT=20 osc.lustre-OST0000-osc-ffff88012e2a8800.idle_timeout=debug osc.lustre-OST0001-osc-ffff88012e2a8800.idle_timeout=debug disable quota as required preparing... Tue Apr 16 15:45:24 EDT 2024 creating 500 files on mds1 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.00880451 s, 119 MB/s prepared Tue Apr 16 15:45:26 EDT 2024. fail_loc=0x193 fail_loc=0 stop mds1 starting MDTs with OI scrub disabled oleg109-server: oleg109-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg109-client: oleg109-server: ssh exited with exit code 1 Starting client: oleg109-client.virtnet: -o user_xattr,flock oleg109-server@tcp:/lustre /mnt/lustre fail_val=3 fail_loc=0x190 File: '/mnt/lustre/d7.sanity-scrub/mds1/f7.sanity-scrub300' Size: 0 Blocks: 0 IO Block: 4194304 regular empty file Device: 2c54f966h/743766374d Inode: 144115373044662655 Links: 1 Access: (0444/-r--r--r--) Uid: ( 0/ root) Gid: ( 0/ root) Access: 2024-04-16 15:45:25.000000000 -0400 Modify: 2024-04-16 15:45:25.000000000 -0400 Change: 2024-04-16 15:45:25.000000000 -0400 Birth: - fail_loc=0 fail_val=0 Waiting 6s for 'completed' Updated after 2s: want 'completed' got 'completed' PASS 7 (32s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == sanity-scrub test 8: Control OI scrub manually ======== 15:45:41 (1713296741) preparing... Tue Apr 16 15:45:43 EDT 2024 creating 128 files on mds1 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0104198 s, 101 MB/s prepared Tue Apr 16 15:45:45 EDT 2024. fail_loc=0x193 fail_loc=0 stop mds1 starting MDTs with OI scrub disabled oleg109-server: oleg109-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg109-client: oleg109-server: ssh exited with exit code 1 fail_val=1 fail_loc=0x190 Started LFSCK on the device lustre-MDT0000: scrub Stopped LFSCK on the device lustre-MDT0000. Started LFSCK on the device lustre-MDT0000: scrub fail_loc=0 fail_val=0 Waiting 6s for 'completed' Updated after 2s: want 'completed' got 'completed' PASS 8 (19s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == sanity-scrub test 9: OI scrub speed control =========== 15:46:02 (1713296762) SKIP: sanity-scrub test_9 test scrub speed only on ldiskfs SKIP 9 (0s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == sanity-scrub test 10a: non-stopped OI scrub should auto restarts after MDS remount (1) ========================================================== 15:46:04 (1713296764) Checking servers environments Checking clients oleg109-client.virtnet environments Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions loading modules on: 'oleg109-server' oleg109-server: oleg109-server.virtnet: executing load_modules_local oleg109-server: Loading modules from /home/green/git/lustre-release/lustre oleg109-server: detected 4 online CPUs by sysfs oleg109-server: Force libcfs to create 2 CPU partitions Setup mgs, mdt, osts Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg109-server: mount.lustre: according to /etc/mtab lustre-mdt1/mdt1 is already mounted on /mnt/lustre-mds1 pdsh@oleg109-client: oleg109-server: ssh exited with exit code 17 Start of lustre-mdt1/mdt1 on mds1 failed 17 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg109-server: mount.lustre: according to /etc/mtab lustre-ost1/ost1 is already mounted on /mnt/lustre-ost1 pdsh@oleg109-client: oleg109-server: ssh exited with exit code 17 seq.cli-lustre-OST0000-super.width=65536 Start of lustre-ost1/ost1 on ost1 failed 17 Starting ost2: -o localrecov lustre-ost2/ost2 /mnt/lustre-ost2 oleg109-server: mount.lustre: according to /etc/mtab lustre-ost2/ost2 is already mounted on /mnt/lustre-ost2 pdsh@oleg109-client: oleg109-server: ssh exited with exit code 17 seq.cli-lustre-OST0001-super.width=65536 Start of lustre-ost2/ost2 on ost2 failed 17 Starting client: oleg109-client.virtnet: -o user_xattr,flock oleg109-server@tcp:/lustre /mnt/lustre Starting client oleg109-client.virtnet: -o user_xattr,flock oleg109-server@tcp:/lustre /mnt/lustre Started clients oleg109-client.virtnet: 192.168.201.109@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Using TIMEOUT=20 osc.lustre-OST0000-osc-ffff8800b6ce3800.idle_timeout=debug osc.lustre-OST0001-osc-ffff8800b6ce3800.idle_timeout=debug disable quota as required preparing... Tue Apr 16 15:46:20 EDT 2024 creating 0 files on mds1 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.00917936 s, 114 MB/s prepared Tue Apr 16 15:46:20 EDT 2024. fail_loc=0x193 fail_loc=0 stop mds1 starting mds with OI scrub disabled (1) oleg109-server: oleg109-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg109-client: oleg109-server: ssh exited with exit code 1 Starting client: oleg109-client.virtnet: -o user_xattr,flock oleg109-server@tcp:/lustre /mnt/lustre fail_val=1 fail_loc=0x190 192.168.201.109@tcp:/lustre /mnt/lustre lustre rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project 0 0 Stopping client oleg109-client.virtnet /mnt/lustre (opts:) stopping mds1 starting MDTs with OI scrub disabled (2) oleg109-server: oleg109-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg109-client: oleg109-server: ssh exited with exit code 1 stopping mds1 starting MDTs without disabling OI scrub oleg109-server: oleg109-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg109-client: oleg109-server: ssh exited with exit code 1 fail_loc=0 fail_val=0 PASS 10a (36s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == sanity-scrub test 11: OI scrub skips the new created objects only once ========================================================== 15:46:42 (1713296802) SKIP: sanity-scrub test_11 ldiskfs special test SKIP 11 (0s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == sanity-scrub test 12: OI scrub can rebuild invalid /O entries ========================================================== 15:46:44 (1713296804) Checking servers environments Checking clients oleg109-client.virtnet environments Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions loading modules on: 'oleg109-server' oleg109-server: oleg109-server.virtnet: executing load_modules_local oleg109-server: Loading modules from /home/green/git/lustre-release/lustre oleg109-server: detected 4 online CPUs by sysfs oleg109-server: Force libcfs to create 2 CPU partitions oleg109-server: libkmod: kmod_module_get_holders: could not open '/sys/module/acpi_cpufreq/holders': No such file or directory Setup mgs, mdt, osts Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg109-server: mount.lustre: according to /etc/mtab lustre-mdt1/mdt1 is already mounted on /mnt/lustre-mds1 pdsh@oleg109-client: oleg109-server: ssh exited with exit code 17 Start of lustre-mdt1/mdt1 on mds1 failed 17 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg109-server: mount.lustre: according to /etc/mtab lustre-ost1/ost1 is already mounted on /mnt/lustre-ost1 pdsh@oleg109-client: oleg109-server: ssh exited with exit code 17 seq.cli-lustre-OST0000-super.width=65536 Start of lustre-ost1/ost1 on ost1 failed 17 Starting ost2: -o localrecov lustre-ost2/ost2 /mnt/lustre-ost2 oleg109-server: mount.lustre: according to /etc/mtab lustre-ost2/ost2 is already mounted on /mnt/lustre-ost2 pdsh@oleg109-client: oleg109-server: ssh exited with exit code 17 seq.cli-lustre-OST0001-super.width=65536 Start of lustre-ost2/ost2 on ost2 failed 17 Starting client: oleg109-client.virtnet: -o user_xattr,flock oleg109-server@tcp:/lustre /mnt/lustre Starting client oleg109-client.virtnet: -o user_xattr,flock oleg109-server@tcp:/lustre /mnt/lustre Started clients oleg109-client.virtnet: 192.168.201.109@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Using TIMEOUT=20 osc.lustre-OST0000-osc-ffff8800aa63f000.idle_timeout=debug osc.lustre-OST0001-osc-ffff8800aa63f000.idle_timeout=debug disable quota as required fail_loc=0x195 total: 64 open/close in 0.15 seconds: 415.03 ops/second 192.168.201.109@tcp:/lustre /mnt/lustre lustre rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project 0 0 Stopping client oleg109-client.virtnet /mnt/lustre (opts:) Stopping /mnt/lustre-ost1 (opts:) on oleg109-server fail_loc=0x233 Starting ost1: -o localrecov -o user_xattr,noscrub lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg109-server: oleg109-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg109-client: oleg109-server: ssh exited with exit code 1 Started lustre-OST0000 Starting client: oleg109-client.virtnet: -o user_xattr,flock oleg109-server@tcp:/lustre /mnt/lustre Started LFSCK on the device lustre-OST0000: scrub fail_loc=0 PASS 12 (203s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == sanity-scrub test 13: OI scrub can rebuild missed /O entries ========================================================== 15:50:09 (1713297009) fail_loc=0x196 total: 64 open/close in 0.14 seconds: 453.14 ops/second fail_loc=0 192.168.201.109@tcp:/lustre /mnt/lustre lustre rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project 0 0 Stopping client oleg109-client.virtnet /mnt/lustre (opts:) Stopping /mnt/lustre-ost1 (opts:) on oleg109-server Starting ost1: -o localrecov -o user_xattr,noscrub lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg109-server: oleg109-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg109-client: oleg109-server: ssh exited with exit code 1 Started lustre-OST0000 Starting client: oleg109-client.virtnet: -o user_xattr,flock oleg109-server@tcp:/lustre /mnt/lustre Started LFSCK on the device lustre-OST0000: scrub PASS 13 (188s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == sanity-scrub test 14: OI scrub can repair OST objects under lost+found ========================================================== 15:53:19 (1713297199) SKIP: sanity-scrub test_14 ldiskfs special test SKIP 14 (0s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == sanity-scrub test 15: Dryrun mode OI scrub ============ 15:53:21 (1713297201) oleg109-server: oleg109-server.virtnet: executing set_hostid oleg109-server: oleg109-server.virtnet: executing load_modules_local oleg109-server: oleg109-server.virtnet: executing load_modules_local oleg109-server: libkmod: kmod_module_get_holders: could not open '/sys/module/pcc_cpufreq/holders': No such file or directory oleg109-server: oleg109-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg109-client: oleg109-server: ssh exited with exit code 1 oleg109-server: oleg109-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg109-client: oleg109-server: ssh exited with exit code 1 oleg109-server: oleg109-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg109-client: oleg109-server: ssh exited with exit code 1 Using TIMEOUT=20 preparing... Tue Apr 16 15:54:11 EDT 2024 creating 20 files on mds1 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.00781264 s, 134 MB/s prepared Tue Apr 16 15:54:12 EDT 2024. fail_loc=0x193 fail_loc=0 stop mds1 starting MDTs with OI scrub disabled oleg109-server: oleg109-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg109-client: oleg109-server: ssh exited with exit code 1 Started LFSCK on the device lustre-MDT0000: scrub Started LFSCK on the device lustre-MDT0000: scrub Started LFSCK on the device lustre-MDT0000: scrub Started LFSCK on the device lustre-MDT0000: scrub PASS 15 (66s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == sanity-scrub test 16: Initial OI scrub can rebuild crashed index objects ========================================================== 15:54:28 (1713297268) Checking servers environments Checking clients oleg109-client.virtnet environments Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions libkmod: kmod_module_get_holders: could not open '/sys/module/acpi_cpufreq/holders': No such file or directory loading modules on: 'oleg109-server' oleg109-server: oleg109-server.virtnet: executing load_modules_local oleg109-server: Loading modules from /home/green/git/lustre-release/lustre oleg109-server: detected 4 online CPUs by sysfs oleg109-server: Force libcfs to create 2 CPU partitions Setup mgs, mdt, osts Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg109-server: mount.lustre: according to /etc/mtab lustre-mdt1/mdt1 is already mounted on /mnt/lustre-mds1 pdsh@oleg109-client: oleg109-server: ssh exited with exit code 17 Start of lustre-mdt1/mdt1 on mds1 failed 17 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 oleg109-server: mount.lustre: according to /etc/mtab lustre-ost1/ost1 is already mounted on /mnt/lustre-ost1 pdsh@oleg109-client: oleg109-server: ssh exited with exit code 17 seq.cli-lustre-OST0000-super.width=65536 Start of lustre-ost1/ost1 on ost1 failed 17 Starting ost2: -o localrecov lustre-ost2/ost2 /mnt/lustre-ost2 oleg109-server: mount.lustre: according to /etc/mtab lustre-ost2/ost2 is already mounted on /mnt/lustre-ost2 pdsh@oleg109-client: oleg109-server: ssh exited with exit code 17 seq.cli-lustre-OST0001-super.width=65536 Start of lustre-ost2/ost2 on ost2 failed 17 Starting client: oleg109-client.virtnet: -o user_xattr,flock oleg109-server@tcp:/lustre /mnt/lustre Starting client oleg109-client.virtnet: -o user_xattr,flock oleg109-server@tcp:/lustre /mnt/lustre Started clients oleg109-client.virtnet: 192.168.201.109@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Using TIMEOUT=20 osc.lustre-OST0000-osc-ffff88012eff5000.idle_timeout=debug osc.lustre-OST0001-osc-ffff88012eff5000.idle_timeout=debug disable quota as required fail_loc=0x199 preparing... Tue Apr 16 15:54:45 EDT 2024 creating 0 files on mds1 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.00958218 s, 109 MB/s prepared Tue Apr 16 15:54:46 EDT 2024. stop mds1 fail_loc=0 starting MDTs without disabling OI scrub oleg109-server: oleg109-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg109-client: oleg109-server: ssh exited with exit code 1 Starting client: oleg109-client.virtnet: -o user_xattr,flock oleg109-server@tcp:/lustre /mnt/lustre PASS 16 (29s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == sanity-scrub test 17a: ENOSPC on OI insert shouldn't leak inodes ========================================================== 15:54:58 (1713297298) SKIP: sanity-scrub test_17a ldiskfs only test SKIP 17a (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == sanity-scrub test 17b: ENOSPC on .. insertion shouldn't leak inodes ========================================================== 15:55:01 (1713297301) SKIP: sanity-scrub test_17b ldiskfs only test SKIP 17b (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == sanity-scrub test 18: test mount -o resetoi to recreate OI files ========================================================== 15:55:04 (1713297304) preparing... Tue Apr 16 15:55:05 EDT 2024 creating 10 files on mds1 1+0 records in 1+0 records out 1048576 bytes (1.0 MB) copied, 0.0145011 s, 72.3 MB/s prepared Tue Apr 16 15:55:06 EDT 2024. stop mds1 oleg109-server: oleg109-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg109-client: oleg109-server: ssh exited with exit code 1 Starting client: oleg109-client.virtnet: -o user_xattr,flock oleg109-server@tcp:/lustre /mnt/lustre oleg109-server: oleg109-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg109-client: oleg109-server: ssh exited with exit code 1 [ 968.139063] Lustre: lustre-MDT0000: reset Object Index mappings Starting client: oleg109-client.virtnet: -o user_xattr,flock oleg109-server@tcp:/lustre /mnt/lustre PASS 18 (17s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == sanity-scrub test 19: LFSCK can fix multiple linked files on OST ========================================================== 15:55:23 (1713297323) total: 64 open/close in 0.16 seconds: 395.45 ops/second stopall link /mnt/lustre-ost1/O/240000400/d2/227 to /mnt/lustre-ost1/O/240000400/d2/226 oleg109-server: oleg109-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg109-client: oleg109-server: ssh exited with exit code 1 Started LFSCK on the device lustre-OST0000: scrub Stopping /mnt/lustre-ost1 (opts:) on oleg109-server File: '/mnt/lustre-ost1/O/240000400/d2/226' Size: 0 Blocks: 1 IO Block: 4096 regular empty file Device: 29h/41d Inode: 714 Links: 1 Access: (7666/-rwSrwSrwT) Uid: ( 0/ root) Gid: ( 0/ root) Access: 1969-12-31 19:00:00.000000000 -0500 Modify: 1969-12-31 19:00:00.000000000 -0500 Change: 2024-04-16 15:55:31.269000000 -0400 Birth: - oleg109-server: oleg109-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg109-client: oleg109-server: ssh exited with exit code 1 PASS 19 (150s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == sanity-scrub test 20: Don't trigger OI scrub for irreparable oi repeatedly ========================================================== 15:57:55 (1713297475) SKIP: sanity-scrub test_20 ldiskfs only test SKIP 20 (1s) debug_raw_pointers=0 debug_raw_pointers=0 debug_raw_pointers=Y debug_raw_pointers=Y == sanity-scrub test 21: don't hang MDS recovery when failed to get update log ========================================================== 15:57:58 (1713297478) SKIP: sanity-scrub test_21 needs >= 2 MDTs SKIP 21 (1s) debug_raw_pointers=0 debug_raw_pointers=0 === sanity-scrub: start setup 15:58:00 (1713297480) === Stopping clients: oleg109-client.virtnet /mnt/lustre (opts:-f) Stopping clients: oleg109-client.virtnet /mnt/lustre2 (opts:-f) Stopping /mnt/lustre-ost1 (opts:-f) on oleg109-server oleg109-server: oleg109-server.virtnet: executing set_hostid Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions loading modules on: 'oleg109-server' oleg109-server: oleg109-server.virtnet: executing load_modules_local oleg109-server: Loading modules from /home/green/git/lustre-release/lustre oleg109-server: detected 4 online CPUs by sysfs oleg109-server: Force libcfs to create 2 CPU partitions oleg109-server: libkmod: kmod_module_get_holders: could not open '/sys/module/pcc_cpufreq/holders': No such file or directory Formatting mgs, mds, osts Format mds1: lustre-mdt1/mdt1 Format ost1: lustre-ost1/ost1 Format ost2: lustre-ost2/ost2 Checking servers environments Checking clients oleg109-client.virtnet environments Loading modules from /home/green/git/lustre-release/lustre detected 4 online CPUs by sysfs Force libcfs to create 2 CPU partitions libkmod: kmod_module_get_holders: could not open '/sys/module/acpi_cpufreq/holders': No such file or directory loading modules on: 'oleg109-server' oleg109-server: oleg109-server.virtnet: executing load_modules_local oleg109-server: Loading modules from /home/green/git/lustre-release/lustre oleg109-server: detected 4 online CPUs by sysfs oleg109-server: Force libcfs to create 2 CPU partitions oleg109-server: libkmod: kmod_module_get_holders: could not open '/sys/module/pcc_cpufreq/holders': No such file or directory Setup mgs, mdt, osts Starting mds1: -o localrecov lustre-mdt1/mdt1 /mnt/lustre-mds1 oleg109-server: oleg109-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg109-client: oleg109-server: ssh exited with exit code 1 Commit the device label on lustre-mdt1/mdt1 Started lustre-MDT0000 Starting ost1: -o localrecov lustre-ost1/ost1 /mnt/lustre-ost1 seq.cli-lustre-OST0000-super.width=65536 oleg109-server: oleg109-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg109-client: oleg109-server: ssh exited with exit code 1 Commit the device label on lustre-ost1/ost1 Started lustre-OST0000 Starting ost2: -o localrecov lustre-ost2/ost2 /mnt/lustre-ost2 seq.cli-lustre-OST0001-super.width=65536 oleg109-server: oleg109-server.virtnet: executing set_default_debug vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck all pdsh@oleg109-client: oleg109-server: ssh exited with exit code 1 Commit the device label on lustre-ost2/ost2 Started lustre-OST0001 Starting client: oleg109-client.virtnet: -o user_xattr,flock oleg109-server@tcp:/lustre /mnt/lustre Starting client oleg109-client.virtnet: -o user_xattr,flock oleg109-server@tcp:/lustre /mnt/lustre Started clients oleg109-client.virtnet: 192.168.201.109@tcp:/lustre on /mnt/lustre type lustre (rw,checksum,flock,user_xattr,lruresize,lazystatfs,nouser_fid2path,verbose,noencrypt,statfs_project) Using TIMEOUT=20 osc.lustre-OST0000-osc-ffff8800b6ce4000.idle_timeout=debug osc.lustre-OST0001-osc-ffff8800b6ce4000.idle_timeout=debug disable quota as required === sanity-scrub: finish setup 15:58:41 (1713297521) === == sanity-scrub test complete, duration 1091 sec ========= 15:58:41 (1713297521) === sanity-scrub: start cleanup 15:58:42 (1713297522) === === sanity-scrub: finish cleanup 15:58:42 (1713297522) ===