[lustre-discuss] Lustre-2.10.6 frequently hangs during OST data migration
Stephane Thiell
sthiell at stanford.edu
Sat Mar 30 10:18:27 PDT 2019
Hi Tung-Han,
Your stack trace looks similar to the one we’ve just seen yesterday on our 2.10.6 system.
I’ve open https://jira.whamcloud.com/browse/LU-12136 to track the issue.
Best,
Stephane
> On Mar 29, 2019, at 8:47 PM, Tung-Han Hsieh <thhsieh at twcp1.phys.ntu.edu.tw> wrote:
>
> Dear All,
>
> Our system was recently upgraded to lustre-2.10.6. We are doing the
> data migration from some almost full OSTs to a newly installed file
> server. But we often encountered file system freezed for about 30 secs,
> and then returned to normal (within 5 mins it may happen several times).
>
> Our procedure is following.
>
> 1. In the MDS, we prevented data writing to the OSTs which are almost full:
> echo 0 > /proc/fs/lustre/osc/chome-OST0000-osc-MDT0000/max_create_count
> echo 0 > /proc/fs/lustre/osc/chome-OST0001-osc-MDT0000/max_create_count
> echo 0 > /proc/fs/lustre/osc/chome-OST0002-osc-MDT0000/max_create_count
> ....
>
> 2. In our system we have 40 OSTs, in which 36 OSTs are almost full so they
> are all marked by the above command. Our total OST size is 286TB. We are
> moving out part of their data to the remaining 4 new OSTs by the following
> standard way:
>
> cp -a /path/to/data /path/to/data.tmp
> mv /path/to/data.tmp /path/to/data
>
> In the beginning, everything looks smoothly. But after one week of running,
> the progress becoming slower and slower. Then we found that the file system
> often got freezed for a while when the data migration is running. However,
> there is almost no any loading in the whole system.
>
> It is also strange that during the past week, we did not see any logs in
> 'dmesg' messages of MDT, OSTs, and the client. Until last night, only MDT
> prompted up these 'dmesg' messages:
>
> ==============================================================================
> [410649.811086] LNet: Service thread pid 3516 was inactive for 200.27s. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes:
> [410649.811175] Pid: 3516, comm: mdt00_003
> [410649.811201]
> [410649.811202] Call Trace:
> [410649.811250] [<ffffffff81033625>] ? check_preempt_curr+0x75/0xa0
> [410649.811278] [<ffffffff8103366b>] ? ttwu_do_wakeup+0x1b/0xa0
> [410649.811306] [<ffffffff810376c1>] ? ttwu_do_activate.constprop.160+0x61/0x70
> [410649.811336] [<ffffffff8103c40a>] ? try_to_wake_up+0x1da/0x280
> [410649.811367] [<ffffffff814106ba>] schedule+0x3a/0x50
> [410649.811393] [<ffffffff81410a95>] schedule_timeout+0x145/0x210
> [410649.811421] [<ffffffff8104d090>] ? process_timeout+0x0/0x10
> [410649.811451] [<ffffffffa0b6dc88>] osp_precreate_reserve+0x328/0x8b0 [osp]
> [410649.811484] [<ffffffffa014f026>] ? do_get_write_access+0x396/0x4d0 [jbd2]
> [410649.811515] [<ffffffff81112a10>] ? __getblk+0x20/0x2e0
> [410649.811542] [<ffffffff8103c4b0>] ? default_wake_function+0x0/0x10
> [410649.811571] [<ffffffffa0b64759>] osp_declare_create+0x1a9/0x680 [osp]
> [410649.811603] [<ffffffffa0ab2a10>] lod_sub_declare_create+0xe0/0x270 [lod]
> [410649.811633] [<ffffffffa0aabdc7>] lod_qos_declare_object_on+0xc7/0x3d0 [lod]
> [410649.811664] [<ffffffffa0aab7fe>] ? lod_statfs_and_check+0xae/0x5b0 [lod]
> [410649.811694] [<ffffffffa0aacfd4>] lod_alloc_qos.constprop.10+0xe64/0x17b0 [lod]
> [410649.811741] [<ffffffffa01741b0>] ? ldiskfs_map_blocks+0x180/0x1e0 [ldiskfs][410649.811772] [<ffffffffa0ab07ea>] lod_qos_prep_create+0x12ea/0x2910 [lod]
> [410649.811803] [<ffffffffa07c8c94>] ? qsd_op_begin+0x114/0x4d0 [lquota]
> [410649.811833] [<ffffffffa0ab23f0>] lod_prepare_create+0x2c0/0x410 [lod]
> [410649.811863] [<ffffffffa0aa7ccd>] lod_declare_striped_create+0x10d/0xa50 [lod]
> [410649.811908] [<ffffffffa0aaa9b9>] lod_declare_create+0x1e9/0x5a0 [lod]
> [410649.811938] [<ffffffffa0b1af26>] mdd_declare_create_object_internal+0x116/0x320 [mdd]
> [410649.811983] [<ffffffffa0b00c9c>] mdd_declare_create_object.isra.19+0x3c/0xbb0 [mdd]
> [410649.812028] [<ffffffffa0b00044>] ? mdd_linkea_prepare+0x294/0x590 [mdd]
> [410649.812058] [<ffffffffa0b0f90e>] mdd_create+0x88e/0x27d0 [mdd]
> [410649.812088] [<ffffffffa08173e0>] ? osd_xattr_get+0x80/0x890 [osd_ldiskfs]
> [410649.812120] [<ffffffffa09fd3ff>] mdt_reint_open+0x225f/0x3890 [mdt]
> [410649.812158] [<ffffffffa0431276>] ? null_alloc_rs+0x186/0x340 [ptlrpc]
> [410649.812191] [<ffffffffa02ea3fa>] ? upcall_cache_get_entry+0x29a/0x890 [obdclass]
> [410649.812237] [<ffffffffa02ef409>] ? lu_ucred+0x19/0x30 [obdclass]
> [410649.812267] [<ffffffffa09df7db>] ? ucred_set_jobid+0x5b/0x70 [mdt]
> [410649.812297] [<ffffffffa09f1810>] mdt_reint_rec+0xa0/0x210 [mdt]
> [410649.812326] [<ffffffffa09de91d>] mdt_reint_internal+0x63d/0xa50 [mdt]
> [410649.812356] [<ffffffffa09df07a>] mdt_intent_reint+0x21a/0x430 [mdt]
> [410649.812385] [<ffffffffa09da7ed>] mdt_intent_policy+0x5bd/0xde0 [mdt]
> [410649.812418] [<ffffffffa03aa257>] ldlm_lock_enqueue+0x3a7/0x9c0 [ptlrpc]
> [410649.812453] [<ffffffffa03d28e3>] ldlm_handle_enqueue0+0x9c3/0x1790 [ptlrpc]
> [410649.812490] [<ffffffffa04211c0>] ? req_capsule_client_get+0x10/0x20 [ptlrpc]
> [410649.812541] [<ffffffffa045d16c>] ? tgt_request_preprocess.isra.17+0x25c/0x1250 [ptlrpc]
> [410649.812593] [<ffffffffa044ee95>] ? tgt_lookup_reply+0x35/0x1c0 [ptlrpc]
> [410649.812629] [<ffffffffa045b74d>] tgt_enqueue+0x5d/0x250 [ptlrpc]
> [410649.812664] [<ffffffffa045ed1d>] tgt_request_handle+0x8ad/0x15a0 [ptlrpc]
> [410649.812701] [<ffffffffa03f84a4>] ? lustre_msg_get_transno+0x84/0x100 [ptlrpc]
> [410649.812752] [<ffffffffa04084e1>] ptlrpc_main+0x1051/0x2a40 [ptlrpc]
> [410649.812780] [<ffffffff8140ff44>] ? __schedule+0x294/0x940
> [410649.812815] [<ffffffffa0407490>] ? ptlrpc_main+0x0/0x2a40 [ptlrpc]
> [410649.812844] [<ffffffff8105c817>] kthread+0x87/0x90
> [410649.812870] [<ffffffff81413c34>] kernel_thread_helper+0x4/0x10
> [410649.812898] [<ffffffff8105c790>] ? kthread+0x0/0x90
> [410649.812923] [<ffffffff81413c30>] ? kernel_thread_helper+0x0/0x10
> [410649.812950]
> [410649.812970] LustreError: dumping log to /tmp/lustre-log.1553888141.3516
> [410649.818730] wanted to write 3985 but wrote 3518
> [410749.275270] LNet: Service thread pid 3516 completed after 299.99s. This indicates the system was overloaded (too many service threads, or there were not enough hardware resources).
> ==============================================================================
>
> Our lustre-2.10.6 was compiled with sles11sp3 kernel
>
> linux-3.0.101-138.gcdbe806
>
> using ldiskfs backend. The hardware spec of our MDS is:
>
> CPU: Intel Xeon E5640 @ 2.67GHz (single CPU)
> RAM: 8 GB
> MGS: 1GB (under RAID 1)
> MDT: 230GB (under RAID 1)
> RAID controller: LSI ServeRAID M1015 SAS/SATA Controller
>
>
> Is there any suggestion to fix this problem ?
>
> Thank you very much.
>
>
> T.H.Hsieh
> _______________________________________________
> lustre-discuss mailing list
> lustre-discuss at lists.lustre.org
> http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
More information about the lustre-discuss
mailing list