[lustre-discuss] FLR mirroring on 2.12.1-1

John Doe ace629426 at gmail.com
Fri May 24 06:28:51 PDT 2019


I have sent the log file to you in a separate email.

Note - I read four 1MB blocks, the first two 1MB blocks were cached.

On Fri, May 24, 2019 at 1:01 AM Jinshan Xiong <jinshan.xiong at gmail.com>
wrote:

> hmm.. This definitely is not expected. As long as ost 1 is down, it should
> be returned immediately from OSC layer and tries to read the 2nd mirror
> that is located on ost 7. For the following blocks, it should not even try
> ost1 but go to 7 directly.
>
> Would you please collect Lustre log and send it to me? You can collect
> logs on client side as follows:
> 0. create mirrored file
> 1. lctl set_param debug=-1 && lctl clear
> 2. lctl mark "======= start ========"
> 3. read the file
> 4. lctl dk > log.txt
>
> and send me the log.txt file. If you can reproduce this problem
> consistently, please use a small file so that it would be easier to check
> the log.
>
> Jinshan
>
> On Mon, May 20, 2019 at 6:20 AM John Doe <ace629426 at gmail.com> wrote:
>
>> It turns out that the read eventually finished and was 1/10th of the
>> performance that I was expecting.
>>
>> As ost idx 1 is unavailable, the client read has to timeout on ost idx 1
>> and then will read from ost idx 7. This happens for each 1MB block, as I am
>> using that as the block size.
>>
>> Is there a tunable to avoid this issue?
>>
>> lfs check osts also takes about 30 seconds as it times out on the
>> unavailable OST.
>>
>> Due to this issue, I am virtually unable to use the mirroring feature.
>>
>> I
>>
>> On Sun, May 19, 2019 at 4:27 PM John Doe <ace629426 at gmail.com> wrote:
>>
>>> After mirroring a file , when one mirror is down, any reads from a
>>> client just hangs. Both server and client are running latest 2.12.1-1.
>>> Client waits for ost idx 1 to come back online.  I am only unmounting ost
>>> idx1 not ost idx 7.
>>>
>>> Has anyone tried this feature?
>>>
>>> Thanks,
>>> John.
>>>
>>> lfs getstripe mirror10
>>> mirror10
>>>   lcm_layout_gen:    5
>>>   lcm_mirror_count:  2
>>>   lcm_entry_count:   2
>>>     lcme_id:             65537
>>>     lcme_mirror_id:      1
>>>     lcme_flags:          init
>>>     lcme_extent.e_start: 0
>>>     lcme_extent.e_end:   EOF
>>>       lmm_stripe_count:  1
>>>       lmm_stripe_size:   1048576
>>>       lmm_pattern:       raid0
>>>       lmm_layout_gen:    0
>>>       lmm_stripe_offset: 1
>>>       lmm_pool:          01
>>>       lmm_objects:
>>>       - 0: { l_ost_idx: 1, l_fid: [0x100010000:0x280a8:0x0] }
>>>
>>>     lcme_id:             131074
>>>     lcme_mirror_id:      2
>>>     lcme_flags:          init
>>>     lcme_extent.e_start: 0
>>>     lcme_extent.e_end:   EOF
>>>       lmm_stripe_count:  1
>>>       lmm_stripe_size:   1048576
>>>       lmm_pattern:       raid0
>>>       lmm_layout_gen:    0
>>>       lmm_stripe_offset: 7
>>>       lmm_pool:          02
>>>       lmm_objects:
>>>       - 0: { l_ost_idx: 7, l_fid: [0x100070000:0x28066:0x0] }
>>>
>> _______________________________________________
>> lustre-discuss mailing list
>> lustre-discuss at lists.lustre.org
>> http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.lustre.org/pipermail/lustre-discuss-lustre.org/attachments/20190524/857ad755/attachment-0001.html>


More information about the lustre-discuss mailing list